Advertisement
Guest User

Untitled

a guest
Sep 20th, 2017
65
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 0.37 KB | None | 0 0
  1. from pyspark.context import SparkContext
  2.  
  3. from pyspark.mllib.clustering import KMeans
  4. from pyspark.mllib.random import RandomRDDs
  5.  
  6.  
  7. if __name__ == "__main__":
  8. sc = SparkContext(appName='kmeansMinimalExample')
  9.  
  10. # same with 10000 points
  11. data = RandomRDDs.uniformVectorRDD(sc, 10000000, 64)
  12. C = KMeans.train(data, 8192, maxIterations=10)
  13.  
  14. sc.stop()
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement