Guest User

Untitled

a guest
Jul 23rd, 2018
84
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 0.93 KB | None | 0 0
  1. import geomesa_pyspark
  2.  
  3. from pyspark.conf import SparkConf
  4.  
  5. conf = SparkConf()
  6.  
  7. # Configure OPTIONS
  8. conf.get('spark.master')
  9. conf.set("spark.yarn.stagingDir","hdfs://ip-10-121-3-100.ec2.internal/user/zeppelin/")
  10.  
  11. from pyspark.conf import SparkConf
  12. from pyspark.sql import SparkSession
  13. from pyspark.sql import SQLContext
  14.  
  15. dsParams = {"hbase.catalog":"geomesa.adsb"}
  16. INCREMENT_SECS = 300
  17.  
  18. sparkSession = SparkSession.builder.config(conf=conf).getOrCreate()
  19.  
  20. sc2 = sparkSession.sparkContext
  21. sc.setLogLevel('DEBUG')
  22.  
  23. TRACKS_NAME = "csmtrackstest17"
  24. parquetTrksDF = (sparkSession.read.format("geomesa").options(**dsParams).option("geomesa.feature", TRACKS_NAME).load())
  25. parquetTrksDF.createOrReplaceTempView (TRACKS_NAME)
  26. parquetTrksDF.printSchema()
  27.  
  28. from geomesa_pyspark.spark import GeoMesaSpark
  29. gms = GeoMesaSpark(spark._sc)
  30. query = "BBOX(geom, -78.0, -34.0, -77.0, 33.0)"
  31.  
  32. feature = "csmtrackstest17"
  33. rdd_dict = gms.apply(dsParams).rdd_dict(feature, query)
Add Comment
Please, Sign In to add comment