Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- import geomesa_pyspark
- from pyspark.conf import SparkConf
- conf = SparkConf()
- # Configure OPTIONS
- conf.get('spark.master')
- conf.set("spark.yarn.stagingDir","hdfs://ip-10-121-3-100.ec2.internal/user/zeppelin/")
- from pyspark.conf import SparkConf
- from pyspark.sql import SparkSession
- from pyspark.sql import SQLContext
- dsParams = {"hbase.catalog":"geomesa.adsb"}
- INCREMENT_SECS = 300
- sparkSession = SparkSession.builder.config(conf=conf).getOrCreate()
- sc2 = sparkSession.sparkContext
- sc.setLogLevel('DEBUG')
- TRACKS_NAME = "csmtrackstest17"
- parquetTrksDF = (sparkSession.read.format("geomesa").options(**dsParams).option("geomesa.feature", TRACKS_NAME).load())
- parquetTrksDF.createOrReplaceTempView (TRACKS_NAME)
- parquetTrksDF.printSchema()
- from geomesa_pyspark.spark import GeoMesaSpark
- gms = GeoMesaSpark(spark._sc)
- query = "BBOX(geom, -78.0, -34.0, -77.0, 33.0)"
- feature = "csmtrackstest17"
- rdd_dict = gms.apply(dsParams).rdd_dict(feature, query)
Add Comment
Please, Sign In to add comment