Advertisement
Guest User

Untitled

a guest
Apr 6th, 2020
185
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
SPARK 1.29 KB | None | 0 0
  1. import org.apache.spark
  2. import org.apache.spark.sql.SparkSession
  3.  
  4. import scala.io.{BufferedSource, Source}
  5. import org.apache.spark.{SparkConf, SparkContext}
  6.  
  7. object extractFile {
  8.   def main(args: Array[String]): Unit = {
  9.     System.setProperty("hadoop.home.dir", "C:\\winutil\\")
  10.     val conf = new SparkConf()
  11.       .setMaster("local")
  12.       .setAppName("loadcsv")
  13.  
  14.     val sc = new SparkContext(conf) // Create Spark Context
  15.     // Load local file data
  16.     val spark : SparkSession = SparkSession.builder
  17.       .appName("test")
  18.       .master("local[2]")
  19.       .getOrCreate()
  20.    
  21.    
  22.     val sourceDf = spark.read.format("csv").option("header", "true").load("C:\\Users\\test\\parking_violations.csv")
  23.     val df1 = sourceDf.select("Plate ID", "Violation description", "House Number","Street Name").toDF()
  24.  
  25.     val randomRow = df1.sample(false,0.1).toDF()
  26.  
  27.     randomRow.show(1)
  28.  
  29.     val plate_id = randomRow.first().getString(0)
  30.     val violation_description = randomRow.first().getString(1)
  31.     val house_number = randomRow.first().getString(2)
  32.     val street_name = randomRow.first().getString(3)
  33.  
  34.     if (plate_id.equals(null)){
  35.       print("alert")
  36.     }
  37.     else{
  38.       print("le plate id est: "+ plate_id )
  39.     }
  40.  
  41.     //df1.where(df1.col("Plate ID").isNull).show()
  42.   }
  43.  
  44. }
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement