Advertisement
Guest User

Untitled

a guest
Jan 11th, 2017
105
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 1.06 KB | None | 0 0
  1. package com.spnotes.spark
  2.  
  3. import java.sql.{Connection, DriverManager, ResultSet}
  4.  
  5. import org.apache.spark.rdd.JdbcRDD
  6. import org.apache.spark.{SparkConf, SparkContext}
  7.  
  8. /**
  9. * Created by sunilpatil on 4/19/16.
  10. */
  11. object JDBCRDDClient {
  12.  
  13. case class Address(addressId: Int, contactId: Int, line1: String, city: String, state: String, zip: String)
  14.  
  15. def main(argv: Array[String]): Unit = {
  16. val sparkConf = new SparkConf().setMaster("local[2]").setAppName("HelloJDBC")
  17. val sparkContext = new SparkContext(sparkConf)
  18.  
  19. val jdbcRdd = new JdbcRDD(sparkContext, getConnection,
  20. "select * from address limit ?,?",
  21. 0, 5, 1, convertToAddress)
  22.  
  23. jdbcRdd.foreach(println)
  24. }
  25.  
  26. def getConnection(): Connection = {
  27. Class.forName("com.mysql.jdbc.Driver")
  28. DriverManager.getConnection("jdbc:mysql://localhost/test1?" + "user=test1&password=test1")
  29. }
  30.  
  31. def convertToAddress(rs: ResultSet): Address = {
  32. new Address(rs.getInt("addressid"), rs.getInt("contactid"), rs.getString("line1"), rs.getString("city"), rs.getString("state"), rs.getString("zip"))
  33. }
  34. }
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement