Advertisement
Guest User

Untitled

a guest
Mar 20th, 2019
131
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 2.06 KB | None | 0 0
  1. import org.apache.spark.{SparkConf, SparkContext}
  2. import org.apache.spark.sql.SparkSession
  3. import org.apache.spark.mllib.linalg.Vectors
  4. import org.apache.spark.rdd.RDD
  5. import org.apache.spark.mllib.stat.{MultivariateStatisticalSummary, Statistics}
  6.  
  7. object computeSxbph_3_30 {
  8. def main(args: Array[String]): Unit = {
  9. val conf = new SparkConf().setAppName("computeSxbph_3_30").setMaster("spark://192.168.41.143:7077")
  10. val sc = new SparkContext(conf)
  11. val input = sc.parallelize(List(
  12. List("001_2019_3_20", 1, 3, 5),
  13. List("001_2019_3_20", 4, 10, 15),
  14. List("001_2019_3_20", 41, 1203, 35),
  15. List("002_2019_3_20", 32, 4, 9),
  16. List("002_2019_3_20", 12, 24, 89),
  17. List("002_2019_3_20", 2, 5, 19),
  18. List("003_2019_3_20", 1, 1, 94),
  19. List("003_2019_3_20", 10, 91, 4),
  20. List("003_2019_3_20", 14, 21, 43)
  21. ), 3)
  22. val maped: RDD[(String, (Double, Double, Double))] = input.map {
  23. x => {
  24. val key = x(0).toString;
  25. val v1 = x(1).toString.toInt;
  26. val v2 = x(2).toString.toInt;
  27. val v3 = x(3).toString.toInt;
  28. (key, (v1, v2, v3))
  29. }
  30. }
  31. val topItem_set = maped.map(ele => (ele._1, (ele._2._1, ele._2._2, ele._2._3))).groupByKey()
  32. topItem_set.map(a=>{
  33. val v1 = Vectors.dense(a._2.toList(0)._1.toDouble,a._2.toList(0)._2.toDouble,a._2.toList(0)._3)
  34. val v2 = Vectors.dense(a._2.toList(1)._1.toDouble,a._2.toList(1)._2.toDouble,a._2.toList(1)._3)
  35. val v3 = Vectors.dense(a._2.toList(2)._1.toDouble,a._2.toList(2)._2.toDouble,a._2.toList(2)._3)
  36. val rows = sc.parallelize(Seq(v1,v2,v3))
  37. val tmp = Seq(v1,v2,v3).toArray
  38. tmp.foreach(println(_))
  39. println("____分割线______")
  40.  
  41. val summary: MultivariateStatisticalSummary = Statistics.colStats(rows)
  42. summary.max
  43. }).collect()
  44.  
  45. }
  46.  
  47. }
  48.  
  49. 19/03/20 23:47:25 ERROR TaskSetManager: Task 2 in stage 11.0 failed 1 times; aborting job
  50. org.apache.spark.SparkException: Job aborted due to stage failure: Task 2 in stage 11.0 failed 1 times, most recent failure:
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement