Advertisement
Guest User

Untitled

a guest
Dec 8th, 2016
169
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 3.48 KB | None | 0 0
  1. message.max.bytes=60000000
  2. replica.fetch.max.bytes=120000000
  3.  
  4. val kafkaParams = Map[String, String](
  5. "security.protocol" -> "SASL_PLAINTEXT",
  6. "group.id" -> groupid,
  7. "metadata.broker.list" -> kafkaBrokerList,
  8. "max.partition.fetch.bytes" -> "60000000")
  9.  
  10. spark-submit
  11. --verbose
  12. --master yarn-cluster
  13. --num-executors 3
  14. --executor-memory 7g
  15. --executor-cores 3
  16. --conf spark.driver.memory=1024m
  17. --conf spark.streaming.backpressure.enabled=false
  18. --conf spark.streaming.kafka.maxRatePerPartition=3
  19. --conf spark.streaming.concurrentJobs=3
  20. --conf spark.speculation=true
  21. --conf spark.hadoop.fs.hdfs.impl.disable.cache=true
  22. --files kafka_jaas.conf#kafka_jaas.conf,user.headless.keytab#user.headless.keytab
  23. --driver-java-options "-Djava.security.auth.login.config=./kafka_jaas.conf -Dhttp.proxyHost=PROXY_IP -Dhttp.proxyPort=8080 -Dhttps.proxyHost=PROXY_IP -Dhttps.proxyPort=8080 -Dlog4j.configuration=file:/home/user/spark-log4j/log4j-topic_name-driver.properties"
  24. --conf "spark.executor.extraJavaOptions=-Djava.security.auth.login.config=./kafka_jaas.conf -Dlog4j.configuration=file:/home/user/spark-log4j/log4j-topic_name-executor.properties"
  25. --class com.spark.demo.StreamProcessor /home/user/demo.jar /tmp/data/out 60 KAFKA_BROKER:6667 "groupid" topic_name
  26.  
  27. User class threw exception: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 3, IP_HOST): java.lang.AssertionError: assertion failed: Ran out of messages before reaching ending offset 197 for topic x_topic_3 partition 24 start 196. This should not happen, and indicates that messages may have been lost
  28. at scala.Predef$.assert(Predef.scala:179)
  29. at org.apache.spark.streaming.kafka.KafkaRDD$KafkaRDDIterator.getNext(KafkaRDD.scala:211)
  30. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  31. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  32. at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:308)
  33. at scala.collection.Iterator$class.foreach(Iterator.scala:727)
  34. at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
  35. at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
  36. at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
  37. at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
  38. at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
  39. at scala.collection.AbstractIterator.to(Iterator.scala:1157)
  40. at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
  41. at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
  42. at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
  43. at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
  44. at org.apache.spark.rdd.RDD$$anonfun$take$1$$anonfun$29.apply(RDD.scala:1335)
  45. at org.apache.spark.rdd.RDD$$anonfun$take$1$$anonfun$29.apply(RDD.scala:1335)
  46. at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1881)
  47. at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1881)
  48. at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
  49. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  50. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
  51. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  52. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  53. at java.lang.Thread.run(Thread.java:745)
  54. Driver stacktrace
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement