Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- /opt/spark/bin/spark-submit --class com.myntelligence.text.tranformation.ScrapperService --master "local[*]" --driver-memory ${SPARK_DRIVER_MEMORY} --driver-java-options="-Dapplication.spark.master=spark://${HOSTNAME}:7077 -Dapplication.amqp.host=${RABBITMQ_HOST} -Dapplication.amqp.port=${RABBITMQ_PORT} -Dapplication.amqp.username=${RABBITMQ_USER} -Dapplication.amqp.password=${RABBITMQ_PASSWORD}" /spark-jobs/myntelligence-text-transformation.jar
- Exception in thread "main" org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 4.0 failed 4 times, most recent failure: Lost task 0.3 in stage 4.0 (TID 75, 172.19.0.8, executor 0): com.typesafe.config.ConfigException$Missing: No configuration setting found for key 'host'
- at com.typesafe.config.impl.SimpleConfig.findKeyOrNull(SimpleConfig.java:152)
- at com.typesafe.config.impl.SimpleConfig.findOrNull(SimpleConfig.java:170)
- at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:184)
- at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:189)
- at com.typesafe.config.impl.SimpleConfig.getString(SimpleConfig.java:246)
- at io.scalac.amqp.ConnectionSettings$$anonfun$apply$1.apply(ConnectionSettings.scala:49)
- at io.scalac.amqp.ConnectionSettings$$anonfun$apply$1.apply(ConnectionSettings.scala:47)
- at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
- at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
- at scala.collection.Iterator$class.foreach(Iterator.scala:893)
- at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
- at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
- at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
- at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
- at scala.collection.AbstractTraversable.map(Traversable.scala:104)
- at io.scalac.amqp.ConnectionSettings$.apply(ConnectionSettings.scala:51)
- at io.scalac.amqp.Connection$.apply(Connection.scala:15)
- at com.myntelligence.text.tranformation.ScrapperService$.handler(ScrapperService.scala:94)
- at com.myntelligence.text.tranformation.ScrapperService$$anonfun$main$1$$anonfun$apply$1.apply(ScrapperService.scala:82)
- at com.myntelligence.text.tranformation.ScrapperService$$anonfun$main$1$$anonfun$apply$1.apply(ScrapperService.scala:82)
- at scala.collection.Iterator$class.foreach(Iterator.scala:893)
- at org.apache.spark.util.CompletionIterator.foreach(CompletionIterator.scala:26)
- at org.apache.spark.rdd.RDD$$anonfun$foreach$1$$anonfun$apply$28.apply(RDD.scala:918)
- at org.apache.spark.rdd.RDD$$anonfun$foreach$1$$anonfun$apply$28.apply(RDD.scala:918)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1951)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1951)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
- at org.apache.spark.scheduler.Task.run(Task.scala:99)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:322)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- Driver stacktrace:
- at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1435)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1423)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1422)
- at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
- at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1422)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:802)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:802)
- at scala.Option.foreach(Option.scala:257)
- at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:802)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1650)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1605)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1594)
- at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:628)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:1925)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:1938)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:1951)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:1965)
- at org.apache.spark.rdd.RDD$$anonfun$foreach$1.apply(RDD.scala:918)
- at org.apache.spark.rdd.RDD$$anonfun$foreach$1.apply(RDD.scala:916)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.RDD.foreach(RDD.scala:916)
- at com.myntelligence.text.tranformation.ScrapperService$$anonfun$main$1.apply(ScrapperService.scala:82)
- at com.myntelligence.text.tranformation.ScrapperService$$anonfun$main$1.apply(ScrapperService.scala:80)
- at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:627)
- at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:627)
- at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ForEachDStream.scala:51)
- at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:51)
- at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:51)
- at org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:415)
- at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:50)
- at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:50)
- at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:50)
- at scala.util.Try$.apply(Try.scala:192)
- at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39)
- at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply$mcV$sp(JobScheduler.scala:256)
- at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:256)
- at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:256)
- at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58)
- at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobScheduler.scala:255)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- Caused by: com.typesafe.config.ConfigException$Missing: No configuration setting found for key 'host'
- at com.typesafe.config.impl.SimpleConfig.findKeyOrNull(SimpleConfig.java:152)
- at com.typesafe.config.impl.SimpleConfig.findOrNull(SimpleConfig.java:170)
- at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:184)
- at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:189)
- at com.typesafe.config.impl.SimpleConfig.getString(SimpleConfig.java:246)
- at io.scalac.amqp.ConnectionSettings$$anonfun$apply$1.apply(ConnectionSettings.scala:49)
- at io.scalac.amqp.ConnectionSettings$$anonfun$apply$1.apply(ConnectionSettings.scala:47)
- at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
- at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
- at scala.collection.Iterator$class.foreach(Iterator.scala:893)
- at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
- at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
- at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
- at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
- at scala.collection.AbstractTraversable.map(Traversable.scala:104)
- at io.scalac.amqp.ConnectionSettings$.apply(ConnectionSettings.scala:51)
- at io.scalac.amqp.Connection$.apply(Connection.scala:15)
- at com.myntelligence.text.tranformation.ScrapperService$.handler(ScrapperService.scala:94)
- at com.myntelligence.text.tranformation.ScrapperService$$anonfun$main$1$$anonfun$apply$1.apply(ScrapperService.scala:82)
- at com.myntelligence.text.tranformation.ScrapperService$$anonfun$main$1$$anonfun$apply$1.apply(ScrapperService.scala:82)
- at scala.collection.Iterator$class.foreach(Iterator.scala:893)
- at org.apache.spark.util.CompletionIterator.foreach(CompletionIterator.scala:26)
- at org.apache.spark.rdd.RDD$$anonfun$foreach$1$$anonfun$apply$28.apply(RDD.scala:918)
- at org.apache.spark.rdd.RDD$$anonfun$foreach$1$$anonfun$apply$28.apply(RDD.scala:918)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1951)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1951)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
- at org.apache.spark.scheduler.Task.run(Task.scala:99)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:322)
- ... 3 more
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement