Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- User class threw exception: org.apache.spark.SparkException: Job aborted.
- at org.apache.spark.internal.io.SparkHadoopWriter$.write(SparkHadoopWriter.scala:96)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1096)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopDataset$1.apply(PairRDDFunctions.scala:1094)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopDataset$1.apply(PairRDDFunctions.scala:1094)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:363)
- at org.apache.spark.rdd.PairRDDFunctions.saveAsHadoopDataset(PairRDDFunctions.scala:1094)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$4.apply$mcV$sp(PairRDDFunctions.scala:1067)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$4.apply(PairRDDFunctions.scala:1032)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$4.apply(PairRDDFunctions.scala:1032)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:363)
- at org.apache.spark.rdd.PairRDDFunctions.saveAsHadoopFile(PairRDDFunctions.scala:1032)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$1.apply$mcV$sp(PairRDDFunctions.scala:958)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$1.apply(PairRDDFunctions.scala:958)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$1.apply(PairRDDFunctions.scala:958)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:363)
- at org.apache.spark.rdd.PairRDDFunctions.saveAsHadoopFile(PairRDDFunctions.scala:957)
- at org.apache.spark.rdd.RDD$$anonfun$saveAsTextFile$1.apply$mcV$sp(RDD.scala:1493)
- at org.apache.spark.rdd.RDD$$anonfun$saveAsTextFile$1.apply(RDD.scala:1472)
- at org.apache.spark.rdd.RDD$$anonfun$saveAsTextFile$1.apply(RDD.scala:1472)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:363)
- at org.apache.spark.rdd.RDD.saveAsTextFile(RDD.scala:1472)
- at org.apache.spark.api.java.JavaRDDLike$class.saveAsTextFile(JavaRDDLike.scala:550)
- at org.apache.spark.api.java.AbstractJavaRDDLike.saveAsTextFile(JavaRDDLike.scala:45)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at clojure.lang.Reflector.invokeMatchingMethod(Reflector.java:93)
- at clojure.lang.Reflector.invokeInstanceMethod(Reflector.java:28)
- at flambo.api$save_as_text_file.invokeStatic(api.clj:434)
- at flambo.api$save_as_text_file.invoke(api.clj:428)
- at place_attach_cli.place_attach_spark$save_txt.invokeStatic(place_attach_spark.clj:141)
- at place_attach_cli.place_attach_spark$save_txt.invoke(place_attach_spark.clj:138)
- at place_attach_cli.place_attach_spark$handle_queries.invokeStatic(place_attach_spark.clj:152)
- at place_attach_cli.place_attach_spark$handle_queries.invoke(place_attach_spark.clj:143)
- at place_attach_cli.spark_cli$_main.invokeStatic(spark_cli.clj:73)
- at place_attach_cli.spark_cli$_main.doInvoke(spark_cli.clj:67)
- at clojure.lang.RestFn.applyTo(RestFn.java:137)
- at place_attach_cli.spark_cli.main(Unknown Source)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.spark.deploy.yarn.ApplicationMaster$$anon$4.run(ApplicationMaster.scala:721)
- Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 4.0 failed 4 times, most recent failure: Lost task 0.3 in stage 4.0 (TID 409, d572.la.prod.factual.com, executor 20): org.apache.spark.SparkException: Task failed while writing rows
- at org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:151)
- at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:79)
- at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:78)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
- at org.apache.spark.scheduler.Task.run(Task.scala:109)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.lang.NoSuchMethodError: org.apache.thrift.TBaseHelper.hashCode(J)I
- at com.factual.tile.api.v3.IDWrapper.hashCode(IDWrapper.java:268)
- at java.util.HashMap.hash(HashMap.java:339)
- at java.util.HashMap.put(HashMap.java:612)
- at java.util.HashSet.add(HashSet.java:220)
- at com.factual.tile.api.v3.FactualPlaceData$FactualPlaceDataStandardScheme.read(FactualPlaceData.java:1928)
- at com.factual.tile.api.v3.FactualPlaceData$FactualPlaceDataStandardScheme.read(FactualPlaceData.java:1751)
- at com.factual.tile.api.v3.FactualPlaceData.read(FactualPlaceData.java:1541)
- at com.factual.tile.api.v3.Feature$FeatureStandardScheme.read(Feature.java:724)
- at com.factual.tile.api.v3.Feature$FeatureStandardScheme.read(Feature.java:709)
- at com.factual.tile.api.v3.Feature.read(Feature.java:589)
- at com.factual.tile.api.v3.Tile$TileStandardScheme.read(Tile.java:424)
- at com.factual.tile.api.v3.Tile$TileStandardScheme.read(Tile.java:395)
- at com.factual.tile.api.v3.Tile.read(Tile.java:332)
- at org.apache.thrift.TDeserializer.deserialize(TDeserializer.java:81)
- at org.apache.thrift.TDeserializer.deserialize(TDeserializer.java:67)
- at com.factual.tile.api.Utils.decodeThrift(Utils.java:42)
- at com.factual.tile.adaptor.v3.TileAdaptor.fromThriftBytes(TileAdaptor.java:44)
- at com.factual.tile.adaptor.v3.TileAdaptor.fromTileBuffer(TileAdaptor.java:55)
- at com.factual.tile.adaptor.TileAdaptor.fromTileBuffer(TileAdaptor.java:28)
- at com.factual.tile.reader.TileSet.lambda$new$0(TileSet.java:19)
- at com.google.common.collect.Iterators$8.transform(Iterators.java:860)
- at com.google.common.collect.TransformedIterator.next(TransformedIterator.java:48)
- at com.factual.placeattach.PlaceAttachmentHDFS.preloadTiles(PlaceAttachmentHDFS.java:92)
- at com.factual.placeattach.PlaceAttachmentHDFS.ensureTiles(PlaceAttachmentHDFS.java:71)
- at com.factual.placeattach.PlaceAttachment.placeAttach(PlaceAttachment.java:153)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at clojure.lang.Reflector.invokeMatchingMethod(Reflector.java:93)
- at clojure.lang.Reflector.invokeInstanceMethod(Reflector.java:28)
- at place_attach_cli.place_attach_spark$perform_pa_single.invokeStatic(place_attach_spark.clj:45)
- at place_attach_cli.place_attach_spark$perform_pa_single.invoke(place_attach_spark.clj:39)
- at clojure.core$partial$fn__4763.invoke(core.clj:2529)
- at place_attach_cli.place_attach_spark$eval1867$place_attach_cli_DOT_place_attach_spark_sfn____1868.invoke(NO_SOURCE_FILE:68)
- at clojure.lang.AFn.applyToHelper(AFn.java:154)
- at clojure.lang.AFn.applyTo(AFn.java:144)
- at clojure.lang.AFunction$1.doInvoke(AFunction.java:29)
- at clojure.lang.RestFn.applyTo(RestFn.java:137)
- at clojure.core$apply.invokeStatic(core.clj:646)
- at clojure.core$apply.invoke(core.clj:641)
- at flambo.function$Function_call.invokeStatic(function.clj:57)
- at flambo.function$Function_call.doInvoke(function.clj:57)
- at clojure.lang.RestFn.invoke(RestFn.java:423)
- at flambo.function.Function.call(Unknown Source)
- at org.apache.spark.api.java.JavaPairRDD$$anonfun$toScalaFunction$1.apply(JavaPairRDD.scala:1040)
- at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
- at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
- at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
- at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:125)
- at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:123)
- at org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1414)
- at org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:135)
- ... 8 more
- Driver stacktrace:
- at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1602)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1590)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1589)
- at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
- at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1589)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:831)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:831)
- at scala.Option.foreach(Option.scala:257)
- at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:831)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1823)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1772)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1761)
- at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:642)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2034)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2055)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2087)
- at org.apache.spark.internal.io.SparkHadoopWriter$.write(SparkHadoopWriter.scala:78)
- ... 51 more
- Caused by: org.apache.spark.SparkException: Task failed while writing rows
- at org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:151)
- at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:79)
- at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:78)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
- at org.apache.spark.scheduler.Task.run(Task.scala:109)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.lang.NoSuchMethodError: org.apache.thrift.TBaseHelper.hashCode(J)I
- at com.factual.tile.api.v3.IDWrapper.hashCode(IDWrapper.java:268)
- at java.util.HashMap.hash(HashMap.java:339)
- at java.util.HashMap.put(HashMap.java:612)
- at java.util.HashSet.add(HashSet.java:220)
- at com.factual.tile.api.v3.FactualPlaceData$FactualPlaceDataStandardScheme.read(FactualPlaceData.java:1928)
- at com.factual.tile.api.v3.FactualPlaceData$FactualPlaceDataStandardScheme.read(FactualPlaceData.java:1751)
- at com.factual.tile.api.v3.FactualPlaceData.read(FactualPlaceData.java:1541)
- at com.factual.tile.api.v3.Feature$FeatureStandardScheme.read(Feature.java:724)
- at com.factual.tile.api.v3.Feature$FeatureStandardScheme.read(Feature.java:709)
- at com.factual.tile.api.v3.Feature.read(Feature.java:589)
- at com.factual.tile.api.v3.Tile$TileStandardScheme.read(Tile.java:424)
- at com.factual.tile.api.v3.Tile$TileStandardScheme.read(Tile.java:395)
- at com.factual.tile.api.v3.Tile.read(Tile.java:332)
- at org.apache.thrift.TDeserializer.deserialize(TDeserializer.java:81)
- at org.apache.thrift.TDeserializer.deserialize(TDeserializer.java:67)
- at com.factual.tile.api.Utils.decodeThrift(Utils.java:42)
- at com.factual.tile.adaptor.v3.TileAdaptor.fromThriftBytes(TileAdaptor.java:44)
- at com.factual.tile.adaptor.v3.TileAdaptor.fromTileBuffer(TileAdaptor.java:55)
- at com.factual.tile.adaptor.TileAdaptor.fromTileBuffer(TileAdaptor.java:28)
- at com.factual.tile.reader.TileSet.lambda$new$0(TileSet.java:19)
- at com.google.common.collect.Iterators$8.transform(Iterators.java:860)
- at com.google.common.collect.TransformedIterator.next(TransformedIterator.java:48)
- at com.factual.placeattach.PlaceAttachmentHDFS.preloadTiles(PlaceAttachmentHDFS.java:92)
- at com.factual.placeattach.PlaceAttachmentHDFS.ensureTiles(PlaceAttachmentHDFS.java:71)
- at com.factual.placeattach.PlaceAttachment.placeAttach(PlaceAttachment.java:153)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at clojure.lang.Reflector.invokeMatchingMethod(Reflector.java:93)
- at clojure.lang.Reflector.invokeInstanceMethod(Reflector.java:28)
- at place_attach_cli.place_attach_spark$perform_pa_single.invokeStatic(place_attach_spark.clj:45)
- at place_attach_cli.place_attach_spark$perform_pa_single.invoke(place_attach_spark.clj:39)
- at clojure.core$partial$fn__4763.invoke(core.clj:2529)
- at place_attach_cli.place_attach_spark$eval1867$place_attach_cli_DOT_place_attach_spark_sfn____1868.invoke(NO_SOURCE_FILE:68)
- at clojure.lang.AFn.applyToHelper(AFn.java:154)
- at clojure.lang.AFn.applyTo(AFn.java:144)
- at clojure.lang.AFunction$1.doInvoke(AFunction.java:29)
- at clojure.lang.RestFn.applyTo(RestFn.java:137)
- at clojure.core$apply.invokeStatic(core.clj:646)
- at clojure.core$apply.invoke(core.clj:641)
- at flambo.function$Function_call.invokeStatic(function.clj:57)
- at flambo.function$Function_call.doInvoke(function.clj:57)
- at clojure.lang.RestFn.invoke(RestFn.java:423)
- at flambo.function.Function.call(Unknown Source)
- at org.apache.spark.api.java.JavaPairRDD$$anonfun$toScalaFunction$1.apply(JavaPairRDD.scala:1040)
- at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
- at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
- at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
- at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:125)
- at org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:123)
- at org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1414)
- at org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:135)
- ... 8 more
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement