Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- java.lang.NullPointerException
- at org.apache.hadoop.hive.ql.udf.generic.GenericUDFLeadLag.evaluate(GenericUDFLeadLag.java:57)
- at org.apache.spark.sql.hive.HiveGenericUDF.eval(hiveUDFs.scala:221)
- at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificMutableProjection.apply(Unknown Source)
- at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:373)
- at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:362)
- at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:141)
- at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:30)
- at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
- at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
- at scala.collection.Iterator$class.foreach(Iterator.scala:727)
- at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
- at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
- at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
- at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
- at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
- at scala.collection.AbstractIterator.to(Iterator.scala:1157)
- at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
- at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
- at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
- at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
- at org.apache.spark.scheduler.Task.run(Task.scala:88)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
- at java.lang.Thread.run(Thread.java:745)
- 15/11/02 15:31:24 INFO TaskSetManager: Starting task 49.0 in stage 5.0 (TID 53, localhost, PROCESS_LOCAL, 1914 bytes)
- 15/11/02 15:31:24 INFO Executor: Running task 49.0 in stage 5.0 (TID 53)
- 15/11/02 15:31:24 WARN TaskSetManager: Lost task 48.0 in stage 5.0 (TID 52, localhost): java.lang.NullPointerException
- at org.apache.hadoop.hive.ql.udf.generic.GenericUDFLeadLag.evaluate(GenericUDFLeadLag.java:57)
- at org.apache.spark.sql.hive.HiveGenericUDF.eval(hiveUDFs.scala:221)
- at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificMutableProjection.apply(Unknown Source)
- at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:373)
- at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:362)
- at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:141)
- at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:30)
- at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
- at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
- at scala.collection.Iterator$class.foreach(Iterator.scala:727)
- at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
- at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
- at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
- at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
- at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
- at scala.collection.AbstractIterator.to(Iterator.scala:1157)
- at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
- at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
- at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
- at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
- at org.apache.spark.scheduler.Task.run(Task.scala:88)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
- at java.lang.Thread.run(Thread.java:745)
- 15/11/02 15:31:24 ERROR TaskSetManager: Task 48 in stage 5.0 failed 1 times; aborting job
- 15/11/02 15:31:24 INFO ShuffleBlockFetcherIterator: Getting 1 non-empty blocks out of 1 blocks
- 15/11/02 15:31:24 INFO ShuffleBlockFetcherIterator: Started 0 remote fetches in 0 ms
- 15/11/02 15:31:24 INFO TaskSchedulerImpl: Cancelling stage 5
- 15/11/02 15:31:24 INFO TaskSchedulerImpl: Stage 5 was cancelled
- 15/11/02 15:31:24 INFO Executor: Executor is trying to kill task 49.0 in stage 5.0 (TID 53)
- 15/11/02 15:31:24 INFO DAGScheduler: ResultStage 5 (show at Main.scala:45) failed in 0.372 s
- 15/11/02 15:31:24 INFO DAGScheduler: Job 3 failed: show at Main.scala:45, took 0.408976 s
- 15/11/02 15:31:24 ERROR Executor: Exception in task 49.0 in stage 5.0 (TID 53)
- java.lang.NullPointerException
- at org.apache.hadoop.hive.ql.udf.generic.GenericUDFLeadLag.evaluate(GenericUDFLeadLag.java:57)
- at org.apache.spark.sql.hive.HiveGenericUDF.eval(hiveUDFs.scala:221)
- at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificMutableProjection.apply(Unknown Source)
- at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:373)
- at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:362)
- at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:141)
- at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:30)
- at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
- at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
- at scala.collection.Iterator$class.foreach(Iterator.scala:727)
- at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
- at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
- at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
- at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
- at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
- at scala.collection.AbstractIterator.to(Iterator.scala:1157)
- at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
- at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
- at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
- at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
- at org.apache.spark.scheduler.Task.run(Task.scala:88)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
- at java.lang.Thread.run(Thread.java:745)
- Exception in thread "main" org.apache.spark.SparkException: Job aborted due to stage failure: Task 48 in stage 5.0 failed 1 times, most recent failure: Lost task 48.0 in stage 5.0 (TID 52, localhost): java.lang.NullPointerException
- at org.apache.hadoop.hive.ql.udf.generic.GenericUDFLeadLag.evaluate(GenericUDFLeadLag.java:57)
- at org.apache.spark.sql.hive.HiveGenericUDF.eval(hiveUDFs.scala:221)
- at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificMutableProjection.apply(Unknown Source)
- at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:373)
- at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:362)
- at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:141)
- at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:30)
- at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
- at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
- at scala.collection.Iterator$class.foreach(Iterator.scala:727)
- at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
- at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
- at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
- at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
- at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
- at scala.collection.AbstractIterator.to(Iterator.scala:1157)
- at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
- at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
- at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
- at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
- at org.apache.spark.scheduler.Task.run(Task.scala:88)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
- at java.lang.Thread.run(Thread.java:745)
- Driver stacktrace:
- at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1283)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1271)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1270)
- at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
- at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
- at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1270)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
- at scala.Option.foreach(Option.scala:236)
- at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:697)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1496)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1458)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1447)
- at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:567)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:1822)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:1835)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:1848)
- at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:215)
- at org.apache.spark.sql.execution.Limit.executeCollect(basicOperators.scala:207)
- at org.apache.spark.sql.DataFrame$$anonfun$collect$1.apply(DataFrame.scala:1385)
- at org.apache.spark.sql.DataFrame$$anonfun$collect$1.apply(DataFrame.scala:1385)
- at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:56)
- at org.apache.spark.sql.DataFrame.withNewExecutionId(DataFrame.scala:1903)
- at org.apache.spark.sql.DataFrame.collect(DataFrame.scala:1384)
- at org.apache.spark.sql.DataFrame.head(DataFrame.scala:1314)
- at org.apache.spark.sql.DataFrame.take(DataFrame.scala:1377)
- at org.apache.spark.sql.DataFrame.showString(DataFrame.scala:178)
- at org.apache.spark.sql.DataFrame.show(DataFrame.scala:401)
- at org.apache.spark.sql.DataFrame.show(DataFrame.scala:362)
- at org.apache.spark.sql.DataFrame.show(DataFrame.scala:370)
- at main.scala.expression.Main$.main(Main.scala:45)
- at main.scala.expression.Main.main(Main.scala)
- Caused by: java.lang.NullPointerException
- at org.apache.hadoop.hive.ql.udf.generic.GenericUDFLeadLag.evaluate(GenericUDFLeadLag.java:57)
- at org.apache.spark.sql.hive.HiveGenericUDF.eval(hiveUDFs.scala:221)
- at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificMutableProjection.apply(Unknown Source)
- at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:373)
- at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:362)
- at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:141)
- at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:30)
- at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
- at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
- at scala.collection.Iterator$class.foreach(Iterator.scala:727)
- at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
- at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
- at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
- at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
- at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
- at scala.collection.AbstractIterator.to(Iterator.scala:1157)
- at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
- at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
- at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
- at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
- at org.apache.spark.scheduler.Task.run(Task.scala:88)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
- at java.lang.Thread.run(Thread.java:745)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement