Advertisement
Guest User

Untitled

a guest
Nov 2nd, 2015
170
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 15.80 KB | None | 0 0
  1. java.lang.NullPointerException
  2. at org.apache.hadoop.hive.ql.udf.generic.GenericUDFLeadLag.evaluate(GenericUDFLeadLag.java:57)
  3. at org.apache.spark.sql.hive.HiveGenericUDF.eval(hiveUDFs.scala:221)
  4. at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificMutableProjection.apply(Unknown Source)
  5. at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:373)
  6. at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:362)
  7. at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:141)
  8. at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:30)
  9. at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
  10. at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
  11. at scala.collection.Iterator$class.foreach(Iterator.scala:727)
  12. at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
  13. at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
  14. at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
  15. at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
  16. at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
  17. at scala.collection.AbstractIterator.to(Iterator.scala:1157)
  18. at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
  19. at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
  20. at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
  21. at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
  22. at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
  23. at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
  24. at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
  25. at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
  26. at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
  27. at org.apache.spark.scheduler.Task.run(Task.scala:88)
  28. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
  29. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  30. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
  31. at java.lang.Thread.run(Thread.java:745)
  32. 15/11/02 15:31:24 INFO TaskSetManager: Starting task 49.0 in stage 5.0 (TID 53, localhost, PROCESS_LOCAL, 1914 bytes)
  33. 15/11/02 15:31:24 INFO Executor: Running task 49.0 in stage 5.0 (TID 53)
  34. 15/11/02 15:31:24 WARN TaskSetManager: Lost task 48.0 in stage 5.0 (TID 52, localhost): java.lang.NullPointerException
  35. at org.apache.hadoop.hive.ql.udf.generic.GenericUDFLeadLag.evaluate(GenericUDFLeadLag.java:57)
  36. at org.apache.spark.sql.hive.HiveGenericUDF.eval(hiveUDFs.scala:221)
  37. at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificMutableProjection.apply(Unknown Source)
  38. at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:373)
  39. at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:362)
  40. at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:141)
  41. at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:30)
  42. at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
  43. at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
  44. at scala.collection.Iterator$class.foreach(Iterator.scala:727)
  45. at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
  46. at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
  47. at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
  48. at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
  49. at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
  50. at scala.collection.AbstractIterator.to(Iterator.scala:1157)
  51. at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
  52. at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
  53. at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
  54. at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
  55. at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
  56. at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
  57. at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
  58. at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
  59. at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
  60. at org.apache.spark.scheduler.Task.run(Task.scala:88)
  61. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
  62. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  63. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
  64. at java.lang.Thread.run(Thread.java:745)
  65.  
  66. 15/11/02 15:31:24 ERROR TaskSetManager: Task 48 in stage 5.0 failed 1 times; aborting job
  67. 15/11/02 15:31:24 INFO ShuffleBlockFetcherIterator: Getting 1 non-empty blocks out of 1 blocks
  68. 15/11/02 15:31:24 INFO ShuffleBlockFetcherIterator: Started 0 remote fetches in 0 ms
  69. 15/11/02 15:31:24 INFO TaskSchedulerImpl: Cancelling stage 5
  70. 15/11/02 15:31:24 INFO TaskSchedulerImpl: Stage 5 was cancelled
  71. 15/11/02 15:31:24 INFO Executor: Executor is trying to kill task 49.0 in stage 5.0 (TID 53)
  72. 15/11/02 15:31:24 INFO DAGScheduler: ResultStage 5 (show at Main.scala:45) failed in 0.372 s
  73. 15/11/02 15:31:24 INFO DAGScheduler: Job 3 failed: show at Main.scala:45, took 0.408976 s
  74. 15/11/02 15:31:24 ERROR Executor: Exception in task 49.0 in stage 5.0 (TID 53)
  75. java.lang.NullPointerException
  76. at org.apache.hadoop.hive.ql.udf.generic.GenericUDFLeadLag.evaluate(GenericUDFLeadLag.java:57)
  77. at org.apache.spark.sql.hive.HiveGenericUDF.eval(hiveUDFs.scala:221)
  78. at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificMutableProjection.apply(Unknown Source)
  79. at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:373)
  80. at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:362)
  81. at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:141)
  82. at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:30)
  83. at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
  84. at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
  85. at scala.collection.Iterator$class.foreach(Iterator.scala:727)
  86. at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
  87. at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
  88. at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
  89. at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
  90. at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
  91. at scala.collection.AbstractIterator.to(Iterator.scala:1157)
  92. at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
  93. at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
  94. at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
  95. at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
  96. at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
  97. at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
  98. at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
  99. at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
  100. at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
  101. at org.apache.spark.scheduler.Task.run(Task.scala:88)
  102. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
  103. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  104. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
  105. at java.lang.Thread.run(Thread.java:745)
  106. Exception in thread "main" org.apache.spark.SparkException: Job aborted due to stage failure: Task 48 in stage 5.0 failed 1 times, most recent failure: Lost task 48.0 in stage 5.0 (TID 52, localhost): java.lang.NullPointerException
  107. at org.apache.hadoop.hive.ql.udf.generic.GenericUDFLeadLag.evaluate(GenericUDFLeadLag.java:57)
  108. at org.apache.spark.sql.hive.HiveGenericUDF.eval(hiveUDFs.scala:221)
  109. at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificMutableProjection.apply(Unknown Source)
  110. at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:373)
  111. at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:362)
  112. at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:141)
  113. at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:30)
  114. at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
  115. at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
  116. at scala.collection.Iterator$class.foreach(Iterator.scala:727)
  117. at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
  118. at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
  119. at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
  120. at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
  121. at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
  122. at scala.collection.AbstractIterator.to(Iterator.scala:1157)
  123. at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
  124. at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
  125. at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
  126. at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
  127. at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
  128. at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
  129. at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
  130. at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
  131. at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
  132. at org.apache.spark.scheduler.Task.run(Task.scala:88)
  133. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
  134. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  135. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
  136. at java.lang.Thread.run(Thread.java:745)
  137.  
  138. Driver stacktrace:
  139. at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1283)
  140. at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1271)
  141. at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1270)
  142. at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
  143. at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
  144. at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1270)
  145. at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
  146. at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
  147. at scala.Option.foreach(Option.scala:236)
  148. at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:697)
  149. at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1496)
  150. at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1458)
  151. at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1447)
  152. at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
  153. at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:567)
  154. at org.apache.spark.SparkContext.runJob(SparkContext.scala:1822)
  155. at org.apache.spark.SparkContext.runJob(SparkContext.scala:1835)
  156. at org.apache.spark.SparkContext.runJob(SparkContext.scala:1848)
  157. at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:215)
  158. at org.apache.spark.sql.execution.Limit.executeCollect(basicOperators.scala:207)
  159. at org.apache.spark.sql.DataFrame$$anonfun$collect$1.apply(DataFrame.scala:1385)
  160. at org.apache.spark.sql.DataFrame$$anonfun$collect$1.apply(DataFrame.scala:1385)
  161. at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:56)
  162. at org.apache.spark.sql.DataFrame.withNewExecutionId(DataFrame.scala:1903)
  163. at org.apache.spark.sql.DataFrame.collect(DataFrame.scala:1384)
  164. at org.apache.spark.sql.DataFrame.head(DataFrame.scala:1314)
  165. at org.apache.spark.sql.DataFrame.take(DataFrame.scala:1377)
  166. at org.apache.spark.sql.DataFrame.showString(DataFrame.scala:178)
  167. at org.apache.spark.sql.DataFrame.show(DataFrame.scala:401)
  168. at org.apache.spark.sql.DataFrame.show(DataFrame.scala:362)
  169. at org.apache.spark.sql.DataFrame.show(DataFrame.scala:370)
  170. at main.scala.expression.Main$.main(Main.scala:45)
  171. at main.scala.expression.Main.main(Main.scala)
  172. Caused by: java.lang.NullPointerException
  173. at org.apache.hadoop.hive.ql.udf.generic.GenericUDFLeadLag.evaluate(GenericUDFLeadLag.java:57)
  174. at org.apache.spark.sql.hive.HiveGenericUDF.eval(hiveUDFs.scala:221)
  175. at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificMutableProjection.apply(Unknown Source)
  176. at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:373)
  177. at org.apache.spark.sql.execution.aggregate.AggregationIterator$$anonfun$32.apply(AggregationIterator.scala:362)
  178. at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:141)
  179. at org.apache.spark.sql.execution.aggregate.SortBasedAggregationIterator.next(SortBasedAggregationIterator.scala:30)
  180. at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
  181. at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
  182. at scala.collection.Iterator$class.foreach(Iterator.scala:727)
  183. at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
  184. at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
  185. at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
  186. at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
  187. at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
  188. at scala.collection.AbstractIterator.to(Iterator.scala:1157)
  189. at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
  190. at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
  191. at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
  192. at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
  193. at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
  194. at org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:215)
  195. at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
  196. at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1848)
  197. at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
  198. at org.apache.spark.scheduler.Task.run(Task.scala:88)
  199. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
  200. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  201. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
  202. at java.lang.Thread.run(Thread.java:745)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement