Guest User

Error

a guest
Sep 14th, 2016
236
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 20.39 KB | None | 0 0
  1. 16/09/14 17:02:42 ERROR StreamingContext: Error starting the context, marking it as stopped
  2. java.io.IOException: org.apache.spark.SparkException: An exception was raised by Python:
  3. Traceback (most recent call last):
  4. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/streaming/util.py", line 105, in dumps
  5. func.func, func.rdd_wrap_func, func.deserializers)))
  6. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/serializers.py", line 428, in dumps
  7. return cloudpickle.dumps(obj, 2)
  8. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 646, in dumps
  9. cp.dump(obj)
  10. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 107, in dump
  11. return Pickler.dump(self, obj)
  12. File "/usr/lib64/python2.7/pickle.py", line 224, in dump
  13. self.save(obj)
  14. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  15. f(self, obj) # Call unbound method with explicit self
  16. File "/usr/lib64/python2.7/pickle.py", line 548, in save_tuple
  17. save(element)
  18. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  19. f(self, obj) # Call unbound method with explicit self
  20. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 193, in save_function
  21. self.save_function_tuple(obj)
  22. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 236, in save_function_tuple
  23. save((code, closure, base_globals))
  24. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  25. f(self, obj) # Call unbound method with explicit self
  26. File "/usr/lib64/python2.7/pickle.py", line 548, in save_tuple
  27. save(element)
  28. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  29. f(self, obj) # Call unbound method with explicit self
  30. File "/usr/lib64/python2.7/pickle.py", line 600, in save_list
  31. self._batch_appends(iter(obj))
  32. File "/usr/lib64/python2.7/pickle.py", line 636, in _batch_appends
  33. save(tmp[0])
  34. File "/usr/lib64/python2.7/pickle.py", line 331, in save
  35. self.save_reduce(obj=obj, *rv)
  36. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 542, in save_reduce
  37. save(state)
  38. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  39. f(self, obj) # Call unbound method with explicit self
  40. File "/usr/lib64/python2.7/pickle.py", line 649, in save_dict
  41. self._batch_setitems(obj.iteritems())
  42. File "/usr/lib64/python2.7/pickle.py", line 681, in _batch_setitems
  43. save(v)
  44. File "/usr/lib64/python2.7/pickle.py", line 306, in save
  45. rv = reduce(self.proto)
  46. File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py", line 813, in __call__
  47. answer, self.gateway_client, self.target_id, self.name)
  48. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/sql/utils.py", line 45, in deco
  49. return f(*a, **kw)
  50. File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/protocol.py", line 312, in get_return_value
  51. format(target_id, ".", name, value))
  52. Py4JError: An error occurred while calling o51.__getnewargs__. Trace:
  53. py4j.Py4JException: Method __getnewargs__([]) does not exist
  54. at py4j.reflection.ReflectionEngine.getMethod(ReflectionEngine.java:335)
  55. at py4j.reflection.ReflectionEngine.getMethod(ReflectionEngine.java:344)
  56. at py4j.Gateway.invoke(Gateway.java:252)
  57. at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
  58. at py4j.commands.CallCommand.execute(CallCommand.java:79)
  59. at py4j.GatewayConnection.run(GatewayConnection.java:209)
  60. at java.lang.Thread.run(Thread.java:745)
  61.  
  62.  
  63.  
  64. at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1207)
  65. at org.apache.spark.streaming.api.python.TransformFunction.writeObject(PythonDStream.scala:100)
  66. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  67. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  68. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  69. at java.lang.reflect.Method.invoke(Method.java:497)
  70. at java.io.ObjectStreamClass.invokeWriteObject(ObjectStreamClass.java:988)
  71. at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1496)
  72. at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
  73. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
  74. at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
  75. at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
  76. at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
  77. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
  78. at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
  79. at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
  80. at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
  81. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
  82. at java.io.ObjectOutputStream.writeArray(ObjectOutputStream.java:1378)
  83. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1174)
  84. at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
  85. at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
  86. at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
  87. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
  88. at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
  89. at java.io.ObjectOutputStream.defaultWriteObject(ObjectOutputStream.java:441)
  90. at org.apache.spark.streaming.DStreamGraph$$anonfun$writeObject$1.apply$mcV$sp(DStreamGraph.scala:180)
  91. at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1204)
  92. at org.apache.spark.streaming.DStreamGraph.writeObject(DStreamGraph.scala:175)
  93. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  94. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  95. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  96. at java.lang.reflect.Method.invoke(Method.java:497)
  97. at java.io.ObjectStreamClass.invokeWriteObject(ObjectStreamClass.java:988)
  98. at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1496)
  99. at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
  100. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
  101. at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
  102. at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
  103. at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
  104. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
  105. at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:348)
  106. at org.apache.spark.streaming.Checkpoint$$anonfun$serialize$1.apply$mcV$sp(Checkpoint.scala:141)
  107. at org.apache.spark.streaming.Checkpoint$$anonfun$serialize$1.apply(Checkpoint.scala:141)
  108. at org.apache.spark.streaming.Checkpoint$$anonfun$serialize$1.apply(Checkpoint.scala:141)
  109. at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1250)
  110. at org.apache.spark.streaming.Checkpoint$.serialize(Checkpoint.scala:142)
  111. at org.apache.spark.streaming.StreamingContext.validate(StreamingContext.scala:554)
  112. at org.apache.spark.streaming.StreamingContext.liftedTree1$1(StreamingContext.scala:601)
  113. at org.apache.spark.streaming.StreamingContext.start(StreamingContext.scala:600)
  114. at org.apache.spark.streaming.api.java.JavaStreamingContext.start(JavaStreamingContext.scala:624)
  115. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  116. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  117. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  118. at java.lang.reflect.Method.invoke(Method.java:497)
  119. at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231)
  120. at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:381)
  121. at py4j.Gateway.invoke(Gateway.java:259)
  122. at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
  123. at py4j.commands.CallCommand.execute(CallCommand.java:79)
  124. at py4j.GatewayConnection.run(GatewayConnection.java:209)
  125. at java.lang.Thread.run(Thread.java:745)
  126. Caused by: org.apache.spark.SparkException: An exception was raised by Python:
  127. Traceback (most recent call last):
  128. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/streaming/util.py", line 105, in dumps
  129. func.func, func.rdd_wrap_func, func.deserializers)))
  130. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/serializers.py", line 428, in dumps
  131. return cloudpickle.dumps(obj, 2)
  132. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 646, in dumps
  133. cp.dump(obj)
  134. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 107, in dump
  135. return Pickler.dump(self, obj)
  136. File "/usr/lib64/python2.7/pickle.py", line 224, in dump
  137. self.save(obj)
  138. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  139. f(self, obj) # Call unbound method with explicit self
  140. File "/usr/lib64/python2.7/pickle.py", line 548, in save_tuple
  141. save(element)
  142. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  143. f(self, obj) # Call unbound method with explicit self
  144. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 193, in save_function
  145. self.save_function_tuple(obj)
  146. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 236, in save_function_tuple
  147. save((code, closure, base_globals))
  148. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  149. f(self, obj) # Call unbound method with explicit self
  150. File "/usr/lib64/python2.7/pickle.py", line 548, in save_tuple
  151. save(element)
  152. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  153. f(self, obj) # Call unbound method with explicit self
  154. File "/usr/lib64/python2.7/pickle.py", line 600, in save_list
  155. self._batch_appends(iter(obj))
  156. File "/usr/lib64/python2.7/pickle.py", line 636, in _batch_appends
  157. save(tmp[0])
  158. File "/usr/lib64/python2.7/pickle.py", line 331, in save
  159. self.save_reduce(obj=obj, *rv)
  160. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 542, in save_reduce
  161. save(state)
  162. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  163. f(self, obj) # Call unbound method with explicit self
  164. File "/usr/lib64/python2.7/pickle.py", line 649, in save_dict
  165. self._batch_setitems(obj.iteritems())
  166. File "/usr/lib64/python2.7/pickle.py", line 681, in _batch_setitems
  167. save(v)
  168. File "/usr/lib64/python2.7/pickle.py", line 306, in save
  169. rv = reduce(self.proto)
  170. File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py", line 813, in __call__
  171. answer, self.gateway_client, self.target_id, self.name)
  172. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/sql/utils.py", line 45, in deco
  173. return f(*a, **kw)
  174. File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/protocol.py", line 312, in get_return_value
  175. format(target_id, ".", name, value))
  176. Py4JError: An error occurred while calling o51.__getnewargs__. Trace:
  177. py4j.Py4JException: Method __getnewargs__([]) does not exist
  178. at py4j.reflection.ReflectionEngine.getMethod(ReflectionEngine.java:335)
  179. at py4j.reflection.ReflectionEngine.getMethod(ReflectionEngine.java:344)
  180. at py4j.Gateway.invoke(Gateway.java:252)
  181. at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
  182. at py4j.commands.CallCommand.execute(CallCommand.java:79)
  183. at py4j.GatewayConnection.run(GatewayConnection.java:209)
  184. at java.lang.Thread.run(Thread.java:745)
  185.  
  186.  
  187.  
  188. at org.apache.spark.streaming.api.python.PythonTransformFunctionSerializer$.serialize(PythonDStream.scala:144)
  189. at org.apache.spark.streaming.api.python.TransformFunction$$anonfun$writeObject$1.apply$mcV$sp(PythonDStream.scala:101)
  190. at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1204)
  191. ... 61 more
  192. Traceback (most recent call last):
  193. File "/usr/hdp/2.4.0.0-169/spark/python/SparkCheckpoint.py", line 78, in <module>
  194. ssc.start()
  195. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/streaming/context.py", line 199, in start
  196. self._jssc.start()
  197. File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py", line 813, in __call__
  198. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/sql/utils.py", line 45, in deco
  199. return f(*a, **kw)
  200. File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/protocol.py", line 308, in get_return_value
  201. py4j.protocol.Py4JJavaError: An error occurred while calling o47.start.
  202. : java.io.IOException: org.apache.spark.SparkException: An exception was raised by Python:
  203. Traceback (most recent call last):
  204. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/streaming/util.py", line 105, in dumps
  205. func.func, func.rdd_wrap_func, func.deserializers)))
  206. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/serializers.py", line 428, in dumps
  207. return cloudpickle.dumps(obj, 2)
  208. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 646, in dumps
  209. cp.dump(obj)
  210. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 107, in dump
  211. return Pickler.dump(self, obj)
  212. File "/usr/lib64/python2.7/pickle.py", line 224, in dump
  213. self.save(obj)
  214. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  215. f(self, obj) # Call unbound method with explicit self
  216. File "/usr/lib64/python2.7/pickle.py", line 548, in save_tuple
  217. save(element)
  218. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  219. f(self, obj) # Call unbound method with explicit self
  220. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 193, in save_function
  221. self.save_function_tuple(obj)
  222. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 236, in save_function_tuple
  223. save((code, closure, base_globals))
  224. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  225. f(self, obj) # Call unbound method with explicit self
  226. File "/usr/lib64/python2.7/pickle.py", line 548, in save_tuple
  227. save(element)
  228. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  229. f(self, obj) # Call unbound method with explicit self
  230. File "/usr/lib64/python2.7/pickle.py", line 600, in save_list
  231. self._batch_appends(iter(obj))
  232. File "/usr/lib64/python2.7/pickle.py", line 636, in _batch_appends
  233. save(tmp[0])
  234. File "/usr/lib64/python2.7/pickle.py", line 331, in save
  235. self.save_reduce(obj=obj, *rv)
  236. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 542, in save_reduce
  237. save(state)
  238. File "/usr/lib64/python2.7/pickle.py", line 286, in save
  239. f(self, obj) # Call unbound method with explicit self
  240. File "/usr/lib64/python2.7/pickle.py", line 649, in save_dict
  241. self._batch_setitems(obj.iteritems())
  242. File "/usr/lib64/python2.7/pickle.py", line 681, in _batch_setitems
  243. save(v)
  244. File "/usr/lib64/python2.7/pickle.py", line 306, in save
  245. rv = reduce(self.proto)
  246. File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py", line 813, in __call__
  247. answer, self.gateway_client, self.target_id, self.name)
  248. File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/sql/utils.py", line 45, in deco
  249. return f(*a, **kw)
  250. File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/protocol.py", line 312, in get_return_value
  251. format(target_id, ".", name, value))
  252. Py4JError: An error occurred while calling o51.__getnewargs__. Trace:
  253. py4j.Py4JException: Method __getnewargs__([]) does not exist
  254. at py4j.reflection.ReflectionEngine.getMethod(ReflectionEngine.java:335)
  255. at py4j.reflection.ReflectionEngine.getMethod(ReflectionEngine.java:344)
  256. at py4j.Gateway.invoke(Gateway.java:252)
  257. at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
  258. at py4j.commands.CallCommand.execute(CallCommand.java:79)
  259. at py4j.GatewayConnection.run(GatewayConnection.java:209)
  260. at java.lang.Thread.run(Thread.java:745)
  261.  
  262.  
  263.  
  264. at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1207)
  265. at org.apache.spark.streaming.api.python.TransformFunction.writeObject(PythonDStream.scala:100)
  266. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  267. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  268. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  269. at java.lang.reflect.Method.invoke(Method.java:497)
  270. at java.io.ObjectStreamClass.invokeWriteObject(ObjectStreamClass.java:988)
  271. at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1496)
  272. at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
  273. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
  274. at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
  275. at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
  276. at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
  277. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
  278. at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
  279. at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
  280. at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
  281. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
  282. at java.io.ObjectOutputStream.writeArray(ObjectOutputStream.java:1378)
  283. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1174)
  284. at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
  285. at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
  286. at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
  287. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
  288. at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
  289. at java.io.ObjectOutputStream.defaultWriteObject(ObjectOutputStream.java:441)
  290. at org.apache.spark.streaming.DStreamGraph$$anonfun$writeObject$1.apply$mcV$sp(DStreamGraph.scala:180)
  291. at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1204)
  292. at org.apache.spark.streaming.DStreamGraph.writeObject(DStreamGraph.scala:175)
  293. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  294. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  295. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  296. at java.lang.reflect.Method.invoke(Method.java:497)
  297. at java.io.ObjectStreamClass.invokeWriteObject(ObjectStreamClass.java:988)
  298. at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1496)
  299. at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
  300. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
  301. at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
  302. at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
  303. at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
  304. at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
  305. at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:348)
  306. at org.apache.spark.streaming.Checkpoint$$anonfun$serialize$1.apply$mcV$sp(Checkpoint.scala:141)
  307. at org.apache.spark.streaming.Checkpoint$$anonfun$serialize$1.apply(Checkpoint.scala:141)
  308. at org.apache.spark.streaming.Checkpoint$$anonfun$serialize$1.apply(Checkpoint.scala:141)
  309. at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1250)
  310. at org.apache.spark.streaming.Checkpoint$.serialize(Checkpoint.scala:142)
  311. at org.apache.spark.streaming.StreamingContext.validate(StreamingContext.scala:554)
  312. at org.apache.spark.streaming.StreamingContext.liftedTree1$1(StreamingContext.scala:601)
  313. at org.apache.spark.streaming.StreamingContext.start(StreamingContext.scala:600)
  314. at org.apache.spark.streaming.api.java.JavaStreamingContext.start(JavaStreamingContext.scala:624)
  315. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  316. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  317. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  318. at java.lang.reflect.Method.invoke(Method.java:497)
  319. at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231)
  320. at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:381)
  321. at py4j.Gateway.invoke(Gateway.java:259)
  322. at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
  323. at py4j.commands.CallCommand.execute(CallCommand.java:79)
  324. at py4j.GatewayConnection.run(GatewayConnection.java:209)
  325. at java.lang.Thread.run(Thread.java:745)
Advertisement
Add Comment
Please, Sign In to add comment