Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- 16/09/14 17:02:42 ERROR StreamingContext: Error starting the context, marking it as stopped
- java.io.IOException: org.apache.spark.SparkException: An exception was raised by Python:
- Traceback (most recent call last):
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/streaming/util.py", line 105, in dumps
- func.func, func.rdd_wrap_func, func.deserializers)))
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/serializers.py", line 428, in dumps
- return cloudpickle.dumps(obj, 2)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 646, in dumps
- cp.dump(obj)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 107, in dump
- return Pickler.dump(self, obj)
- File "/usr/lib64/python2.7/pickle.py", line 224, in dump
- self.save(obj)
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/lib64/python2.7/pickle.py", line 548, in save_tuple
- save(element)
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 193, in save_function
- self.save_function_tuple(obj)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 236, in save_function_tuple
- save((code, closure, base_globals))
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/lib64/python2.7/pickle.py", line 548, in save_tuple
- save(element)
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/lib64/python2.7/pickle.py", line 600, in save_list
- self._batch_appends(iter(obj))
- File "/usr/lib64/python2.7/pickle.py", line 636, in _batch_appends
- save(tmp[0])
- File "/usr/lib64/python2.7/pickle.py", line 331, in save
- self.save_reduce(obj=obj, *rv)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 542, in save_reduce
- save(state)
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/lib64/python2.7/pickle.py", line 649, in save_dict
- self._batch_setitems(obj.iteritems())
- File "/usr/lib64/python2.7/pickle.py", line 681, in _batch_setitems
- save(v)
- File "/usr/lib64/python2.7/pickle.py", line 306, in save
- rv = reduce(self.proto)
- File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py", line 813, in __call__
- answer, self.gateway_client, self.target_id, self.name)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/sql/utils.py", line 45, in deco
- return f(*a, **kw)
- File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/protocol.py", line 312, in get_return_value
- format(target_id, ".", name, value))
- Py4JError: An error occurred while calling o51.__getnewargs__. Trace:
- py4j.Py4JException: Method __getnewargs__([]) does not exist
- at py4j.reflection.ReflectionEngine.getMethod(ReflectionEngine.java:335)
- at py4j.reflection.ReflectionEngine.getMethod(ReflectionEngine.java:344)
- at py4j.Gateway.invoke(Gateway.java:252)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:209)
- at java.lang.Thread.run(Thread.java:745)
- at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1207)
- at org.apache.spark.streaming.api.python.TransformFunction.writeObject(PythonDStream.scala:100)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:497)
- at java.io.ObjectStreamClass.invokeWriteObject(ObjectStreamClass.java:988)
- at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1496)
- at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
- at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
- at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
- at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
- at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
- at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
- at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
- at java.io.ObjectOutputStream.writeArray(ObjectOutputStream.java:1378)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1174)
- at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
- at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
- at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
- at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
- at java.io.ObjectOutputStream.defaultWriteObject(ObjectOutputStream.java:441)
- at org.apache.spark.streaming.DStreamGraph$$anonfun$writeObject$1.apply$mcV$sp(DStreamGraph.scala:180)
- at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1204)
- at org.apache.spark.streaming.DStreamGraph.writeObject(DStreamGraph.scala:175)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:497)
- at java.io.ObjectStreamClass.invokeWriteObject(ObjectStreamClass.java:988)
- at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1496)
- at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
- at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
- at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
- at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
- at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:348)
- at org.apache.spark.streaming.Checkpoint$$anonfun$serialize$1.apply$mcV$sp(Checkpoint.scala:141)
- at org.apache.spark.streaming.Checkpoint$$anonfun$serialize$1.apply(Checkpoint.scala:141)
- at org.apache.spark.streaming.Checkpoint$$anonfun$serialize$1.apply(Checkpoint.scala:141)
- at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1250)
- at org.apache.spark.streaming.Checkpoint$.serialize(Checkpoint.scala:142)
- at org.apache.spark.streaming.StreamingContext.validate(StreamingContext.scala:554)
- at org.apache.spark.streaming.StreamingContext.liftedTree1$1(StreamingContext.scala:601)
- at org.apache.spark.streaming.StreamingContext.start(StreamingContext.scala:600)
- at org.apache.spark.streaming.api.java.JavaStreamingContext.start(JavaStreamingContext.scala:624)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:497)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:381)
- at py4j.Gateway.invoke(Gateway.java:259)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:209)
- at java.lang.Thread.run(Thread.java:745)
- Caused by: org.apache.spark.SparkException: An exception was raised by Python:
- Traceback (most recent call last):
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/streaming/util.py", line 105, in dumps
- func.func, func.rdd_wrap_func, func.deserializers)))
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/serializers.py", line 428, in dumps
- return cloudpickle.dumps(obj, 2)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 646, in dumps
- cp.dump(obj)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 107, in dump
- return Pickler.dump(self, obj)
- File "/usr/lib64/python2.7/pickle.py", line 224, in dump
- self.save(obj)
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/lib64/python2.7/pickle.py", line 548, in save_tuple
- save(element)
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 193, in save_function
- self.save_function_tuple(obj)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 236, in save_function_tuple
- save((code, closure, base_globals))
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/lib64/python2.7/pickle.py", line 548, in save_tuple
- save(element)
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/lib64/python2.7/pickle.py", line 600, in save_list
- self._batch_appends(iter(obj))
- File "/usr/lib64/python2.7/pickle.py", line 636, in _batch_appends
- save(tmp[0])
- File "/usr/lib64/python2.7/pickle.py", line 331, in save
- self.save_reduce(obj=obj, *rv)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 542, in save_reduce
- save(state)
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/lib64/python2.7/pickle.py", line 649, in save_dict
- self._batch_setitems(obj.iteritems())
- File "/usr/lib64/python2.7/pickle.py", line 681, in _batch_setitems
- save(v)
- File "/usr/lib64/python2.7/pickle.py", line 306, in save
- rv = reduce(self.proto)
- File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py", line 813, in __call__
- answer, self.gateway_client, self.target_id, self.name)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/sql/utils.py", line 45, in deco
- return f(*a, **kw)
- File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/protocol.py", line 312, in get_return_value
- format(target_id, ".", name, value))
- Py4JError: An error occurred while calling o51.__getnewargs__. Trace:
- py4j.Py4JException: Method __getnewargs__([]) does not exist
- at py4j.reflection.ReflectionEngine.getMethod(ReflectionEngine.java:335)
- at py4j.reflection.ReflectionEngine.getMethod(ReflectionEngine.java:344)
- at py4j.Gateway.invoke(Gateway.java:252)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:209)
- at java.lang.Thread.run(Thread.java:745)
- at org.apache.spark.streaming.api.python.PythonTransformFunctionSerializer$.serialize(PythonDStream.scala:144)
- at org.apache.spark.streaming.api.python.TransformFunction$$anonfun$writeObject$1.apply$mcV$sp(PythonDStream.scala:101)
- at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1204)
- ... 61 more
- Traceback (most recent call last):
- File "/usr/hdp/2.4.0.0-169/spark/python/SparkCheckpoint.py", line 78, in <module>
- ssc.start()
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/streaming/context.py", line 199, in start
- self._jssc.start()
- File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py", line 813, in __call__
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/sql/utils.py", line 45, in deco
- return f(*a, **kw)
- File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/protocol.py", line 308, in get_return_value
- py4j.protocol.Py4JJavaError: An error occurred while calling o47.start.
- : java.io.IOException: org.apache.spark.SparkException: An exception was raised by Python:
- Traceback (most recent call last):
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/streaming/util.py", line 105, in dumps
- func.func, func.rdd_wrap_func, func.deserializers)))
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/serializers.py", line 428, in dumps
- return cloudpickle.dumps(obj, 2)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 646, in dumps
- cp.dump(obj)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 107, in dump
- return Pickler.dump(self, obj)
- File "/usr/lib64/python2.7/pickle.py", line 224, in dump
- self.save(obj)
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/lib64/python2.7/pickle.py", line 548, in save_tuple
- save(element)
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 193, in save_function
- self.save_function_tuple(obj)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 236, in save_function_tuple
- save((code, closure, base_globals))
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/lib64/python2.7/pickle.py", line 548, in save_tuple
- save(element)
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/lib64/python2.7/pickle.py", line 600, in save_list
- self._batch_appends(iter(obj))
- File "/usr/lib64/python2.7/pickle.py", line 636, in _batch_appends
- save(tmp[0])
- File "/usr/lib64/python2.7/pickle.py", line 331, in save
- self.save_reduce(obj=obj, *rv)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/cloudpickle.py", line 542, in save_reduce
- save(state)
- File "/usr/lib64/python2.7/pickle.py", line 286, in save
- f(self, obj) # Call unbound method with explicit self
- File "/usr/lib64/python2.7/pickle.py", line 649, in save_dict
- self._batch_setitems(obj.iteritems())
- File "/usr/lib64/python2.7/pickle.py", line 681, in _batch_setitems
- save(v)
- File "/usr/lib64/python2.7/pickle.py", line 306, in save
- rv = reduce(self.proto)
- File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py", line 813, in __call__
- answer, self.gateway_client, self.target_id, self.name)
- File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/sql/utils.py", line 45, in deco
- return f(*a, **kw)
- File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/protocol.py", line 312, in get_return_value
- format(target_id, ".", name, value))
- Py4JError: An error occurred while calling o51.__getnewargs__. Trace:
- py4j.Py4JException: Method __getnewargs__([]) does not exist
- at py4j.reflection.ReflectionEngine.getMethod(ReflectionEngine.java:335)
- at py4j.reflection.ReflectionEngine.getMethod(ReflectionEngine.java:344)
- at py4j.Gateway.invoke(Gateway.java:252)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:209)
- at java.lang.Thread.run(Thread.java:745)
- at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1207)
- at org.apache.spark.streaming.api.python.TransformFunction.writeObject(PythonDStream.scala:100)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:497)
- at java.io.ObjectStreamClass.invokeWriteObject(ObjectStreamClass.java:988)
- at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1496)
- at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
- at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
- at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
- at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
- at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
- at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
- at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
- at java.io.ObjectOutputStream.writeArray(ObjectOutputStream.java:1378)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1174)
- at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
- at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
- at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
- at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
- at java.io.ObjectOutputStream.defaultWriteObject(ObjectOutputStream.java:441)
- at org.apache.spark.streaming.DStreamGraph$$anonfun$writeObject$1.apply$mcV$sp(DStreamGraph.scala:180)
- at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1204)
- at org.apache.spark.streaming.DStreamGraph.writeObject(DStreamGraph.scala:175)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:497)
- at java.io.ObjectStreamClass.invokeWriteObject(ObjectStreamClass.java:988)
- at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1496)
- at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
- at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
- at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
- at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
- at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:348)
- at org.apache.spark.streaming.Checkpoint$$anonfun$serialize$1.apply$mcV$sp(Checkpoint.scala:141)
- at org.apache.spark.streaming.Checkpoint$$anonfun$serialize$1.apply(Checkpoint.scala:141)
- at org.apache.spark.streaming.Checkpoint$$anonfun$serialize$1.apply(Checkpoint.scala:141)
- at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1250)
- at org.apache.spark.streaming.Checkpoint$.serialize(Checkpoint.scala:142)
- at org.apache.spark.streaming.StreamingContext.validate(StreamingContext.scala:554)
- at org.apache.spark.streaming.StreamingContext.liftedTree1$1(StreamingContext.scala:601)
- at org.apache.spark.streaming.StreamingContext.start(StreamingContext.scala:600)
- at org.apache.spark.streaming.api.java.JavaStreamingContext.start(JavaStreamingContext.scala:624)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:497)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:381)
- at py4j.Gateway.invoke(Gateway.java:259)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:209)
- at java.lang.Thread.run(Thread.java:745)
Advertisement
Add Comment
Please, Sign In to add comment