Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- 18/03/11 18:45:23 INFO org.spark_project.jetty.util.log: Logging initialized @2766ms
- 18/03/11 18:45:23 INFO org.spark_project.jetty.server.Server: jetty-9.3.z-SNAPSHOT
- 18/03/11 18:45:23 INFO org.spark_project.jetty.server.Server: Started @2857ms
- 18/03/11 18:45:23 INFO org.spark_project.jetty.server.AbstractConnector: Started ServerConnector@46b7c1a0{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
- 18/03/11 18:45:24 INFO com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystemBase: GHFS version: 1.6.3-hadoop2
- 18/03/11 18:45:25 INFO org.apache.hadoop.yarn.client.RMProxy: Connecting to ResourceManager at cluster-main-m/<ip>:8032
- 18/03/11 18:45:27 WARN org.apache.hadoop.hdfs.DataStreamer: Caught exception
- java.lang.InterruptedException
- at java.lang.Object.wait(Native Method)
- at java.lang.Thread.join(Thread.java:1252)
- at java.lang.Thread.join(Thread.java:1326)
- at org.apache.hadoop.hdfs.DataStreamer.closeResponder(DataStreamer.java:973)
- at org.apache.hadoop.hdfs.DataStreamer.endBlock(DataStreamer.java:624)
- at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:801)
- 18/03/11 18:45:27 INFO org.apache.hadoop.yarn.client.api.impl.YarnClientImpl: Submitted application application_1519879216511_0014
- Traceback (most recent call last):
- File "/tmp/job-eb92287a/streaming.py", line 17, in <module>
- .option("startingOffsets", "earliest") \
- File "/usr/lib/spark/python/lib/pyspark.zip/pyspark/sql/streaming.py", line 397, in load
- File "/usr/lib/spark/python/lib/py4j-0.10.4-src.zip/py4j/java_gateway.py", line 1133, in __call__
- File "/usr/lib/spark/python/lib/pyspark.zip/pyspark/sql/utils.py", line 63, in deco
- File "/usr/lib/spark/python/lib/py4j-0.10.4-src.zip/py4j/protocol.py", line 319, in get_return_value
- py4j.protocol.Py4JJavaError: An error occurred while calling o58.load.
- : java.lang.ClassNotFoundException: Failed to find data source: kafka. Please find packages at http://spark.apache.org/third-party-projects.html
- at org.apache.spark.sql.execution.datasources.DataSource$.lookupDataSource(DataSource.scala:549)
- at org.apache.spark.sql.execution.datasources.DataSource.providingClass$lzycompute(DataSource.scala:86)
- at org.apache.spark.sql.execution.datasources.DataSource.providingClass(DataSource.scala:86)
- at org.apache.spark.sql.execution.datasources.DataSource.sourceSchema(DataSource.scala:195)
- at org.apache.spark.sql.execution.datasources.DataSource.sourceInfo$lzycompute(DataSource.scala:87)
- at org.apache.spark.sql.execution.datasources.DataSource.sourceInfo(DataSource.scala:87)
- at org.apache.spark.sql.execution.streaming.StreamingRelation$.apply(StreamingRelation.scala:30)
- at org.apache.spark.sql.streaming.DataStreamReader.load(DataStreamReader.scala:150)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
- at py4j.Gateway.invoke(Gateway.java:280)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:214)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.lang.ClassNotFoundException: kafka.DefaultSource
- at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
- at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
- at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
- at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$21$$anonfun$apply$12.apply(DataSource.scala:533)
- at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$21$$anonfun$apply$12.apply(DataSource.scala:533)
- at scala.util.Try$.apply(Try.scala:192)
- at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$21.apply(DataSource.scala:533)
- at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$21.apply(DataSource.scala:533)
- at scala.util.Try.orElse(Try.scala:84)
- at org.apache.spark.sql.execution.datasources.DataSource$.lookupDataSource(DataSource.scala:533)
- ... 18 more
- 18/03/11 18:45:36 INFO org.spark_project.jetty.server.AbstractConnector: Stopped Spark@46b7c1a0{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
- 18/03/11 18:45:36 WARN org.apache.hadoop.ipc.Client: interrupted waiting to send rpc request to server
- java.lang.InterruptedException
- at java.util.concurrent.FutureTask.awaitDone(FutureTask.java:404)
- at java.util.concurrent.FutureTask.get(FutureTask.java:191)
- at org.apache.hadoop.ipc.Client$Connection.sendRpcRequest(Client.java:1135)
- at org.apache.hadoop.ipc.Client.call(Client.java:1384)
- at org.apache.hadoop.ipc.Client.call(Client.java:1342)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:227)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
- at com.sun.proxy.$Proxy15.getApplicationReport(Unknown Source)
- at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getApplicationReport(ApplicationClientProtocolPBClientImpl.java:228)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:409)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:163)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:155)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:346)
- at com.sun.proxy.$Proxy16.getApplicationReport(Unknown Source)
- at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getApplicationReport(YarnClientImpl.java:480)
- at org.apache.spark.deploy.yarn.Client.getApplicationReport(Client.scala:284)
- at org.apache.spark.deploy.yarn.Client.monitorApplication(Client.scala:998)
- at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:105)
- 18/03/11 18:45:36 ERROR org.apache.spark.deploy.yarn.Client: Failed to contact YARN for application application_1519879216511_0014.
- java.io.IOException: java.lang.InterruptedException
- at org.apache.hadoop.ipc.Client.call(Client.java:1390)
- at org.apache.hadoop.ipc.Client.call(Client.java:1342)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:227)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
- at com.sun.proxy.$Proxy15.getApplicationReport(Unknown Source)
- at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getApplicationReport(ApplicationClientProtocolPBClientImpl.java:228)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:409)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:163)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:155)
- at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:346)
- at com.sun.proxy.$Proxy16.getApplicationReport(Unknown Source)
- at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getApplicationReport(YarnClientImpl.java:480)
- at org.apache.spark.deploy.yarn.Client.getApplicationReport(Client.scala:284)
- at org.apache.spark.deploy.yarn.Client.monitorApplication(Client.scala:998)
- at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:105)
- Caused by: java.lang.InterruptedException
- at java.util.concurrent.FutureTask.awaitDone(FutureTask.java:404)
- at java.util.concurrent.FutureTask.get(FutureTask.java:191)
- at org.apache.hadoop.ipc.Client$Connection.sendRpcRequest(Client.java:1135)
- at org.apache.hadoop.ipc.Client.call(Client.java:1384)
- ... 19 more
- 18/03/11 18:45:36 ERROR org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend: Yarn application has already exited with state FAILED!
- Job output is complete
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement