Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- vagrant@kmaster:~/spark-2.4.5-bin-hadoop2.7$ kubectl logs spark-pi-1587656135431-driver
- ++ id -u
- + myuid=0
- ++ id -g
- + mygid=0
- + set +e
- ++ getent passwd 0
- + uidentry=root:x:0:0:root:/root:/bin/bash
- + set -e
- + '[' -z root:x:0:0:root:/root:/bin/bash ']'
- + SPARK_K8S_CMD=driver
- + case "$SPARK_K8S_CMD" in
- + shift 1
- + SPARK_CLASSPATH=':/opt/spark/jars/*'
- + env
- + sed 's/[^=]*=\(.*\)/\1/g'
- + sort -t_ -k4 -n
- + grep SPARK_JAVA_OPT_
- + readarray -t SPARK_EXECUTOR_JAVA_OPTS
- + '[' -n '' ']'
- + '[' -n '' ']'
- + PYSPARK_ARGS=
- + '[' -n '' ']'
- + R_ARGS=
- + '[' -n '' ']'
- + '[' '' == 2 ']'
- + '[' '' == 3 ']'
- + case "$SPARK_K8S_CMD" in
- + CMD=("$SPARK_HOME/bin/spark-submit" --conf "spark.driver.bindAddress=$SPARK_DRIVER_BIND_ADDRESS" --deploy-mode client "$@")
- + exec /usr/bin/tini -s -- /opt/spark/bin/spark-submit --conf spark.driver.bindAddress=192.168.41.156 --deploy-mode client --properties-file /opt/spark/conf/spark.properties --class org.apache.spark.examples.SparkPi spark-internal 10000000
- 20/04/23 15:35:59 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
- Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
- 20/04/23 15:36:00 INFO SparkContext: Running Spark version 2.4.5
- 20/04/23 15:36:00 INFO SparkContext: Submitted application: Spark Pi
- 20/04/23 15:36:00 INFO SecurityManager: Changing view acls to: root
- 20/04/23 15:36:00 INFO SecurityManager: Changing modify acls to: root
- 20/04/23 15:36:00 INFO SecurityManager: Changing view acls groups to:
- 20/04/23 15:36:00 INFO SecurityManager: Changing modify acls groups to:
- 20/04/23 15:36:00 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); groups with view permissions: Set(); users with modify permissions: Set(root); groups with modify permissions: Set()
- 20/04/23 15:36:00 INFO Utils: Successfully started service 'sparkDriver' on port 7078.
- 20/04/23 15:36:00 INFO SparkEnv: Registering MapOutputTracker
- 20/04/23 15:36:00 INFO SparkEnv: Registering BlockManagerMaster
- 20/04/23 15:36:00 INFO BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
- 20/04/23 15:36:00 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
- 20/04/23 15:36:00 INFO DiskBlockManager: Created local directory at /var/data/spark-6ef68f6e-878f-4924-bb80-4d7d8561989c/blockmgr-852d1b3b-dd24-4108-a6da-42446e8aef49
- 20/04/23 15:36:00 INFO MemoryStore: MemoryStore started with capacity 413.9 MB
- 20/04/23 15:36:00 INFO SparkEnv: Registering OutputCommitCoordinator
- 20/04/23 15:36:01 INFO Utils: Successfully started service 'SparkUI' on port 4040.
- 20/04/23 15:36:01 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at http://spark-pi-1587656135431-driver-svc.default.svc:4040
- 20/04/23 15:36:01 INFO SparkContext: Added JAR file:///opt/spark/examples/jars/spark-examples_2.11-2.4.5.jar at spark://spark-pi-1587656135431-driver-svc.default.svc:7078/jars/spark-examples_2.11-2.4.5.jar with timestamp 1587656161394
- 20/04/23 15:36:02 ERROR SparkContext: Error initializing SparkContext.
- org.apache.spark.SparkException: External scheduler cannot be instantiated
- at org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:2794)
- at org.apache.spark.SparkContext.<init>(SparkContext.scala:493)
- at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2520)
- at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:935)
- at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:926)
- at scala.Option.getOrElse(Option.scala:121)
- at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:926)
- at org.apache.spark.examples.SparkPi$.main(SparkPi.scala:31)
- at org.apache.spark.examples.SparkPi.main(SparkPi.scala)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
- at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:845)
- at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:161)
- at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:184)
- at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
- at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:920)
- at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:929)
- at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
- Caused by: io.fabric8.kubernetes.client.KubernetesClientException: Operation: [get] for kind: [Pod] with name: [spark-pi-1587656135431-driver] in namespace: [default] failed.
- at io.fabric8.kubernetes.client.KubernetesClientException.launderThrowable(KubernetesClientException.java:64)
- at io.fabric8.kubernetes.client.KubernetesClientException.launderThrowable(KubernetesClientException.java:72)
- at io.fabric8.kubernetes.client.dsl.base.BaseOperation.getMandatory(BaseOperation.java:237)
- at io.fabric8.kubernetes.client.dsl.base.BaseOperation.get(BaseOperation.java:170)
- at org.apache.spark.scheduler.cluster.k8s.ExecutorPodsAllocator$$anonfun$1.apply(ExecutorPodsAllocator.scala:57)
- at org.apache.spark.scheduler.cluster.k8s.ExecutorPodsAllocator$$anonfun$1.apply(ExecutorPodsAllocator.scala:55)
- at scala.Option.map(Option.scala:146)
- at org.apache.spark.scheduler.cluster.k8s.ExecutorPodsAllocator.<init>(ExecutorPodsAllocator.scala:55)
- at org.apache.spark.scheduler.cluster.k8s.KubernetesClusterManager.createSchedulerBackend(KubernetesClusterManager.scala:89)
- at org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:2788)
- ... 20 more
- Caused by: java.net.SocketException: Broken pipe (Write failed)
- at java.net.SocketOutputStream.socketWrite0(Native Method)
- at java.net.SocketOutputStream.socketWrite(SocketOutputStream.java:111)
- at java.net.SocketOutputStream.write(SocketOutputStream.java:155)
- at sun.security.ssl.OutputRecord.writeBuffer(OutputRecord.java:431)
- at sun.security.ssl.OutputRecord.write(OutputRecord.java:417)
- at sun.security.ssl.SSLSocketImpl.writeRecordInternal(SSLSocketImpl.java:894)
- ... many more (redacted to fit body)
- 20/04/23 15:36:02 INFO SparkUI: Stopped Spark web UI at http://spark-pi-1587656135431-driver-svc.default.svc:4040
- 20/04/23 15:36:03 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
- 20/04/23 15:36:03 INFO MemoryStore: MemoryStore cleared
- 20/04/23 15:36:03 INFO BlockManager: BlockManager stopped
- 20/04/23 15:36:03 INFO BlockManagerMaster: BlockManagerMaster stopped
- 20/04/23 15:36:03 WARN MetricsSystem: Stopping a MetricsSystem that is not running
- 20/04/23 15:36:03 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
- 20/04/23 15:36:03 INFO SparkContext: Successfully stopped SparkContext
- Exception in thread "main" org.apache.spark.SparkException: External scheduler cannot be instantiated
- at org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:2794)
- at org.apache.spark.SparkContext.<init>(SparkContext.scala:493)
- at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2520)
- at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:935)
- at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:926)
- at scala.Option.getOrElse(Option.scala:121)
- ... many more (redacted to fit body)
- Caused by: io.fabric8.kubernetes.client.KubernetesClientException: Operation: [get] for kind: [Pod] with name: [spark-pi-1587656135431-driver] in namespace: [default] failed.
- at io.fabric8.kubernetes.client.KubernetesClientException.launderThrowable(KubernetesClientException.java:64)
- at io.fabric8.kubernetes.client.KubernetesClientException.launderThrowable(KubernetesClientException.java:72)
- at io.fabric8.kubernetes.client.dsl.base.BaseOperation.getMandatory(BaseOperation.java:237)
- at io.fabric8.kubernetes.client.dsl.base.BaseOperation.get(BaseOperation.java:170)
- at org.apache.spark.scheduler.cluster.k8s.ExecutorPodsAllocator$$anonfun$1.apply(ExecutorPodsAllocator.scala:57)
- at org.apache.spark.scheduler.cluster.k8s.ExecutorPodsAllocator$$anonfun$1.apply(ExecutorPodsAllocator.scala:55)
- at scala.Option.map(Option.scala:146)
- at org.apache.spark.scheduler.cluster.k8s.ExecutorPodsAllocator.<init>(ExecutorPodsAllocator.scala:55)
- at org.apache.spark.scheduler.cluster.k8s.KubernetesClusterManager.createSchedulerBackend(KubernetesClusterManager.scala:89)
- at org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:2788)
- ... 20 more
- Caused by: java.net.SocketException: Broken pipe (Write failed)
- at java.net.SocketOutputStream.socketWrite0(Native Method)
- at java.net.SocketOutputStream.socketWrite(SocketOutputStream.java:111)
- at java.net.SocketOutputStream.write(SocketOutputStream.java:155)
- at sun.security.ssl.OutputRecord.writeBuffer(OutputRecord.java:431)
- at sun.security.ssl.OutputRecord.write(OutputRecord.java:417)
- at sun.security.ssl.SSLSocketImpl.writeRecordInternal(SSLSocketImpl.java:894)
- at sun.security.ssl.SSLSocketImpl.writeRecord(SSLSocketImpl.java:865)
- at sun.security.ssl.AppOutputStream.write(AppOutputStream.java:123)
- at okio.Okio$1.write(Okio.java:79)
- ... many more (redacted to fit body)
- 20/04/23 15:36:03 INFO ShutdownHookManager: Shutdown hook called
- 20/04/23 15:36:03 INFO ShutdownHookManager: Deleting directory /var/data/spark-6ef68f6e-878f-4924-bb80-4d7d8561989c/spark-07474c84-2ebb-49a0-a71d-3d716c95839b
- 20/04/23 15:36:03 INFO ShutdownHookManager: Deleting directory /tmp/spark-901da352-f688-4010-a937-f87d169be25e
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement