Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- vagrant@kmaster:~/spark$ kubectl logs spark-pi-75235d719d31da3f-driver
- ++ id -u
- + myuid=185
- ++ id -g
- + mygid=0
- + set +e
- ++ getent passwd 185
- + uidentry=
- + set -e
- + '[' -z '' ']'
- + '[' -w /etc/passwd ']'
- + echo '185:x:185:0:anonymous uid:/opt/spark:/bin/false'
- + SPARK_CLASSPATH=':/opt/spark/jars/*'
- + env
- + grep SPARK_JAVA_OPT_
- + sort -t_ -k4 -n
- + sed 's/[^=]*=\(.*\)/\1/g'
- + readarray -t SPARK_EXECUTOR_JAVA_OPTS
- + '[' -n '' ']'
- + '[' '' == 2 ']'
- + '[' '' == 3 ']'
- + '[' -n '' ']'
- + '[' -z ']'
- + case "$1" in
- + shift 1
- + CMD=("$SPARK_HOME/bin/spark-submit" --conf "spark.driver.bindAddress=$SPARK_DRIVER_BIND_ADDRESS" --deploy-mode client "$@")
- + exec /usr/bin/tini -s -- /opt/spark/bin/spark-submit --conf spark.driver.bindAddress=192.168.41.181 --deploy-mode client --properties-file /opt/spark/conf/spark.properties --class org.apache.spark.examples.SparkPi local:///opt/spark/examples/jars/spark-examples_2.12-3.1.0-SNAPSHOT.jar 10000000
- 20/04/21 14:43:50 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
- Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
- 20/04/21 14:43:50 INFO SparkContext: Running Spark version 3.1.0-SNAPSHOT
- 20/04/21 14:43:50 INFO ResourceUtils: ==============================================================
- 20/04/21 14:43:50 INFO ResourceUtils: No custom resources configured for spark.driver.
- 20/04/21 14:43:50 INFO ResourceUtils: ==============================================================
- 20/04/21 14:43:50 INFO SparkContext: Submitted application: Spark Pi
- 20/04/21 14:43:50 INFO ResourceProfile: Default ResourceProfile created, executor resources: Map(cores -> name: cores, amount: 1, script: , vendor: , memory -> name: memory, amount: 1024, script: , vendor: ), task resources: Map(cpus -> name: cpus, amount: 1.0)
- 20/04/21 14:43:50 INFO ResourceProfile: Limiting resource is cpus at 1 tasks per executor
- 20/04/21 14:43:50 INFO ResourceProfileManager: Added ResourceProfile id: 0
- 20/04/21 14:43:50 INFO SecurityManager: Changing view acls to: 185,vagrant
- 20/04/21 14:43:50 INFO SecurityManager: Changing modify acls to: 185,vagrant
- 20/04/21 14:43:50 INFO SecurityManager: Changing view acls groups to:
- 20/04/21 14:43:50 INFO SecurityManager: Changing modify acls groups to:
- 20/04/21 14:43:50 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(185, vagrant); groups with view permissions: Set(); users with modify permissions: Set(185, vagrant); groups with modify permissions: Set()
- 20/04/21 14:43:50 INFO Utils: Successfully started service 'sparkDriver' on port 7078.
- 20/04/21 14:43:50 INFO SparkEnv: Registering MapOutputTracker
- 20/04/21 14:43:51 INFO SparkEnv: Registering BlockManagerMaster
- 20/04/21 14:43:51 INFO BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
- 20/04/21 14:43:51 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
- 20/04/21 14:43:51 INFO SparkEnv: Registering BlockManagerMasterHeartbeat
- 20/04/21 14:43:51 INFO DiskBlockManager: Created local directory at /var/data/spark-4fdc6242-3ab9-4ea4-864b-3b0ffb981e91/blockmgr-d8363b86-9490-4c0c-b4b5-93a59535bbb0
- 20/04/21 14:43:51 INFO MemoryStore: MemoryStore started with capacity 413.9 MiB
- 20/04/21 14:43:51 INFO SparkEnv: Registering OutputCommitCoordinator
- 20/04/21 14:43:51 INFO Utils: Successfully started service 'SparkUI' on port 4040.
- 20/04/21 14:43:51 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at http://spark-pi-75235d719d31da3f-driver-svc.default.svc:4040
- 20/04/21 14:43:51 INFO SparkContext: Added JAR local:///opt/spark/examples/jars/spark-examples_2.12-3.1.0-SNAPSHOT.jar at file:/opt/spark/examples/jars/spark-examples_2.12-3.1.0-SNAPSHOT.jar with timestamp 1587480231678
- 20/04/21 14:43:51 WARN SparkContext: The jar local:///opt/spark/examples/jars/spark-examples_2.12-3.1.0-SNAPSHOT.jar has been added already. Overwriting of added jars is not supported in the current version.
- 20/04/21 14:43:51 INFO SparkKubernetesClientFactory: Auto-configuring K8S client using current context from users K8S config file
- 20/04/21 14:43:52 ERROR SparkContext: Error initializing SparkContext.
- org.apache.spark.SparkException: External scheduler cannot be instantiated
- at org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:2859)
- at org.apache.spark.SparkContext.<init>(SparkContext.scala:533)
- at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2570)
- at org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$5(SparkSession.scala:937)
- at scala.Option.getOrElse(Option.scala:189)
- at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:928)
- at org.apache.spark.examples.SparkPi$.main(SparkPi.scala:30)
- at org.apache.spark.examples.SparkPi.main(SparkPi.scala)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
- at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:934)
- at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
- at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
- at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
- at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1013)
- at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1022)
- at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
- Caused by: io.fabric8.kubernetes.client.KubernetesClientException: Operation: [get] for kind: [Pod] with name: [spark-pi-75235d719d31da3f-driver] in namespace: [default] failed.
- at io.fabric8.kubernetes.client.KubernetesClientException.launderThrowable(KubernetesClientException.java:64)
- at io.fabric8.kubernetes.client.KubernetesClientException.launderThrowable(KubernetesClientException.java:72)
- at io.fabric8.kubernetes.client.dsl.base.BaseOperation.getMandatory(BaseOperation.java:225)
- at io.fabric8.kubernetes.client.dsl.base.BaseOperation.get(BaseOperation.java:168)
- at org.apache.spark.scheduler.cluster.k8s.ExecutorPodsAllocator.$anonfun$driverPod$1(ExecutorPodsAllocator.scala:59)
- at scala.Option.map(Option.scala:230)
- at org.apache.spark.scheduler.cluster.k8s.ExecutorPodsAllocator.<init>(ExecutorPodsAllocator.scala:58)
- at org.apache.spark.scheduler.cluster.k8s.KubernetesClusterManager.createSchedulerBackend(KubernetesClusterManager.scala:113)
- at org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:2853)
- ... 19 more
- Caused by: okhttp3.internal.http2.ConnectionShutdownException
- at okhttp3.internal.http2.Http2Connection.newStream(Http2Connection.java:247)
- at okhttp3.internal.http2.Http2Connection.newStream(Http2Connection.java:230)
- at okhttp3.internal.http2.Http2Codec.writeRequestHeaders(Http2Codec.java:117)
- at okhttp3.internal.http.CallServerInterceptor.intercept(CallServerInterceptor.java:50)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.connection.ConnectInterceptor.intercept(ConnectInterceptor.java:45)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121)
- at okhttp3.internal.cache.CacheInterceptor.intercept(CacheInterceptor.java:93)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121)
- at okhttp3.internal.http.BridgeInterceptor.intercept(BridgeInterceptor.java:93)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RetryAndFollowUpInterceptor.intercept(RetryAndFollowUpInterceptor.java:127)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121)
- at io.fabric8.kubernetes.client.utils.BackwardsCompatibilityInterceptor.intercept(BackwardsCompatibilityInterceptor.java:134)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121)
- at io.fabric8.kubernetes.client.utils.ImpersonatorInterceptor.intercept(ImpersonatorInterceptor.java:68)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121)
- at io.fabric8.kubernetes.client.utils.HttpClientUtils.lambda$createHttpClient$3(HttpClientUtils.java:112)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121)
- at okhttp3.RealCall.getResponseWithInterceptorChain(RealCall.java:257)
- at okhttp3.RealCall.execute(RealCall.java:93)
- at io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleResponse(OperationSupport.java:469)
- at io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleResponse(OperationSupport.java:430)
- at io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleGet(OperationSupport.java:395)
- at io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleGet(OperationSupport.java:376)
- at io.fabric8.kubernetes.client.dsl.base.BaseOperation.handleGet(BaseOperation.java:832)
- at io.fabric8.kubernetes.client.dsl.base.BaseOperation.getMandatory(BaseOperation.java:214)
- ... 25 more
- 20/04/21 14:43:52 INFO SparkUI: Stopped Spark web UI at http://spark-pi-75235d719d31da3f-driver-svc.default.svc:4040
- 20/04/21 14:43:52 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
- 20/04/21 14:43:52 INFO MemoryStore: MemoryStore cleared
- 20/04/21 14:43:52 INFO BlockManager: BlockManager stopped
- 20/04/21 14:43:52 INFO BlockManagerMaster: BlockManagerMaster stopped
- 20/04/21 14:43:52 WARN MetricsSystem: Stopping a MetricsSystem that is not running
- 20/04/21 14:43:52 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
- 20/04/21 14:43:52 INFO SparkContext: Successfully stopped SparkContext
- Exception in thread "main" org.apache.spark.SparkException: External scheduler cannot be instantiated
- at org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:2859)
- at org.apache.spark.SparkContext.<init>(SparkContext.scala:533)
- at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2570)
- at org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$5(SparkSession.scala:937)
- at scala.Option.getOrElse(Option.scala:189)
- at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:928)
- at org.apache.spark.examples.SparkPi$.main(SparkPi.scala:30)
- at org.apache.spark.examples.SparkPi.main(SparkPi.scala)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
- at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:934)
- at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
- at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
- at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
- at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1013)
- at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1022)
- at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
- Caused by: io.fabric8.kubernetes.client.KubernetesClientException: Operation: [get] for kind: [Pod] with name: [spark-pi-75235d719d31da3f-driver] in namespace: [default] failed.
- at io.fabric8.kubernetes.client.KubernetesClientException.launderThrowable(KubernetesClientException.java:64)
- at io.fabric8.kubernetes.client.KubernetesClientException.launderThrowable(KubernetesClientException.java:72)
- at io.fabric8.kubernetes.client.dsl.base.BaseOperation.getMandatory(BaseOperation.java:225)
- at io.fabric8.kubernetes.client.dsl.base.BaseOperation.get(BaseOperation.java:168)
- at org.apache.spark.scheduler.cluster.k8s.ExecutorPodsAllocator.$anonfun$driverPod$1(ExecutorPodsAllocator.scala:59)
- at scala.Option.map(Option.scala:230)
- at org.apache.spark.scheduler.cluster.k8s.ExecutorPodsAllocator.<init>(ExecutorPodsAllocator.scala:58)
- at org.apache.spark.scheduler.cluster.k8s.KubernetesClusterManager.createSchedulerBackend(KubernetesClusterManager.scala:113)
- at org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:2853)
- ... 19 more
- Caused by: okhttp3.internal.http2.ConnectionShutdownException
- at okhttp3.internal.http2.Http2Connection.newStream(Http2Connection.java:247)
- at okhttp3.internal.http2.Http2Connection.newStream(Http2Connection.java:230)
- at okhttp3.internal.http2.Http2Codec.writeRequestHeaders(Http2Codec.java:117)
- at okhttp3.internal.http.CallServerInterceptor.intercept(CallServerInterceptor.java:50)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.connection.ConnectInterceptor.intercept(ConnectInterceptor.java:45)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121)
- at okhttp3.internal.cache.CacheInterceptor.intercept(CacheInterceptor.java:93)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121)
- at okhttp3.internal.http.BridgeInterceptor.intercept(BridgeInterceptor.java:93)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RetryAndFollowUpInterceptor.intercept(RetryAndFollowUpInterceptor.java:127)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121)
- at io.fabric8.kubernetes.client.utils.BackwardsCompatibilityInterceptor.intercept(BackwardsCompatibilityInterceptor.java:134)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121)
- at io.fabric8.kubernetes.client.utils.ImpersonatorInterceptor.intercept(ImpersonatorInterceptor.java:68)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121)
- at io.fabric8.kubernetes.client.utils.HttpClientUtils.lambda$createHttpClient$3(HttpClientUtils.java:112)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147)
- at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121)
- at okhttp3.RealCall.getResponseWithInterceptorChain(RealCall.java:257)
- at okhttp3.RealCall.execute(RealCall.java:93)
- at io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleResponse(OperationSupport.java:469)
- at io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleResponse(OperationSupport.java:430)
- at io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleGet(OperationSupport.java:395)
- at io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleGet(OperationSupport.java:376)
- at io.fabric8.kubernetes.client.dsl.base.BaseOperation.handleGet(BaseOperation.java:832)
- at io.fabric8.kubernetes.client.dsl.base.BaseOperation.getMandatory(BaseOperation.java:214)
- ... 25 more
- 20/04/21 14:43:52 INFO ShutdownHookManager: Shutdown hook called
- 20/04/21 14:43:52 INFO ShutdownHookManager: Deleting directory /tmp/spark-8d462ebb-91cc-4089-b61c-a274781c369a
- 20/04/21 14:43:52 INFO ShutdownHookManager: Deleting directory /var/data/spark-4fdc6242-3ab9-4ea4-864b-3b0ffb981e91/spark-373cdf3b-231a-4216-aae0-47bd3ca65874
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement