Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- at org.apache.spark.util.ClosureCleaner$.ensureSerializable(ClosureCleaner.scala:416)
- at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:406)
- at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:162)
- at org.apache.spark.SparkContext.clean(SparkContext.scala:2572)
- at org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndex$1(RDD.scala:893)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:125)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:395)
- at org.apache.spark.rdd.RDD.mapPartitionsWithIndex(RDD.scala:892)
- at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:725)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$3(SparkPlan.scala:247)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243)
- at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:187)
- at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.inputRDD$lzycompute(ShuffleExchangeExec.scala:64)
- at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.inputRDD(ShuffleExchangeExec.scala:64)
- at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.shuffleDependency$lzycompute(ShuffleExchangeExec.scala:83)
- at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.shuffleDependency(ShuffleExchangeExec.scala:81)
- at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.$anonfun$doExecute$1(ShuffleExchangeExec.scala:98)
- at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:52)
- at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.doExecute(ShuffleExchangeExec.scala:95)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$3(SparkPlan.scala:247)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243)
- at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:187)
- at org.apache.spark.sql.execution.InputAdapter.inputRDD(WholeStageCodegenExec.scala:527)
- at org.apache.spark.sql.execution.InputRDDCodegen.inputRDDs(WholeStageCodegenExec.scala:455)
- at org.apache.spark.sql.execution.InputRDDCodegen.inputRDDs$(WholeStageCodegenExec.scala:454)
- at org.apache.spark.sql.execution.InputAdapter.inputRDDs(WholeStageCodegenExec.scala:498)
- at org.apache.spark.sql.execution.aggregate.HashAggregateExec.inputRDDs(HashAggregateExec.scala:208)
- at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:722)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$3(SparkPlan.scala:247)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243)
- at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:187)
- at org.apache.spark.sql.execution.collect.Collector$.collect(Collector.scala:79)
- at org.apache.spark.sql.execution.collect.Collector$.collect(Collector.scala:88)
- at org.apache.spark.sql.execution.ResultCacheManager.getOrComputeResult(ResultCacheManager.scala:508)
- at org.apache.spark.sql.execution.ResultCacheManager.getOrComputeResult(ResultCacheManager.scala:480)
- at org.apache.spark.sql.execution.SparkPlan.executeCollectResult(SparkPlan.scala:396)
- at org.apache.spark.sql.Dataset.collectResult(Dataset.scala:2984)
- at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:3690)
- at org.apache.spark.sql.Dataset.$anonfun$collect$1(Dataset.scala:2951)
- at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3682)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withCustomExecutionEnv$5(SQLExecution.scala:115)
- at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:246)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withCustomExecutionEnv$1(SQLExecution.scala:100)
- at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:828)
- at org.apache.spark.sql.execution.SQLExecution$.withCustomExecutionEnv(SQLExecution.scala:76)
- at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:196)
- at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3680)
- at org.apache.spark.sql.Dataset.collect(Dataset.scala:2951)
- at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw$$iw.<init>(command-2523001380624274:18)
- at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw.<init>(command-2523001380624274:65)
- at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw.<init>(command-2523001380624274:67)
- at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw.<init>(command-2523001380624274:69)
- at line348719c49257466aa07acec6b546446427.$read$$iw$$iw.<init>(command-2523001380624274:71)
- at line348719c49257466aa07acec6b546446427.$read$$iw.<init>(command-2523001380624274:73)
- at line348719c49257466aa07acec6b546446427.$read.<init>(command-2523001380624274:75)
- at line348719c49257466aa07acec6b546446427.$read$.<init>(command-2523001380624274:79)
- at line348719c49257466aa07acec6b546446427.$read$.<clinit>(command-2523001380624274)
- at line348719c49257466aa07acec6b546446427.$eval$.$print$lzycompute(<notebook>:7)
- at line348719c49257466aa07acec6b546446427.$eval$.$print(<notebook>:6)
- at line348719c49257466aa07acec6b546446427.$eval.$print(<notebook>)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:745)
- at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1021)
- at scala.tools.nsc.interpreter.IMain.$anonfun$interpret$1(IMain.scala:574)
- at scala.reflect.internal.util.ScalaClassLoader.asContext(ScalaClassLoader.scala:41)
- at scala.reflect.internal.util.ScalaClassLoader.asContext$(ScalaClassLoader.scala:37)
- at scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:41)
- at scala.tools.nsc.interpreter.IMain.loadAndRunReq$1(IMain.scala:573)
- at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:600)
- at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:570)
- at com.databricks.backend.daemon.driver.DriverILoop.execute(DriverILoop.scala:215)
- at com.databricks.backend.daemon.driver.ScalaDriverLocal.$anonfun$repl$1(ScalaDriverLocal.scala:202)
- at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
- at com.databricks.backend.daemon.driver.DriverLocal$TrapExitInternal$.trapExit(DriverLocal.scala:714)
- at com.databricks.backend.daemon.driver.DriverLocal$TrapExit$.apply(DriverLocal.scala:667)
- at com.databricks.backend.daemon.driver.ScalaDriverLocal.repl(ScalaDriverLocal.scala:202)
- at com.databricks.backend.daemon.driver.DriverLocal.$anonfun$execute$10(DriverLocal.scala:396)
- at com.databricks.logging.UsageLogging.$anonfun$withAttributionContext$1(UsageLogging.scala:238)
- at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)
- at com.databricks.logging.UsageLogging.withAttributionContext(UsageLogging.scala:233)
- at com.databricks.logging.UsageLogging.withAttributionContext$(UsageLogging.scala:230)
- at com.databricks.backend.daemon.driver.DriverLocal.withAttributionContext(DriverLocal.scala:49)
- at com.databricks.logging.UsageLogging.withAttributionTags(UsageLogging.scala:275)
- at com.databricks.logging.UsageLogging.withAttributionTags$(UsageLogging.scala:268)
- at com.databricks.backend.daemon.driver.DriverLocal.withAttributionTags(DriverLocal.scala:49)
- at com.databricks.backend.daemon.driver.DriverLocal.execute(DriverLocal.scala:373)
- at com.databricks.backend.daemon.driver.DriverWrapper.$anonfun$tryExecutingCommand$1(DriverWrapper.scala:653)
- at scala.util.Try$.apply(Try.scala:213)
- at com.databricks.backend.daemon.driver.DriverWrapper.tryExecutingCommand(DriverWrapper.scala:645)
- at com.databricks.backend.daemon.driver.DriverWrapper.getCommandOutputAndError(DriverWrapper.scala:486)
- at com.databricks.backend.daemon.driver.DriverWrapper.executeCommand(DriverWrapper.scala:598)
- at com.databricks.backend.daemon.driver.DriverWrapper.runInnerLoop(DriverWrapper.scala:391)
- at com.databricks.backend.daemon.driver.DriverWrapper.runInner(DriverWrapper.scala:337)
- at com.databricks.backend.daemon.driver.DriverWrapper.run(DriverWrapper.scala:219)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.io.NotSerializableException: org.apache.spark.sql.TypedColumn
- Serialization stack:
- - object not serializable (class: org.apache.spark.sql.TypedColumn, value: $anon$1(boundreference() AS `value`, value, unresolveddeserializer(assertnotnull(upcast(getcolumnbyordinal(0, IntegerType), IntegerType, - root class: "scala.Int")), value#4), boundreference() AS `value`))
- - field (class: line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw$$iw, name: customSummer, type: class org.apache.spark.sql.TypedColumn)
- - object (class line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw$$iw, line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw$$iw@463f93db)
- - element of array (index: 2)
- - array (class [Ljava.lang.Object;, size 5)
- - element of array (index: 1)
- - array (class [Ljava.lang.Object;, size 3)
- - field (class: java.lang.invoke.SerializedLambda, name: capturedArgs, type: class [Ljava.lang.Object;)
- - object (class java.lang.invoke.SerializedLambda, SerializedLambda[capturingClass=class org.apache.spark.sql.execution.WholeStageCodegenExec, functionalInterfaceMethod=scala/Function2.apply:(Ljava/lang/Object;Ljava/lang/Object;)Ljava/lang/Object;, implementation=invokeStatic org/apache/spark/sql/execution/WholeStageCodegenExec.$anonfun$doExecute$4$adapted:(Lorg/apache/spark/sql/catalyst/expressions/codegen/CodeAndComment;[Ljava/lang/Object;Lorg/apache/spark/sql/execution/metric/SQLMetric;Ljava/lang/Object;Lscala/collection/Iterator;)Lscala/collection/Iterator;, instantiatedMethodType=(Ljava/lang/Object;Lscala/collection/Iterator;)Lscala/collection/Iterator;, numCaptured=3])
- - writeReplace data (class: java.lang.invoke.SerializedLambda)
- - object (class org.apache.spark.sql.execution.WholeStageCodegenExec$$Lambda$3961/49809591, org.apache.spark.sql.execution.WholeStageCodegenExec$$Lambda$3961/49809591@68c54c5e)
- at org.apache.spark.serializer.SerializationDebugger$.improveException(SerializationDebugger.scala:41)
- at org.apache.spark.serializer.JavaSerializationStream.writeObject(JavaSerializer.scala:47)
- at org.apache.spark.serializer.JavaSerializerInstance.serialize(JavaSerializer.scala:101)
- at org.apache.spark.util.ClosureCleaner$.ensureSerializable(ClosureCleaner.scala:413)
- at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:406)
- at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:162)
- at org.apache.spark.SparkContext.clean(SparkContext.scala:2572)
- at org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndex$1(RDD.scala:893)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:125)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:395)
- at org.apache.spark.rdd.RDD.mapPartitionsWithIndex(RDD.scala:892)
- at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:725)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$3(SparkPlan.scala:247)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243)
- at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:187)
- at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.inputRDD$lzycompute(ShuffleExchangeExec.scala:64)
- at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.inputRDD(ShuffleExchangeExec.scala:64)
- at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.shuffleDependency$lzycompute(ShuffleExchangeExec.scala:83)
- at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.shuffleDependency(ShuffleExchangeExec.scala:81)
- at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.$anonfun$doExecute$1(ShuffleExchangeExec.scala:98)
- at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:52)
- at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.doExecute(ShuffleExchangeExec.scala:95)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$3(SparkPlan.scala:247)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243)
- at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:187)
- at org.apache.spark.sql.execution.InputAdapter.inputRDD(WholeStageCodegenExec.scala:527)
- at org.apache.spark.sql.execution.InputRDDCodegen.inputRDDs(WholeStageCodegenExec.scala:455)
- at org.apache.spark.sql.execution.InputRDDCodegen.inputRDDs$(WholeStageCodegenExec.scala:454)
- at org.apache.spark.sql.execution.InputAdapter.inputRDDs(WholeStageCodegenExec.scala:498)
- at org.apache.spark.sql.execution.aggregate.HashAggregateExec.inputRDDs(HashAggregateExec.scala:208)
- at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:722)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195)
- at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$3(SparkPlan.scala:247)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243)
- at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:187)
- at org.apache.spark.sql.execution.collect.Collector$.collect(Collector.scala:79)
- at org.apache.spark.sql.execution.collect.Collector$.collect(Collector.scala:88)
- at org.apache.spark.sql.execution.ResultCacheManager.getOrComputeResult(ResultCacheManager.scala:508)
- at org.apache.spark.sql.execution.ResultCacheManager.getOrComputeResult(ResultCacheManager.scala:480)
- at org.apache.spark.sql.execution.SparkPlan.executeCollectResult(SparkPlan.scala:396)
- at org.apache.spark.sql.Dataset.collectResult(Dataset.scala:2984)
- at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:3690)
- at org.apache.spark.sql.Dataset.$anonfun$collect$1(Dataset.scala:2951)
- at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3682)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withCustomExecutionEnv$5(SQLExecution.scala:115)
- at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:246)
- at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withCustomExecutionEnv$1(SQLExecution.scala:100)
- at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:828)
- at org.apache.spark.sql.execution.SQLExecution$.withCustomExecutionEnv(SQLExecution.scala:76)
- at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:196)
- at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3680)
- at org.apache.spark.sql.Dataset.collect(Dataset.scala:2951)
- at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw$$iw.<init>(command-2523001380624274:18)
- at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw.<init>(command-2523001380624274:65)
- at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw.<init>(command-2523001380624274:67)
- at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw.<init>(command-2523001380624274:69)
- at line348719c49257466aa07acec6b546446427.$read$$iw$$iw.<init>(command-2523001380624274:71)
- at line348719c49257466aa07acec6b546446427.$read$$iw.<init>(command-2523001380624274:73)
- at line348719c49257466aa07acec6b546446427.$read.<init>(command-2523001380624274:75)
- at line348719c49257466aa07acec6b546446427.$read$.<init>(command-2523001380624274:79)
- at line348719c49257466aa07acec6b546446427.$read$.<clinit>(command-2523001380624274)
- at line348719c49257466aa07acec6b546446427.$eval$.$print$lzycompute(<notebook>:7)
- at line348719c49257466aa07acec6b546446427.$eval$.$print(<notebook>:6)
- at line348719c49257466aa07acec6b546446427.$eval.$print(<notebook>)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:745)
- at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1021)
- at scala.tools.nsc.interpreter.IMain.$anonfun$interpret$1(IMain.scala:574)
- at scala.reflect.internal.util.ScalaClassLoader.asContext(ScalaClassLoader.scala:41)
- at scala.reflect.internal.util.ScalaClassLoader.asContext$(ScalaClassLoader.scala:37)
- at scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:41)
- at scala.tools.nsc.interpreter.IMain.loadAndRunReq$1(IMain.scala:573)
- at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:600)
- at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:570)
- at com.databricks.backend.daemon.driver.DriverILoop.execute(DriverILoop.scala:215)
- at com.databricks.backend.daemon.driver.ScalaDriverLocal.$anonfun$repl$1(ScalaDriverLocal.scala:202)
- at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
- at com.databricks.backend.daemon.driver.DriverLocal$TrapExitInternal$.trapExit(DriverLocal.scala:714)
- at com.databricks.backend.daemon.driver.DriverLocal$TrapExit$.apply(DriverLocal.scala:667)
- at com.databricks.backend.daemon.driver.ScalaDriverLocal.repl(ScalaDriverLocal.scala:202)
- at com.databricks.backend.daemon.driver.DriverLocal.$anonfun$execute$10(DriverLocal.scala:396)
- at com.databricks.logging.UsageLogging.$anonfun$withAttributionContext$1(UsageLogging.scala:238)
- at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)
- at com.databricks.logging.UsageLogging.withAttributionContext(UsageLogging.scala:233)
- at com.databricks.logging.UsageLogging.withAttributionContext$(UsageLogging.scala:230)
- at com.databricks.backend.daemon.driver.DriverLocal.withAttributionContext(DriverLocal.scala:49)
- at com.databricks.logging.UsageLogging.withAttributionTags(UsageLogging.scala:275)
- at com.databricks.logging.UsageLogging.withAttributionTags$(UsageLogging.scala:268)
- at com.databricks.backend.daemon.driver.DriverLocal.withAttributionTags(DriverLocal.scala:49)
- at com.databricks.backend.daemon.driver.DriverLocal.execute(DriverLocal.scala:373)
- at com.databricks.backend.daemon.driver.DriverWrapper.$anonfun$tryExecutingCommand$1(DriverWrapper.scala:653)
- at scala.util.Try$.apply(Try.scala:213)
- at com.databricks.backend.daemon.driver.DriverWrapper.tryExecutingCommand(DriverWrapper.scala:645)
- at com.databricks.backend.daemon.driver.DriverWrapper.getCommandOutputAndError(DriverWrapper.scala:486)
- at com.databricks.backend.daemon.driver.DriverWrapper.executeCommand(DriverWrapper.scala:598)
- at com.databricks.backend.daemon.driver.DriverWrapper.runInnerLoop(DriverWrapper.scala:391)
- at com.databricks.backend.daemon.driver.DriverWrapper.runInner(DriverWrapper.scala:337)
- at com.databricks.backend.daemon.driver.DriverWrapper.run(DriverWrapper.scala:219)
- at java.lang.Thread.run(Thread.java:748)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement