Advertisement
Guest User

Spark Aggregator Stack Trace

a guest
Jul 8th, 2020
78
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 20.08 KB | None | 0 0
  1. at org.apache.spark.util.ClosureCleaner$.ensureSerializable(ClosureCleaner.scala:416)
  2. at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:406)
  3. at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:162)
  4. at org.apache.spark.SparkContext.clean(SparkContext.scala:2572)
  5. at org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndex$1(RDD.scala:893)
  6. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
  7. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:125)
  8. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
  9. at org.apache.spark.rdd.RDD.withScope(RDD.scala:395)
  10. at org.apache.spark.rdd.RDD.mapPartitionsWithIndex(RDD.scala:892)
  11. at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:725)
  12. at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195)
  13. at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$3(SparkPlan.scala:247)
  14. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
  15. at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243)
  16. at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:187)
  17. at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.inputRDD$lzycompute(ShuffleExchangeExec.scala:64)
  18. at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.inputRDD(ShuffleExchangeExec.scala:64)
  19. at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.shuffleDependency$lzycompute(ShuffleExchangeExec.scala:83)
  20. at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.shuffleDependency(ShuffleExchangeExec.scala:81)
  21. at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.$anonfun$doExecute$1(ShuffleExchangeExec.scala:98)
  22. at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:52)
  23. at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.doExecute(ShuffleExchangeExec.scala:95)
  24. at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195)
  25. at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$3(SparkPlan.scala:247)
  26. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
  27. at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243)
  28. at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:187)
  29. at org.apache.spark.sql.execution.InputAdapter.inputRDD(WholeStageCodegenExec.scala:527)
  30. at org.apache.spark.sql.execution.InputRDDCodegen.inputRDDs(WholeStageCodegenExec.scala:455)
  31. at org.apache.spark.sql.execution.InputRDDCodegen.inputRDDs$(WholeStageCodegenExec.scala:454)
  32. at org.apache.spark.sql.execution.InputAdapter.inputRDDs(WholeStageCodegenExec.scala:498)
  33. at org.apache.spark.sql.execution.aggregate.HashAggregateExec.inputRDDs(HashAggregateExec.scala:208)
  34. at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:722)
  35. at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195)
  36. at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$3(SparkPlan.scala:247)
  37. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
  38. at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243)
  39. at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:187)
  40. at org.apache.spark.sql.execution.collect.Collector$.collect(Collector.scala:79)
  41. at org.apache.spark.sql.execution.collect.Collector$.collect(Collector.scala:88)
  42. at org.apache.spark.sql.execution.ResultCacheManager.getOrComputeResult(ResultCacheManager.scala:508)
  43. at org.apache.spark.sql.execution.ResultCacheManager.getOrComputeResult(ResultCacheManager.scala:480)
  44. at org.apache.spark.sql.execution.SparkPlan.executeCollectResult(SparkPlan.scala:396)
  45. at org.apache.spark.sql.Dataset.collectResult(Dataset.scala:2984)
  46. at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:3690)
  47. at org.apache.spark.sql.Dataset.$anonfun$collect$1(Dataset.scala:2951)
  48. at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3682)
  49. at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withCustomExecutionEnv$5(SQLExecution.scala:115)
  50. at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:246)
  51. at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withCustomExecutionEnv$1(SQLExecution.scala:100)
  52. at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:828)
  53. at org.apache.spark.sql.execution.SQLExecution$.withCustomExecutionEnv(SQLExecution.scala:76)
  54. at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:196)
  55. at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3680)
  56. at org.apache.spark.sql.Dataset.collect(Dataset.scala:2951)
  57. at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw$$iw.<init>(command-2523001380624274:18)
  58. at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw.<init>(command-2523001380624274:65)
  59. at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw.<init>(command-2523001380624274:67)
  60. at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw.<init>(command-2523001380624274:69)
  61. at line348719c49257466aa07acec6b546446427.$read$$iw$$iw.<init>(command-2523001380624274:71)
  62. at line348719c49257466aa07acec6b546446427.$read$$iw.<init>(command-2523001380624274:73)
  63. at line348719c49257466aa07acec6b546446427.$read.<init>(command-2523001380624274:75)
  64. at line348719c49257466aa07acec6b546446427.$read$.<init>(command-2523001380624274:79)
  65. at line348719c49257466aa07acec6b546446427.$read$.<clinit>(command-2523001380624274)
  66. at line348719c49257466aa07acec6b546446427.$eval$.$print$lzycompute(<notebook>:7)
  67. at line348719c49257466aa07acec6b546446427.$eval$.$print(<notebook>:6)
  68. at line348719c49257466aa07acec6b546446427.$eval.$print(<notebook>)
  69. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  70. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  71. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  72. at java.lang.reflect.Method.invoke(Method.java:498)
  73. at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:745)
  74. at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1021)
  75. at scala.tools.nsc.interpreter.IMain.$anonfun$interpret$1(IMain.scala:574)
  76. at scala.reflect.internal.util.ScalaClassLoader.asContext(ScalaClassLoader.scala:41)
  77. at scala.reflect.internal.util.ScalaClassLoader.asContext$(ScalaClassLoader.scala:37)
  78. at scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:41)
  79. at scala.tools.nsc.interpreter.IMain.loadAndRunReq$1(IMain.scala:573)
  80. at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:600)
  81. at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:570)
  82. at com.databricks.backend.daemon.driver.DriverILoop.execute(DriverILoop.scala:215)
  83. at com.databricks.backend.daemon.driver.ScalaDriverLocal.$anonfun$repl$1(ScalaDriverLocal.scala:202)
  84. at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
  85. at com.databricks.backend.daemon.driver.DriverLocal$TrapExitInternal$.trapExit(DriverLocal.scala:714)
  86. at com.databricks.backend.daemon.driver.DriverLocal$TrapExit$.apply(DriverLocal.scala:667)
  87. at com.databricks.backend.daemon.driver.ScalaDriverLocal.repl(ScalaDriverLocal.scala:202)
  88. at com.databricks.backend.daemon.driver.DriverLocal.$anonfun$execute$10(DriverLocal.scala:396)
  89. at com.databricks.logging.UsageLogging.$anonfun$withAttributionContext$1(UsageLogging.scala:238)
  90. at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)
  91. at com.databricks.logging.UsageLogging.withAttributionContext(UsageLogging.scala:233)
  92. at com.databricks.logging.UsageLogging.withAttributionContext$(UsageLogging.scala:230)
  93. at com.databricks.backend.daemon.driver.DriverLocal.withAttributionContext(DriverLocal.scala:49)
  94. at com.databricks.logging.UsageLogging.withAttributionTags(UsageLogging.scala:275)
  95. at com.databricks.logging.UsageLogging.withAttributionTags$(UsageLogging.scala:268)
  96. at com.databricks.backend.daemon.driver.DriverLocal.withAttributionTags(DriverLocal.scala:49)
  97. at com.databricks.backend.daemon.driver.DriverLocal.execute(DriverLocal.scala:373)
  98. at com.databricks.backend.daemon.driver.DriverWrapper.$anonfun$tryExecutingCommand$1(DriverWrapper.scala:653)
  99. at scala.util.Try$.apply(Try.scala:213)
  100. at com.databricks.backend.daemon.driver.DriverWrapper.tryExecutingCommand(DriverWrapper.scala:645)
  101. at com.databricks.backend.daemon.driver.DriverWrapper.getCommandOutputAndError(DriverWrapper.scala:486)
  102. at com.databricks.backend.daemon.driver.DriverWrapper.executeCommand(DriverWrapper.scala:598)
  103. at com.databricks.backend.daemon.driver.DriverWrapper.runInnerLoop(DriverWrapper.scala:391)
  104. at com.databricks.backend.daemon.driver.DriverWrapper.runInner(DriverWrapper.scala:337)
  105. at com.databricks.backend.daemon.driver.DriverWrapper.run(DriverWrapper.scala:219)
  106. at java.lang.Thread.run(Thread.java:748)
  107. Caused by: java.io.NotSerializableException: org.apache.spark.sql.TypedColumn
  108. Serialization stack:
  109. - object not serializable (class: org.apache.spark.sql.TypedColumn, value: $anon$1(boundreference() AS `value`, value, unresolveddeserializer(assertnotnull(upcast(getcolumnbyordinal(0, IntegerType), IntegerType, - root class: "scala.Int")), value#4), boundreference() AS `value`))
  110. - field (class: line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw$$iw, name: customSummer, type: class org.apache.spark.sql.TypedColumn)
  111. - object (class line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw$$iw, line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw$$iw@463f93db)
  112. - element of array (index: 2)
  113. - array (class [Ljava.lang.Object;, size 5)
  114. - element of array (index: 1)
  115. - array (class [Ljava.lang.Object;, size 3)
  116. - field (class: java.lang.invoke.SerializedLambda, name: capturedArgs, type: class [Ljava.lang.Object;)
  117. - object (class java.lang.invoke.SerializedLambda, SerializedLambda[capturingClass=class org.apache.spark.sql.execution.WholeStageCodegenExec, functionalInterfaceMethod=scala/Function2.apply:(Ljava/lang/Object;Ljava/lang/Object;)Ljava/lang/Object;, implementation=invokeStatic org/apache/spark/sql/execution/WholeStageCodegenExec.$anonfun$doExecute$4$adapted:(Lorg/apache/spark/sql/catalyst/expressions/codegen/CodeAndComment;[Ljava/lang/Object;Lorg/apache/spark/sql/execution/metric/SQLMetric;Ljava/lang/Object;Lscala/collection/Iterator;)Lscala/collection/Iterator;, instantiatedMethodType=(Ljava/lang/Object;Lscala/collection/Iterator;)Lscala/collection/Iterator;, numCaptured=3])
  118. - writeReplace data (class: java.lang.invoke.SerializedLambda)
  119. - object (class org.apache.spark.sql.execution.WholeStageCodegenExec$$Lambda$3961/49809591, org.apache.spark.sql.execution.WholeStageCodegenExec$$Lambda$3961/49809591@68c54c5e)
  120. at org.apache.spark.serializer.SerializationDebugger$.improveException(SerializationDebugger.scala:41)
  121. at org.apache.spark.serializer.JavaSerializationStream.writeObject(JavaSerializer.scala:47)
  122. at org.apache.spark.serializer.JavaSerializerInstance.serialize(JavaSerializer.scala:101)
  123. at org.apache.spark.util.ClosureCleaner$.ensureSerializable(ClosureCleaner.scala:413)
  124. at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:406)
  125. at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:162)
  126. at org.apache.spark.SparkContext.clean(SparkContext.scala:2572)
  127. at org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndex$1(RDD.scala:893)
  128. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
  129. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:125)
  130. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
  131. at org.apache.spark.rdd.RDD.withScope(RDD.scala:395)
  132. at org.apache.spark.rdd.RDD.mapPartitionsWithIndex(RDD.scala:892)
  133. at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:725)
  134. at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195)
  135. at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$3(SparkPlan.scala:247)
  136. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
  137. at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243)
  138. at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:187)
  139. at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.inputRDD$lzycompute(ShuffleExchangeExec.scala:64)
  140. at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.inputRDD(ShuffleExchangeExec.scala:64)
  141. at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.shuffleDependency$lzycompute(ShuffleExchangeExec.scala:83)
  142. at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.shuffleDependency(ShuffleExchangeExec.scala:81)
  143. at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.$anonfun$doExecute$1(ShuffleExchangeExec.scala:98)
  144. at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:52)
  145. at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.doExecute(ShuffleExchangeExec.scala:95)
  146. at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195)
  147. at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$3(SparkPlan.scala:247)
  148. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
  149. at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243)
  150. at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:187)
  151. at org.apache.spark.sql.execution.InputAdapter.inputRDD(WholeStageCodegenExec.scala:527)
  152. at org.apache.spark.sql.execution.InputRDDCodegen.inputRDDs(WholeStageCodegenExec.scala:455)
  153. at org.apache.spark.sql.execution.InputRDDCodegen.inputRDDs$(WholeStageCodegenExec.scala:454)
  154. at org.apache.spark.sql.execution.InputAdapter.inputRDDs(WholeStageCodegenExec.scala:498)
  155. at org.apache.spark.sql.execution.aggregate.HashAggregateExec.inputRDDs(HashAggregateExec.scala:208)
  156. at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:722)
  157. at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:195)
  158. at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$3(SparkPlan.scala:247)
  159. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:165)
  160. at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:243)
  161. at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:187)
  162. at org.apache.spark.sql.execution.collect.Collector$.collect(Collector.scala:79)
  163. at org.apache.spark.sql.execution.collect.Collector$.collect(Collector.scala:88)
  164. at org.apache.spark.sql.execution.ResultCacheManager.getOrComputeResult(ResultCacheManager.scala:508)
  165. at org.apache.spark.sql.execution.ResultCacheManager.getOrComputeResult(ResultCacheManager.scala:480)
  166. at org.apache.spark.sql.execution.SparkPlan.executeCollectResult(SparkPlan.scala:396)
  167. at org.apache.spark.sql.Dataset.collectResult(Dataset.scala:2984)
  168. at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:3690)
  169. at org.apache.spark.sql.Dataset.$anonfun$collect$1(Dataset.scala:2951)
  170. at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3682)
  171. at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withCustomExecutionEnv$5(SQLExecution.scala:115)
  172. at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:246)
  173. at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withCustomExecutionEnv$1(SQLExecution.scala:100)
  174. at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:828)
  175. at org.apache.spark.sql.execution.SQLExecution$.withCustomExecutionEnv(SQLExecution.scala:76)
  176. at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:196)
  177. at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3680)
  178. at org.apache.spark.sql.Dataset.collect(Dataset.scala:2951)
  179. at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw$$iw.<init>(command-2523001380624274:18)
  180. at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw$$iw.<init>(command-2523001380624274:65)
  181. at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw$$iw.<init>(command-2523001380624274:67)
  182. at line348719c49257466aa07acec6b546446427.$read$$iw$$iw$$iw.<init>(command-2523001380624274:69)
  183. at line348719c49257466aa07acec6b546446427.$read$$iw$$iw.<init>(command-2523001380624274:71)
  184. at line348719c49257466aa07acec6b546446427.$read$$iw.<init>(command-2523001380624274:73)
  185. at line348719c49257466aa07acec6b546446427.$read.<init>(command-2523001380624274:75)
  186. at line348719c49257466aa07acec6b546446427.$read$.<init>(command-2523001380624274:79)
  187. at line348719c49257466aa07acec6b546446427.$read$.<clinit>(command-2523001380624274)
  188. at line348719c49257466aa07acec6b546446427.$eval$.$print$lzycompute(<notebook>:7)
  189. at line348719c49257466aa07acec6b546446427.$eval$.$print(<notebook>:6)
  190. at line348719c49257466aa07acec6b546446427.$eval.$print(<notebook>)
  191. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  192. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  193. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  194. at java.lang.reflect.Method.invoke(Method.java:498)
  195. at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:745)
  196. at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1021)
  197. at scala.tools.nsc.interpreter.IMain.$anonfun$interpret$1(IMain.scala:574)
  198. at scala.reflect.internal.util.ScalaClassLoader.asContext(ScalaClassLoader.scala:41)
  199. at scala.reflect.internal.util.ScalaClassLoader.asContext$(ScalaClassLoader.scala:37)
  200. at scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:41)
  201. at scala.tools.nsc.interpreter.IMain.loadAndRunReq$1(IMain.scala:573)
  202. at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:600)
  203. at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:570)
  204. at com.databricks.backend.daemon.driver.DriverILoop.execute(DriverILoop.scala:215)
  205. at com.databricks.backend.daemon.driver.ScalaDriverLocal.$anonfun$repl$1(ScalaDriverLocal.scala:202)
  206. at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
  207. at com.databricks.backend.daemon.driver.DriverLocal$TrapExitInternal$.trapExit(DriverLocal.scala:714)
  208. at com.databricks.backend.daemon.driver.DriverLocal$TrapExit$.apply(DriverLocal.scala:667)
  209. at com.databricks.backend.daemon.driver.ScalaDriverLocal.repl(ScalaDriverLocal.scala:202)
  210. at com.databricks.backend.daemon.driver.DriverLocal.$anonfun$execute$10(DriverLocal.scala:396)
  211. at com.databricks.logging.UsageLogging.$anonfun$withAttributionContext$1(UsageLogging.scala:238)
  212. at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)
  213. at com.databricks.logging.UsageLogging.withAttributionContext(UsageLogging.scala:233)
  214. at com.databricks.logging.UsageLogging.withAttributionContext$(UsageLogging.scala:230)
  215. at com.databricks.backend.daemon.driver.DriverLocal.withAttributionContext(DriverLocal.scala:49)
  216. at com.databricks.logging.UsageLogging.withAttributionTags(UsageLogging.scala:275)
  217. at com.databricks.logging.UsageLogging.withAttributionTags$(UsageLogging.scala:268)
  218. at com.databricks.backend.daemon.driver.DriverLocal.withAttributionTags(DriverLocal.scala:49)
  219. at com.databricks.backend.daemon.driver.DriverLocal.execute(DriverLocal.scala:373)
  220. at com.databricks.backend.daemon.driver.DriverWrapper.$anonfun$tryExecutingCommand$1(DriverWrapper.scala:653)
  221. at scala.util.Try$.apply(Try.scala:213)
  222. at com.databricks.backend.daemon.driver.DriverWrapper.tryExecutingCommand(DriverWrapper.scala:645)
  223. at com.databricks.backend.daemon.driver.DriverWrapper.getCommandOutputAndError(DriverWrapper.scala:486)
  224. at com.databricks.backend.daemon.driver.DriverWrapper.executeCommand(DriverWrapper.scala:598)
  225. at com.databricks.backend.daemon.driver.DriverWrapper.runInnerLoop(DriverWrapper.scala:391)
  226. at com.databricks.backend.daemon.driver.DriverWrapper.runInner(DriverWrapper.scala:337)
  227. at com.databricks.backend.daemon.driver.DriverWrapper.run(DriverWrapper.scala:219)
  228. at java.lang.Thread.run(Thread.java:748)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement