Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- Traceback (most recent call last):
- File "<stdin>", line 1, in <module>
- File "/Users/marwanbaghdad/miniconda3/envs/pyspark/lib/python3.6/site-packages/pyspark/sql/dataframe.py", line 378, in show
- print(self._jdf.showString(n, 20, vertical))
- File "/Users/marwanbaghdad/miniconda3/envs/pyspark/lib/python3.6/site-packages/pyspark/python/lib/py4j-0.10.7-src.zip/py4j/java_gateway.py", line 1257, in __call__
- File "/Users/marwanbaghdad/miniconda3/envs/pyspark/lib/python3.6/site-packages/pyspark/sql/utils.py", line 63, in deco
- return f(*a, **kw)
- File "/Users/marwanbaghdad/miniconda3/envs/pyspark/lib/python3.6/site-packages/pyspark/python/lib/py4j-0.10.7-src.zip/py4j/protocol.py", line 328, in get_return_value
- py4j.protocol.Py4JJavaError: An error occurred while calling o302.showString.
- : org.apache.spark.SparkException: Exception thrown in awaitResult:
- at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:226)
- at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec.doExecuteBroadcast(BroadcastExchangeExec.scala:146)
- at org.apache.spark.sql.execution.InputAdapter.doExecuteBroadcast(WholeStageCodegenExec.scala:387)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeBroadcast$1.apply(SparkPlan.scala:144)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeBroadcast$1.apply(SparkPlan.scala:140)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
- at org.apache.spark.sql.execution.SparkPlan.executeBroadcast(SparkPlan.scala:140)
- at org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.prepareBroadcast(BroadcastHashJoinExec.scala:117)
- at org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.codegenInner(BroadcastHashJoinExec.scala:211)
- at org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doConsume(BroadcastHashJoinExec.scala:101)
- at org.apache.spark.sql.execution.CodegenSupport$class.constructDoConsumeFunction(WholeStageCodegenExec.scala:216)
- at org.apache.spark.sql.execution.CodegenSupport$class.consume(WholeStageCodegenExec.scala:187)
- at org.apache.spark.sql.execution.ProjectExec.consume(basicPhysicalOperators.scala:35)
- at org.apache.spark.sql.execution.ProjectExec.doConsume(basicPhysicalOperators.scala:65)
- at org.apache.spark.sql.execution.CodegenSupport$class.consume(WholeStageCodegenExec.scala:189)
- at org.apache.spark.sql.execution.GlobalLimitExec.consume(limit.scala:108)
- at org.apache.spark.sql.execution.BaseLimitExec$class.doConsume(limit.scala:87)
- at org.apache.spark.sql.execution.GlobalLimitExec.doConsume(limit.scala:108)
- at org.apache.spark.sql.execution.CodegenSupport$class.consume(WholeStageCodegenExec.scala:189)
- at org.apache.spark.sql.execution.InputAdapter.consume(WholeStageCodegenExec.scala:374)
- at org.apache.spark.sql.execution.InputAdapter.doProduce(WholeStageCodegenExec.scala:403)
- at org.apache.spark.sql.execution.CodegenSupport$$anonfun$produce$1.apply(WholeStageCodegenExec.scala:90)
- at org.apache.spark.sql.execution.CodegenSupport$$anonfun$produce$1.apply(WholeStageCodegenExec.scala:85)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
- at org.apache.spark.sql.execution.CodegenSupport$class.produce(WholeStageCodegenExec.scala:85)
- at org.apache.spark.sql.execution.InputAdapter.produce(WholeStageCodegenExec.scala:374)
- at org.apache.spark.sql.execution.BaseLimitExec$class.doProduce(limit.scala:70)
- at org.apache.spark.sql.execution.GlobalLimitExec.doProduce(limit.scala:108)
- at org.apache.spark.sql.execution.CodegenSupport$$anonfun$produce$1.apply(WholeStageCodegenExec.scala:90)
- at org.apache.spark.sql.execution.CodegenSupport$$anonfun$produce$1.apply(WholeStageCodegenExec.scala:85)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
- at org.apache.spark.sql.execution.CodegenSupport$class.produce(WholeStageCodegenExec.scala:85)
- at org.apache.spark.sql.execution.GlobalLimitExec.produce(limit.scala:108)
- at org.apache.spark.sql.execution.ProjectExec.doProduce(basicPhysicalOperators.scala:45)
- at org.apache.spark.sql.execution.CodegenSupport$$anonfun$produce$1.apply(WholeStageCodegenExec.scala:90)
- at org.apache.spark.sql.execution.CodegenSupport$$anonfun$produce$1.apply(WholeStageCodegenExec.scala:85)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
- at org.apache.spark.sql.execution.CodegenSupport$class.produce(WholeStageCodegenExec.scala:85)
- at org.apache.spark.sql.execution.ProjectExec.produce(basicPhysicalOperators.scala:35)
- at org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doProduce(BroadcastHashJoinExec.scala:96)
- at org.apache.spark.sql.execution.CodegenSupport$$anonfun$produce$1.apply(WholeStageCodegenExec.scala:90)
- at org.apache.spark.sql.execution.CodegenSupport$$anonfun$produce$1.apply(WholeStageCodegenExec.scala:85)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
- at org.apache.spark.sql.execution.CodegenSupport$class.produce(WholeStageCodegenExec.scala:85)
- at org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.produce(BroadcastHashJoinExec.scala:40)
- at org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.doProduce(BroadcastHashJoinExec.scala:96)
- at org.apache.spark.sql.execution.CodegenSupport$$anonfun$produce$1.apply(WholeStageCodegenExec.scala:90)
- at org.apache.spark.sql.execution.CodegenSupport$$anonfun$produce$1.apply(WholeStageCodegenExec.scala:85)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
- at org.apache.spark.sql.execution.CodegenSupport$class.produce(WholeStageCodegenExec.scala:85)
- at org.apache.spark.sql.execution.joins.BroadcastHashJoinExec.produce(BroadcastHashJoinExec.scala:40)
- at org.apache.spark.sql.execution.ProjectExec.doProduce(basicPhysicalOperators.scala:45)
- at org.apache.spark.sql.execution.CodegenSupport$$anonfun$produce$1.apply(WholeStageCodegenExec.scala:90)
- at org.apache.spark.sql.execution.CodegenSupport$$anonfun$produce$1.apply(WholeStageCodegenExec.scala:85)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
- at org.apache.spark.sql.execution.CodegenSupport$class.produce(WholeStageCodegenExec.scala:85)
- at org.apache.spark.sql.execution.ProjectExec.produce(basicPhysicalOperators.scala:35)
- at org.apache.spark.sql.execution.WholeStageCodegenExec.doCodeGen(WholeStageCodegenExec.scala:544)
- at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:598)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127)
- at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
- at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127)
- at org.apache.spark.sql.execution.SparkPlan.getByteArrayRdd(SparkPlan.scala:247)
- at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:339)
- at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:38)
- at org.apache.spark.sql.Dataset.org$apache$spark$sql$Dataset$$collectFromPlan(Dataset.scala:3383)
- at org.apache.spark.sql.Dataset$$anonfun$head$1.apply(Dataset.scala:2544)
- at org.apache.spark.sql.Dataset$$anonfun$head$1.apply(Dataset.scala:2544)
- at org.apache.spark.sql.Dataset$$anonfun$53.apply(Dataset.scala:3364)
- at org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:78)
- at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:125)
- at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:73)
- at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3363)
- at org.apache.spark.sql.Dataset.head(Dataset.scala:2544)
- at org.apache.spark.sql.Dataset.take(Dataset.scala:2758)
- at org.apache.spark.sql.Dataset.getRows(Dataset.scala:254)
- at org.apache.spark.sql.Dataset.showString(Dataset.scala:291)
- at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.base/java.lang.reflect.Method.invoke(Method.java:567)
- at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
- at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
- at py4j.Gateway.invoke(Gateway.java:282)
- at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
- at py4j.commands.CallCommand.execute(CallCommand.java:79)
- at py4j.GatewayConnection.run(GatewayConnection.java:238)
- at java.base/java.lang.Thread.run(Thread.java:835)
- Caused by: java.lang.IllegalArgumentException: Unsupported class file major version 56
- at org.apache.xbean.asm6.ClassReader.<init>(ClassReader.java:166)
- at org.apache.xbean.asm6.ClassReader.<init>(ClassReader.java:148)
- at org.apache.xbean.asm6.ClassReader.<init>(ClassReader.java:136)
- at org.apache.xbean.asm6.ClassReader.<init>(ClassReader.java:237)
- at org.apache.spark.util.ClosureCleaner$.getClassReader(ClosureCleaner.scala:49)
- at org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:517)
- at org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:500)
- at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
- at scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:134)
- at scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:134)
- at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:236)
- at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
- at scala.collection.mutable.HashMap$$anon$1.foreach(HashMap.scala:134)
- at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
- at org.apache.spark.util.FieldAccessFinder$$anon$3.visitMethodInsn(ClosureCleaner.scala:500)
- at org.apache.xbean.asm6.ClassReader.readCode(ClassReader.java:2175)
- at org.apache.xbean.asm6.ClassReader.readMethod(ClassReader.java:1238)
- at org.apache.xbean.asm6.ClassReader.accept(ClassReader.java:631)
- at org.apache.xbean.asm6.ClassReader.accept(ClassReader.java:355)
- at org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:307)
- at org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:306)
- at scala.collection.immutable.List.foreach(List.scala:392)
- at org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:306)
- at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:162)
- at org.apache.spark.SparkContext.clean(SparkContext.scala:2326)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2100)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2126)
- at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:945)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:363)
- at org.apache.spark.rdd.RDD.collect(RDD.scala:944)
- at org.apache.spark.sql.execution.SparkPlan.executeCollectIterator(SparkPlan.scala:306)
- at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1$$anonfun$apply$1.apply(BroadcastExchangeExec.scala:79)
- at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1$$anonfun$apply$1.apply(BroadcastExchangeExec.scala:76)
- at org.apache.spark.sql.execution.SQLExecution$$anonfun$withExecutionId$1.apply(SQLExecution.scala:101)
- at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:125)
- at org.apache.spark.sql.execution.SQLExecution$.withExecutionId(SQLExecution.scala:98)
- at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1.apply(BroadcastExchangeExec.scala:75)
- at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1.apply(BroadcastExchangeExec.scala:75)
- at scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
- at scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
- at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
- at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
- ... 1 more
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement