Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- 14/10/25 16:35:36 ERROR Executor: Exception in task 3.0 in stage 4.0 (TID 27)
- org.apache.spark.api.python.PythonException: Traceback (most recent call last):
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/worker.py", line 79, in main
- serializer.dump_stream(func(split_index, iterator), outfile)
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1990, in pipeline_func
- return func(split, prev_func(split, iterator))
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1990, in pipeline_func
- return func(split, prev_func(split, iterator))
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1990, in pipeline_func
- return func(split, prev_func(split, iterator))
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1990, in pipeline_func
- return func(split, prev_func(split, iterator))
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 352, in func
- return f(iterator)
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1585, in _mergeCombiners
- merger.mergeCombiners(iterator)
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/shuffle.py", line 282, in mergeCombiners
- d[k] = comb(d[k], v) if k in d else v
- File "/home/cc/cs61c/fa14/class/cs61c-ef/proj2/SlidingBfsSpark.py", line 10, in bfs_reduce
- if value1[1] < value2[1]:
- TypeError: 'int' object has no attribute '__getitem__'
- at org.apache.spark.api.python.PythonRDD$$anon$1.read(PythonRDD.scala:124)
- at org.apache.spark.api.python.PythonRDD$$anon$1.<init>(PythonRDD.scala:154)
- at org.apache.spark.api.python.PythonRDD.compute(PythonRDD.scala:87)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
- at org.apache.spark.api.python.PairwiseRDD.compute(PythonRDD.scala:265)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
- at org.apache.spark.scheduler.Task.run(Task.scala:54)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
- at java.lang.Thread.run(Thread.java:745)
- 14/10/25 16:35:36 INFO TaskSetManager: Finished task 6.0 in stage 4.0 (TID 30) in 41 ms on localhost (2/8)
- 14/10/25 16:35:36 INFO Executor: Finished task 7.0 in stage 4.0 (TID 31). 1014 bytes result sent to driver
- 14/10/25 16:35:36 INFO TaskSetManager: Finished task 0.0 in stage 4.0 (TID 24) in 60 ms on localhost (3/8)
- 14/10/25 16:35:36 INFO TaskSetManager: Finished task 2.0 in stage 4.0 (TID 26) in 65 ms on localhost (4/8)
- 14/10/25 16:35:36 INFO TaskSetManager: Finished task 4.0 in stage 4.0 (TID 28) in 67 ms on localhost (5/8)
- 14/10/25 16:35:36 INFO TaskSetManager: Finished task 5.0 in stage 4.0 (TID 29) in 73 ms on localhost (6/8)
- 14/10/25 16:35:36 INFO TaskSetManager: Finished task 7.0 in stage 4.0 (TID 31) in 76 ms on localhost (7/8)
- 14/10/25 16:35:36 WARN TaskSetManager: Lost task 3.0 in stage 4.0 (TID 27, localhost): org.apache.spark.api.python.PythonException: Traceback (most recent call last):
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/worker.py", line 79, in main
- serializer.dump_stream(func(split_index, iterator), outfile)
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1990, in pipeline_func
- return func(split, prev_func(split, iterator))
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1990, in pipeline_func
- return func(split, prev_func(split, iterator))
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1990, in pipeline_func
- return func(split, prev_func(split, iterator))
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1990, in pipeline_func
- return func(split, prev_func(split, iterator))
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 352, in func
- return f(iterator)
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1585, in _mergeCombiners
- merger.mergeCombiners(iterator)
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/shuffle.py", line 282, in mergeCombiners
- d[k] = comb(d[k], v) if k in d else v
- File "/home/cc/cs61c/fa14/class/cs61c-ef/proj2/SlidingBfsSpark.py", line 10, in bfs_reduce
- if value1[1] < value2[1]:
- TypeError: 'int' object has no attribute '__getitem__'
- org.apache.spark.api.python.PythonRDD$$anon$1.read(PythonRDD.scala:124)
- org.apache.spark.api.python.PythonRDD$$anon$1.<init>(PythonRDD.scala:154)
- org.apache.spark.api.python.PythonRDD.compute(PythonRDD.scala:87)
- org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
- org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
- org.apache.spark.api.python.PairwiseRDD.compute(PythonRDD.scala:265)
- org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
- org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
- org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
- org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
- org.apache.spark.scheduler.Task.run(Task.scala:54)
- org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177)
- java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
- java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
- java.lang.Thread.run(Thread.java:745)
- 14/10/25 16:35:36 ERROR TaskSetManager: Task 3 in stage 4.0 failed 1 times; aborting job
- 14/10/25 16:35:36 INFO TaskSchedulerImpl: Removed TaskSet 4.0, whose tasks have all completed, from pool
- 14/10/25 16:35:36 INFO TaskSchedulerImpl: Cancelling stage 4
- 14/10/25 16:35:36 INFO DAGScheduler: Failed to run collect at /home/cc/cs61c/fa14/class/cs61c-ef/proj2/SlidingBfsSpark.py:53
- Traceback (most recent call last):
- File "/home/cc/cs61c/fa14/class/cs61c-ef/proj2/SlidingBfsSpark.py", line 95, in <module>
- main()
- File "/home/cc/cs61c/fa14/class/cs61c-ef/proj2/SlidingBfsSpark.py", line 88, in main
- solve_sliding_puzzle(args.master, writer, args.height, args.width)
- File "/home/cc/cs61c/fa14/class/cs61c-ef/proj2/SlidingBfsSpark.py", line 53, in solve_sliding_puzzle
- print(gameBoard.collect())
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 723, in collect
- bytesInJava = self._jrdd.collect().iterator()
- File "/home/ff/cs61c/spark-1.1.0/python/lib/py4j-0.8.2.1-src.zip/py4j/java_gateway.py", line 538, in __call__
- File "/home/ff/cs61c/spark-1.1.0/python/lib/py4j-0.8.2.1-src.zip/py4j/protocol.py", line 300, in get_return_value
- py4j.protocol.Py4JJavaError: An error occurred while calling o133.collect.
- : org.apache.spark.SparkException: Job aborted due to stage failure: Task 3 in stage 4.0 failed 1 times, most recent failure: Lost task 3.0 in stage 4.0 (TID 27, localhost): org.apache.spark.api.python.PythonException: Traceback (most recent call last):
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/worker.py", line 79, in main
- serializer.dump_stream(func(split_index, iterator), outfile)
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1990, in pipeline_func
- return func(split, prev_func(split, iterator))
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1990, in pipeline_func
- return func(split, prev_func(split, iterator))
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1990, in pipeline_func
- return func(split, prev_func(split, iterator))
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1990, in pipeline_func
- return func(split, prev_func(split, iterator))
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 352, in func
- return f(iterator)
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/rdd.py", line 1585, in _mergeCombiners
- merger.mergeCombiners(iterator)
- File "/home/ff/cs61c/spark-1.1.0/python/pyspark/shuffle.py", line 282, in mergeCombiners
- d[k] = comb(d[k], v) if k in d else v
- File "/home/cc/cs61c/fa14/class/cs61c-ef/proj2/SlidingBfsSpark.py", line 10, in bfs_reduce
- if value1[1] < value2[1]:
- TypeError: 'int' object has no attribute '__getitem__'
- org.apache.spark.api.python.PythonRDD$$anon$1.read(PythonRDD.scala:124)
- org.apache.spark.api.python.PythonRDD$$anon$1.<init>(PythonRDD.scala:154)
- org.apache.spark.api.python.PythonRDD.compute(PythonRDD.scala:87)
- org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
- org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
- org.apache.spark.api.python.PairwiseRDD.compute(PythonRDD.scala:265)
- org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
- org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
- org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
- org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
- org.apache.spark.scheduler.Task.run(Task.scala:54)
- org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177)
- java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
- java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
- java.lang.Thread.run(Thread.java:745)
- Driver stacktrace:
- at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1185)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1174)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1173)
- at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
- at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
- at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1173)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:688)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:688)
- at scala.Option.foreach(Option.scala:236)
- at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:688)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1391)
- at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)
- at akka.actor.ActorCell.invoke(ActorCell.scala:456)
- at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)
- at akka.dispatch.Mailbox.run(Mailbox.scala:219)
- at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
- at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
- at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
- at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
- at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
- Command exited with non-zero status 1
- 7.28user 0.46system 0:05.20elapsed 148%CPU (0avgtext+0avgdata 917168maxresident)k
- 0inputs+264outputs (0major+96509minor)pagefaults 0swaps
- gmake: *** [run-small] Error 1
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement