Advertisement
aironman

exception hadoop vmware

Apr 27th, 2016
293
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 54.97 KB | None | 0 0
  1. scala> val textFile = sc.textFile("hdfs://192.168.30.157:8020/cervantes/novela/quijote.txt").flatMap(line=>line.split(" ")).map(word=>(word,1)).reduceByKey(_+_).saveAsTextFile("output-quijote")
  2. 16/04/27 18:59:44 WARN BlockReaderFactory: I/O error constructing remote block reader.
  3. java.net.ConnectException: Connection refused
  4. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  5. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  6. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  7. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  8. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  9. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  10. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  11. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  12. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  13. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  14. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  15. at java.io.DataInputStream.read(DataInputStream.java:100)
  16. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  17. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  18. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  19. at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
  20. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
  21. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
  22. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:246)
  23. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
  24. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  25. at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
  26. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  27. at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
  28. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  29. at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:189)
  30. at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64)
  31. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  32. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  33. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  34. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  35. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  36. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  37. at java.lang.Thread.run(Thread.java:745)
  38. 16/04/27 18:59:44 WARN DFSClient: Failed to connect to /127.0.0.1:50010 for block, add to deadNodes and continue. java.net.ConnectException: Connection refused
  39. java.net.ConnectException: Connection refused
  40. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  41. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  42. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  43. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  44. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  45. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  46. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  47. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  48. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  49. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  50. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  51. at java.io.DataInputStream.read(DataInputStream.java:100)
  52. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  53. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  54. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  55. at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
  56. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
  57. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
  58. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:246)
  59. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
  60. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  61. at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
  62. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  63. at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
  64. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  65. at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:189)
  66. at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64)
  67. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  68. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  69. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  70. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  71. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  72. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  73. at java.lang.Thread.run(Thread.java:745)
  74. 16/04/27 18:59:44 WARN DFSClient: DFS chooseDataNode: got # 1 IOException, will wait for 1306.5790151229824 msec.
  75. 16/04/27 18:59:44 WARN BlockReaderFactory: I/O error constructing remote block reader.
  76. java.net.ConnectException: Connection refused
  77. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  78. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  79. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  80. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  81. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  82. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  83. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  84. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  85. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  86. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  87. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  88. at java.io.DataInputStream.read(DataInputStream.java:100)
  89. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  90. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  91. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  92. at org.apache.hadoop.mapred.LineRecordReader.<init>(LineRecordReader.java:134)
  93. at org.apache.hadoop.mapred.TextInputFormat.getRecordReader(TextInputFormat.java:67)
  94. at org.apache.spark.rdd.HadoopRDD$$anon$1.<init>(HadoopRDD.scala:237)
  95. at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:208)
  96. at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:101)
  97. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  98. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  99. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  100. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  101. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  102. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  103. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  104. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  105. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  106. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  107. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  108. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  109. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  110. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  111. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  112. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  113. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  114. at java.lang.Thread.run(Thread.java:745)
  115. 16/04/27 18:59:44 WARN DFSClient: Failed to connect to /127.0.0.1:50010 for block, add to deadNodes and continue. java.net.ConnectException: Connection refused
  116. java.net.ConnectException: Connection refused
  117. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  118. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  119. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  120. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  121. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  122. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  123. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  124. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  125. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  126. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  127. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  128. at java.io.DataInputStream.read(DataInputStream.java:100)
  129. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  130. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  131. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  132. at org.apache.hadoop.mapred.LineRecordReader.<init>(LineRecordReader.java:134)
  133. at org.apache.hadoop.mapred.TextInputFormat.getRecordReader(TextInputFormat.java:67)
  134. at org.apache.spark.rdd.HadoopRDD$$anon$1.<init>(HadoopRDD.scala:237)
  135. at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:208)
  136. at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:101)
  137. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  138. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  139. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  140. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  141. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  142. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  143. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  144. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  145. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  146. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  147. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  148. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  149. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  150. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  151. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  152. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  153. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  154. at java.lang.Thread.run(Thread.java:745)
  155. 16/04/27 18:59:44 WARN DFSClient: DFS chooseDataNode: got # 1 IOException, will wait for 1316.8481425690604 msec.
  156. [Stage 6:> (0 + 2) / 2]16/04/27 18:59:45 WARN BlockReaderFactory: I/O error constructing remote block reader.
  157. java.net.ConnectException: Connection refused
  158. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  159. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  160. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  161. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  162. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  163. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  164. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  165. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  166. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  167. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  168. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  169. at java.io.DataInputStream.read(DataInputStream.java:100)
  170. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  171. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  172. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  173. at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
  174. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
  175. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
  176. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:246)
  177. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
  178. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  179. at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
  180. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  181. at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
  182. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  183. at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:189)
  184. at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64)
  185. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  186. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  187. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  188. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  189. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  190. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  191. at java.lang.Thread.run(Thread.java:745)
  192. 16/04/27 18:59:45 WARN DFSClient: Failed to connect to /127.0.0.1:50010 for block, add to deadNodes and continue. java.net.ConnectException: Connection refused
  193. java.net.ConnectException: Connection refused
  194. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  195. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  196. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  197. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  198. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  199. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  200. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  201. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  202. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  203. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  204. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  205. at java.io.DataInputStream.read(DataInputStream.java:100)
  206. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  207. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  208. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  209. at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
  210. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
  211. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
  212. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:246)
  213. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
  214. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  215. at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
  216. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  217. at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
  218. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  219. at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:189)
  220. at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64)
  221. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  222. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  223. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  224. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  225. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  226. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  227. at java.lang.Thread.run(Thread.java:745)
  228. 16/04/27 18:59:45 WARN DFSClient: DFS chooseDataNode: got # 2 IOException, will wait for 5344.531746769184 msec.
  229. 16/04/27 18:59:45 WARN BlockReaderFactory: I/O error constructing remote block reader.
  230. java.net.ConnectException: Connection refused
  231. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  232. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  233. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  234. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  235. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  236. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  237. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  238. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  239. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  240. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  241. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  242. at java.io.DataInputStream.read(DataInputStream.java:100)
  243. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  244. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  245. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  246. at org.apache.hadoop.mapred.LineRecordReader.<init>(LineRecordReader.java:134)
  247. at org.apache.hadoop.mapred.TextInputFormat.getRecordReader(TextInputFormat.java:67)
  248. at org.apache.spark.rdd.HadoopRDD$$anon$1.<init>(HadoopRDD.scala:237)
  249. at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:208)
  250. at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:101)
  251. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  252. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  253. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  254. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  255. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  256. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  257. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  258. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  259. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  260. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  261. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  262. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  263. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  264. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  265. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  266. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  267. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  268. at java.lang.Thread.run(Thread.java:745)
  269. 16/04/27 18:59:45 WARN DFSClient: Failed to connect to /127.0.0.1:50010 for block, add to deadNodes and continue. java.net.ConnectException: Connection refused
  270. java.net.ConnectException: Connection refused
  271. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  272. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  273. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  274. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  275. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  276. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  277. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  278. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  279. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  280. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  281. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  282. at java.io.DataInputStream.read(DataInputStream.java:100)
  283. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  284. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  285. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  286. at org.apache.hadoop.mapred.LineRecordReader.<init>(LineRecordReader.java:134)
  287. at org.apache.hadoop.mapred.TextInputFormat.getRecordReader(TextInputFormat.java:67)
  288. at org.apache.spark.rdd.HadoopRDD$$anon$1.<init>(HadoopRDD.scala:237)
  289. at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:208)
  290. at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:101)
  291. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  292. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  293. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  294. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  295. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  296. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  297. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  298. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  299. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  300. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  301. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  302. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  303. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  304. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  305. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  306. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  307. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  308. at java.lang.Thread.run(Thread.java:745)
  309. 16/04/27 18:59:45 WARN DFSClient: DFS chooseDataNode: got # 2 IOException, will wait for 3568.022412691614 msec.
  310. 16/04/27 18:59:49 WARN BlockReaderFactory: I/O error constructing remote block reader.
  311. java.net.ConnectException: Connection refused
  312. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  313. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  314. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  315. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  316. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  317. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  318. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  319. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  320. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  321. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  322. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  323. at java.io.DataInputStream.read(DataInputStream.java:100)
  324. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  325. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  326. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  327. at org.apache.hadoop.mapred.LineRecordReader.<init>(LineRecordReader.java:134)
  328. at org.apache.hadoop.mapred.TextInputFormat.getRecordReader(TextInputFormat.java:67)
  329. at org.apache.spark.rdd.HadoopRDD$$anon$1.<init>(HadoopRDD.scala:237)
  330. at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:208)
  331. at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:101)
  332. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  333. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  334. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  335. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  336. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  337. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  338. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  339. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  340. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  341. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  342. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  343. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  344. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  345. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  346. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  347. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  348. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  349. at java.lang.Thread.run(Thread.java:745)
  350. 16/04/27 18:59:49 WARN DFSClient: Failed to connect to /127.0.0.1:50010 for block, add to deadNodes and continue. java.net.ConnectException: Connection refused
  351. java.net.ConnectException: Connection refused
  352. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  353. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  354. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  355. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  356. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  357. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  358. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  359. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  360. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  361. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  362. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  363. at java.io.DataInputStream.read(DataInputStream.java:100)
  364. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  365. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  366. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  367. at org.apache.hadoop.mapred.LineRecordReader.<init>(LineRecordReader.java:134)
  368. at org.apache.hadoop.mapred.TextInputFormat.getRecordReader(TextInputFormat.java:67)
  369. at org.apache.spark.rdd.HadoopRDD$$anon$1.<init>(HadoopRDD.scala:237)
  370. at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:208)
  371. at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:101)
  372. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  373. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  374. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  375. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  376. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  377. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  378. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  379. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  380. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  381. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
  382. at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
  383. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  384. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  385. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  386. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  387. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  388. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  389. at java.lang.Thread.run(Thread.java:745)
  390. 16/04/27 18:59:49 WARN DFSClient: DFS chooseDataNode: got # 3 IOException, will wait for 13606.690016648003 msec.
  391. 16/04/27 18:59:51 WARN BlockReaderFactory: I/O error constructing remote block reader.
  392. java.net.ConnectException: Connection refused
  393. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  394. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  395. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  396. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  397. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  398. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  399. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  400. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  401. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  402. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  403. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  404. at java.io.DataInputStream.read(DataInputStream.java:100)
  405. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  406. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  407. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  408. at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
  409. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
  410. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
  411. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:246)
  412. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
  413. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  414. at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
  415. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  416. at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
  417. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  418. at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:189)
  419. at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64)
  420. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  421. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  422. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  423. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  424. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  425. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  426. at java.lang.Thread.run(Thread.java:745)
  427. 16/04/27 18:59:51 WARN DFSClient: Failed to connect to /127.0.0.1:50010 for block, add to deadNodes and continue. java.net.ConnectException: Connection refused
  428. java.net.ConnectException: Connection refused
  429. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  430. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  431. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  432. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  433. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  434. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  435. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  436. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  437. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  438. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  439. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  440. at java.io.DataInputStream.read(DataInputStream.java:100)
  441. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  442. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  443. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  444. at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
  445. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
  446. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
  447. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:246)
  448. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
  449. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  450. at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
  451. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  452. at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
  453. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  454. at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:189)
  455. at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64)
  456. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  457. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  458. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  459. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  460. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  461. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  462. at java.lang.Thread.run(Thread.java:745)
  463. 16/04/27 18:59:51 WARN DFSClient: DFS chooseDataNode: got # 3 IOException, will wait for 9560.09839904081 msec.
  464. 16/04/27 19:00:00 WARN BlockReaderFactory: I/O error constructing remote block reader.
  465. java.net.ConnectException: Connection refused
  466. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  467. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  468. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  469. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  470. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  471. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  472. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  473. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  474. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  475. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  476. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  477. at java.io.DataInputStream.read(DataInputStream.java:100)
  478. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  479. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  480. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  481. at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
  482. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
  483. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
  484. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:246)
  485. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
  486. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  487. at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
  488. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  489. at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
  490. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  491. at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:189)
  492. at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64)
  493. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  494. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  495. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  496. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  497. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  498. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  499. at java.lang.Thread.run(Thread.java:745)
  500. 16/04/27 19:00:00 WARN DFSClient: Failed to connect to /127.0.0.1:50010 for block, add to deadNodes and continue. java.net.ConnectException: Connection refused
  501. java.net.ConnectException: Connection refused
  502. at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
  503. at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
  504. at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
  505. at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
  506. at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
  507. at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
  508. at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
  509. at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
  510. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
  511. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  512. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  513. at java.io.DataInputStream.read(DataInputStream.java:100)
  514. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  515. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  516. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  517. at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
  518. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
  519. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
  520. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:246)
  521. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
  522. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  523. at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
  524. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  525. at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
  526. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  527. at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:189)
  528. at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64)
  529. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  530. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  531. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  532. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  533. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  534. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  535. at java.lang.Thread.run(Thread.java:745)
  536. 16/04/27 19:00:00 WARN DFSClient: Could not obtain block: BP-286282631-127.0.0.1-1433865208026:blk_1073743637_2859 file=/cervantes/novela/quijote.txt No live nodes contain current block Block locations: 127.0.0.1:50010 Dead nodes: 127.0.0.1:50010. Throwing a BlockMissingException
  537. 16/04/27 19:00:00 WARN DFSClient: Could not obtain block: BP-286282631-127.0.0.1-1433865208026:blk_1073743637_2859 file=/cervantes/novela/quijote.txt No live nodes contain current block Block locations: 127.0.0.1:50010 Dead nodes: 127.0.0.1:50010. Throwing a BlockMissingException
  538. 16/04/27 19:00:00 WARN DFSClient: DFS Read
  539. org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block: BP-286282631-127.0.0.1-1433865208026:blk_1073743637_2859 file=/cervantes/novela/quijote.txt
  540. at org.apache.hadoop.hdfs.DFSInputStream.chooseDataNode(DFSInputStream.java:888)
  541. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:568)
  542. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  543. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  544. at java.io.DataInputStream.read(DataInputStream.java:100)
  545. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  546. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  547. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  548. at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
  549. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
  550. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
  551. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:246)
  552. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
  553. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  554. at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
  555. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  556. at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
  557. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  558. at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:189)
  559. at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64)
  560. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  561. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  562. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  563. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  564. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  565. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  566. at java.lang.Thread.run(Thread.java:745)
  567. 16/04/27 19:00:00 ERROR Executor: Exception in task 0.0 in stage 6.0 (TID 15)
  568. org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block: BP-286282631-127.0.0.1-1433865208026:blk_1073743637_2859 file=/cervantes/novela/quijote.txt
  569. at org.apache.hadoop.hdfs.DFSInputStream.chooseDataNode(DFSInputStream.java:888)
  570. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:568)
  571. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  572. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  573. at java.io.DataInputStream.read(DataInputStream.java:100)
  574. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  575. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  576. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  577. at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
  578. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
  579. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
  580. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:246)
  581. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
  582. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  583. at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
  584. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  585. at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
  586. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  587. at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:189)
  588. at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64)
  589. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  590. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  591. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  592. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  593. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  594. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  595. at java.lang.Thread.run(Thread.java:745)
  596. 16/04/27 19:00:00 WARN TaskSetManager: Lost task 0.0 in stage 6.0 (TID 15, localhost): org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block: BP-286282631-127.0.0.1-1433865208026:blk_1073743637_2859 file=/cervantes/novela/quijote.txt
  597. at org.apache.hadoop.hdfs.DFSInputStream.chooseDataNode(DFSInputStream.java:888)
  598. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:568)
  599. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  600. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  601. at java.io.DataInputStream.read(DataInputStream.java:100)
  602. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  603. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  604. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  605. at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
  606. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
  607. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
  608. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:246)
  609. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
  610. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  611. at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
  612. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  613. at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
  614. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  615. at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:189)
  616. at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64)
  617. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  618. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  619. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  620. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  621. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  622. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  623. at java.lang.Thread.run(Thread.java:745)
  624.  
  625. 16/04/27 19:00:00 ERROR TaskSetManager: Task 0 in stage 6.0 failed 1 times; aborting job
  626. org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 6.0 failed 1 times, most recent failure: Lost task 0.0 in stage 6.0 (TID 15, localhost): org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block: BP-286282631-127.0.0.1-1433865208026:blk_1073743637_2859 file=/cervantes/novela/quijote.txt
  627. at org.apache.hadoop.hdfs.DFSInputStream.chooseDataNode(DFSInputStream.java:888)
  628. at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:568)
  629. at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
  630. at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
  631. at java.io.DataInputStream.read(DataInputStream.java:100)
  632. at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
  633. at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
  634. at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
  635. at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
  636. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
  637. at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
  638. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:246)
  639. at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:208)
  640. at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
  641. at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
  642. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  643. at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
  644. at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
  645. at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:189)
  646. at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:64)
  647. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  648. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  649. at org.apache.spark.scheduler.Task.run(Task.scala:89)
  650. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
  651. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  652. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  653. at java.lang.Thread.run(Thread.java:745)
  654.  
  655. Driver stacktrace:
  656. at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
  657. at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
  658. at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
  659. at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
  660. at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
  661. at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
  662. at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
  663. at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
  664. at scala.Option.foreach(Option.scala:236)
  665. at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
  666. at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640)
  667. at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599)
  668. at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588)
  669. at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
  670. at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)
  671. at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
  672. at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
  673. at org.apache.spark.SparkContext.runJob(SparkContext.scala:1922)
  674. at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1213)
  675. at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopDataset$1.apply(PairRDDFunctions.scala:1156)
  676. at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopDataset$1.apply(PairRDDFunctions.scala:1156)
  677. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
  678. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
  679. at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
  680. at org.apache.spark.rdd.PairRDDFunctions.saveAsHadoopDataset(PairRDDFunctions.scala:1156)
  681. at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$4.apply$mcV$sp(PairRDDFunctions.scala:1060)
  682. at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$4.apply(PairRDDFunctions.scala:1026)
  683. at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$4.apply(PairRDDFunctions.scala:1026)
  684. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
  685. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
  686. at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
  687. at org.apache.spark.rdd.PairRDDFunctions.saveAsHadoopFile(PairRDDFunctions.scala:1026)
  688. at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$1.apply$mcV$sp(PairRDDFunctions.scala:952)
  689. at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$1.apply(PairRDDFunctions.scala:952)
  690. at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsHadoopFile$1.apply(PairRDDFunctions.scala:952)
  691. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
  692. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
  693. at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
  694. at org.apache.spark.rdd.PairRDDFunctions.saveAsHadoopFile(PairRDDFunctions.scala:951)
  695. at org.apache.spark.rdd.RDD$$anonfun$saveAsTextFile$1.apply$mcV$sp(RDD.scala:1443)
  696. at org.apache.spark.rdd.RDD$$anonfun$saveAsTextFile$1.apply(RDD.scala:1422)
  697. at org.apache.spark.rdd.RDD$$anonfun$saveAsTextFile$1.apply(RDD.scala:1422)
  698. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
  699. at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
  700. at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
  701. at org.apache.spark.rdd.RDD.saveAsTextFile(RDD.scala:1422)
  702. at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:27)
  703. at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:32)
  704. at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:34)
  705. at $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:36)
  706. at $iwC$$iwC$$iwC$$iwC.<init>(<console>:38)
  707. at $iwC$$iwC$$iwC.<init>(<console>:40)
  708. at $iwC$$iwC.<init>(<console>:42)
  709. at $iwC.<init>(<console>:44)
  710. at <init>(<console>:46)
  711. at .<init>(<console>:50)
  712. at .<clinit>(<console>)
  713. at .<init>(<console>:7)
  714. at .<clinit>(<console>)
  715. at $print(<console>)
  716. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  717. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  718. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  719. at java.lang.reflect.Method.invoke(Method.java:498)
  720. at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065)
  721. at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346)
  722. at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840)
  723. at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871)
  724. at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819)
  725. at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857)
  726. at org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902)
  727. at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814)
  728. at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657)
  729. at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665)
  730. at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670)
  731. at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997)
  732. at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement