Advertisement
Guest User

Untitled

a guest
May 27th, 2015
247
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 18.00 KB | None | 0 0
  1. INFO : WriteAheadLogManager for ReceivedBlockHandlerMaster - Attempting to clear 0 old log files in hdfs://spark-akhil-master:9000/checkpointed/receivedBlockMetadata older than 1429262834000:
  2. INFO : WriteAheadLogManager for ReceivedBlockHandlerMaster - Cleared log files in hdfs://spark-akhil-master:9000/checkpointed/receivedBlockMetadata older than 1429262834000
  3. [Stage 10:> (0 + 2) / 2]INFO : WriteAheadLogManager for ReceivedBlockHandlerMaster - Attempting to clear 0 old log files in hdfs://spark-akhil-master:9000/checkpointed/receivedBlockMetadata older than 1429262974000:
  4. INFO : WriteAheadLogManager for ReceivedBlockHandlerMaster - Cleared log files in hdfs://spark-akhil-master:9000/checkpointed/receivedBlockMetadata older than 1429262974000
  5. [error] (run-main-0) org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 981.0 failed 4 times, most recent failure: Lost task 0.3 in stage 981.0 (TID 1330, spark-akhil-slave1.c.neat-axis-616.internal): java.io.FileNotFoundException: File does not exist: /checkpointed/receivedData/0/log-1429262779712-1429262839712
  6. [error] at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:66)
  7. [error] at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:56)
  8. [error] at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsUpdateTimes(FSNamesystem.java:1891)
  9. [error] at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1832)
  10. [error] at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1812)
  11. [error] at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1784)
  12. [error] at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:542)
  13. [error] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:362)
  14. [error] at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
  15. [error] at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
  16. [error] at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962)
  17. [error] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039)
  18. [error] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035)
  19. [error] at java.security.AccessController.doPrivileged(Native Method)
  20. [error] at javax.security.auth.Subject.doAs(Subject.java:415)
  21. [error] at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
  22. [error] at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033)
  23. [error]
  24. [error] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
  25. [error] at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
  26. [error] at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
  27. [error] at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
  28. [error] at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
  29. [error] at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)
  30. [error] at org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:1144)
  31. [error] at org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:1132)
  32. [error] at org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:1122)
  33. [error] at org.apache.hadoop.hdfs.DFSInputStream.fetchLocatedBlocksAndGetLastBlockLength(DFSInputStream.java:264)
  34. [error] at org.apache.hadoop.hdfs.DFSInputStream.openInfo(DFSInputStream.java:231)
  35. [error] at org.apache.hadoop.hdfs.DFSInputStream.<init>(DFSInputStream.java:224)
  36. [error] at org.apache.hadoop.hdfs.DFSClient.open(DFSClient.java:1295)
  37. [error] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:300)
  38. [error] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:296)
  39. [error] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
  40. [error] at org.apache.hadoop.hdfs.DistributedFileSystem.open(DistributedFileSystem.java:296)
  41. [error] at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:764)
  42. [error] at org.apache.spark.streaming.util.HdfsUtils$.getInputStream(HdfsUtils.scala:45)
  43. [error] at org.apache.spark.streaming.util.WriteAheadLogRandomReader.<init>(WriteAheadLogRandomReader.scala:32)
  44. [error] at org.apache.spark.streaming.rdd.WriteAheadLogBackedBlockRDD.compute(WriteAheadLogBackedBlockRDD.scala:99)
  45. [error] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
  46. [error] at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
  47. [error] at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
  48. [error] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
  49. [error] at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
  50. [error] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
  51. [error] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
  52. [error] at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
  53. [error] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
  54. [error] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
  55. [error] at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
  56. [error] at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
  57. [error] at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
  58. [error] at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
  59. [error] at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
  60. [error] at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  61. [error] at org.apache.spark.scheduler.Task.run(Task.scala:64)
  62. [error] at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203)
  63. [error] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  64. [error] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
  65. [error] at java.lang.Thread.run(Thread.java:745)
  66. [error] Caused by: org.apache.hadoop.ipc.RemoteException(java.io.FileNotFoundException): File does not exist: /checkpointed/receivedData/0/log-1429262779712-1429262839712
  67. [error] at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:66)
  68. [error] at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:56)
  69. [error] at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsUpdateTimes(FSNamesystem.java:1891)
  70. [error] at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1832)
  71. [error] at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1812)
  72. [error] at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1784)
  73. [error] at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:542)
  74. [error] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:362)
  75. [error] at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
  76. [error] at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
  77. [error] at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962)
  78. [error] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039)
  79. [error] at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035)
  80. [error] at java.security.AccessController.doPrivileged(Native Method)
  81. [error] at javax.security.auth.Subject.doAs(Subject.java:415)
  82. [error] at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
  83. [error] at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033)
  84. [error]
  85. [error] at org.apache.hadoop.ipc.Client.call(Client.java:1410)
  86. [error] at org.apache.hadoop.ipc.Client.call(Client.java:1363)
  87. [error] at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
  88. [error] at com.sun.proxy.$Proxy15.getBlockLocations(Unknown Source)
  89. [error] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  90. [error] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
  91. [error] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  92. [error] at java.lang.reflect.Method.invoke(Method.java:606)
  93. [error] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:190)
  94. [error] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:103)
  95. [error] at com.sun.proxy.$Proxy15.getBlockLocations(Unknown Source)
  96. [error] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getBlockLocations(ClientNamenodeProtocolTranslatorPB.java:219)
  97. [error] at org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:1142)
  98. [error] ... 35 more
  99. [error]
  100. [error] Driver stacktrace:
  101. org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 981.0 failed 4 times, most recent failure: Lost task 0.3 in stage 981.0 (TID 1330, spark-akhil-slave1.c.neat-axis-616.internal): java.io.FileNotFoundException: File does not exist: /checkpointed/receivedData/0/log-1429262779712-1429262839712
  102. at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:66)
  103. at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:56)
  104. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsUpdateTimes(FSNamesystem.java:1891)
  105. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1832)
  106. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1812)
  107. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1784)
  108. at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:542)
  109. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:362)
  110. at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
  111. at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
  112. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962)
  113. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039)
  114. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035)
  115. at java.security.AccessController.doPrivileged(Native Method)
  116. at javax.security.auth.Subject.doAs(Subject.java:415)
  117. at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
  118. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033)
  119.  
  120. at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
  121. at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
  122. at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
  123. at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
  124. at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
  125. at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)
  126. at org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:1144)
  127. at org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:1132)
  128. at org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:1122)
  129. at org.apache.hadoop.hdfs.DFSInputStream.fetchLocatedBlocksAndGetLastBlockLength(DFSInputStream.java:264)
  130. at org.apache.hadoop.hdfs.DFSInputStream.openInfo(DFSInputStream.java:231)
  131. at org.apache.hadoop.hdfs.DFSInputStream.<init>(DFSInputStream.java:224)
  132. at org.apache.hadoop.hdfs.DFSClient.open(DFSClient.java:1295)
  133. at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:300)
  134. at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:296)
  135. at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
  136. at org.apache.hadoop.hdfs.DistributedFileSystem.open(DistributedFileSystem.java:296)
  137. at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:764)
  138. at org.apache.spark.streaming.util.HdfsUtils$.getInputStream(HdfsUtils.scala:45)
  139. at org.apache.spark.streaming.util.WriteAheadLogRandomReader.<init>(WriteAheadLogRandomReader.scala:32)
  140. at org.apache.spark.streaming.rdd.WriteAheadLogBackedBlockRDD.compute(WriteAheadLogBackedBlockRDD.scala:99)
  141. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
  142. at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
  143. at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:87)
  144. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
  145. at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
  146. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
  147. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
  148. at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
  149. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
  150. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
  151. at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
  152. at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
  153. at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
  154. at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
  155. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
  156. at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  157. at org.apache.spark.scheduler.Task.run(Task.scala:64)
  158. at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203)
  159. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
  160. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
  161. at java.lang.Thread.run(Thread.java:745)
  162. Caused by: org.apache.hadoop.ipc.RemoteException(java.io.FileNotFoundException): File does not exist: /checkpointed/receivedData/0/log-1429262779712-1429262839712
  163. at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:66)
  164. at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:56)
  165. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsUpdateTimes(FSNamesystem.java:1891)
  166. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1832)
  167. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1812)
  168. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1784)
  169. at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:542)
  170. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:362)
  171. at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
  172. at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
  173. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962)
  174. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039)
  175. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035)
  176. at java.security.AccessController.doPrivileged(Native Method)
  177. at javax.security.auth.Subject.doAs(Subject.java:415)
  178. at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
  179. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033)
  180.  
  181. at org.apache.hadoop.ipc.Client.call(Client.java:1410)
  182. at org.apache.hadoop.ipc.Client.call(Client.java:1363)
  183. at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
  184. at com.sun.proxy.$Proxy15.getBlockLocations(Unknown Source)
  185. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  186. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
  187. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  188. at java.lang.reflect.Method.invoke(Method.java:606)
  189. at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:190)
  190. at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:103)
  191. at com.sun.proxy.$Proxy15.getBlockLocations(Unknown Source)
  192. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getBlockLocations(ClientNamenodeProtocolTranslatorPB.java:219)
  193. at org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:1142)
  194. ... 35 more
  195.  
  196. Driver stacktrace:
  197. at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1203)
  198. at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1192)
  199. at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1191)
  200. at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
  201. at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
  202. at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1191)
  203. at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:693)
  204. at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:693)
  205. at scala.Option.foreach(Option.scala:236)
  206. at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:693)
  207. at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1393)
  208. at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1354)
  209. at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
  210. [trace] Stack trace suppressed: run last compile:run for the full output.
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement