Advertisement
Guest User

Namenode log

a guest
Feb 26th, 2014
90
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 9.76 KB | None | 0 0
  1. 2014-02-21 13:29:44,710 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* NameSystem.allocateBlock: /hbase/.logs/Tapi7,60020,1392967864961/Tapi7%2C60020%2C1392967864961.1393007384753. blk_-6695300470410774365_837638
  2. .
  3. .
  4. .
  5.  
  6.  
  7. 2014-02-21 13:38:25,730 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* blk_-6695300470410774365_837638 recovery started, primary=10.0.0.151:50010
  8. 2014-02-21 13:38:25,730 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 54310, call recoverLease(/hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753, DFSClient_hb_rs_Tapi6,60020,1392967864513) from 10.0.0.108:57767: error: org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to create file /hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753 for DFSClient_hb_rs_Tapi6,60020,1392967864513 on client 10.0.0.108, because this file is already being created by NN_Recovery on 10.0.0.151
  9. org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to create file /hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753 for DFSClient_hb_rs_Tapi6,60020,1392967864513 on client 10.0.0.108, because this file is already being created by NN_Recovery on 10.0.0.151
  10. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLeaseInternal(FSNamesystem.java:1202)
  11. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLease(FSNamesystem.java:1157)
  12. at org.apache.hadoop.hdfs.server.namenode.NameNode.recoverLease(NameNode.java:404)
  13. at sun.reflect.GeneratedMethodAccessor28.invoke(Unknown Source)
  14. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
  15. at java.lang.reflect.Method.invoke(Method.java:597)
  16. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
  17. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:961)
  18. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:957)
  19. at java.security.AccessController.doPrivileged(Native Method)
  20. at javax.security.auth.Subject.doAs(Subject.java:396)
  21. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:955)
  22.  
  23.  
  24.  
  25. .
  26. .
  27. .
  28. 2014-02-21 13:39:26,304 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* blk_-6695300470410774365_837638 recovery started, primary=10.0.0.108:50010
  29. 2014-02-21 13:39:26,304 INFO org.apache.hadoop.ipc.Server: IPC Server handler 5 on 54310, call recoverLease(/hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753, DFSClient_hb_rs_Tapi4,60020,1392967864339) from 10.0.0.96:52778: error: org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to create file /hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753 for DFSClient_hb_rs_Tapi4,60020,1392967864339 on client 10.0.0.96, because this file is already being created by NN_Recovery on 10.0.0.151
  30. org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to create file /hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753 for DFSClient_hb_rs_Tapi4,60020,1392967864339 on client 10.0.0.96, because this file is already being created by NN_Recovery on 10.0.0.151
  31. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLeaseInternal(FSNamesystem.java:1202)
  32. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLease(FSNamesystem.java:1157)
  33. at org.apache.hadoop.hdfs.server.namenode.NameNode.recoverLease(NameNode.java:404)
  34. at sun.reflect.GeneratedMethodAccessor28.invoke(Unknown Source)
  35. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
  36. at java.lang.reflect.Method.invoke(Method.java:597)
  37. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
  38. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:961)
  39. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:957)
  40. at java.security.AccessController.doPrivileged(Native Method)
  41. at javax.security.auth.Subject.doAs(Subject.java:396)
  42. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:955)
  43.  
  44.  
  45. .
  46. .
  47. .
  48. 2014-02-21 13:40:26,335 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* blk_-6695300470410774365_837638 recovery started, primary=10.0.0.96:50010
  49. 2014-02-21 13:40:26,335 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 54310, call recoverLease(/hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753, DFSClient_hb_rs_Tapi4,60020,1392967864339) from 10.0.0.96:52778: error: org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to create file /hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753 for DFSClient_hb_rs_Tapi4,60020,1392967864339 on client 10.0.0.96, because this file is already being created by NN_Recovery on 10.0.0.151
  50. org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to create file /hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753 for DFSClient_hb_rs_Tapi4,60020,1392967864339 on client 10.0.0.96, because this file is already being created by NN_Recovery on 10.0.0.151
  51. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLeaseInternal(FSNamesystem.java:1202)
  52. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLease(FSNamesystem.java:1157)
  53. at org.apache.hadoop.hdfs.server.namenode.NameNode.recoverLease(NameNode.java:404)
  54. at sun.reflect.GeneratedMethodAccessor28.invoke(Unknown Source)
  55. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
  56. at java.lang.reflect.Method.invoke(Method.java:597)
  57. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
  58. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:961)
  59. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:957)
  60. at java.security.AccessController.doPrivileged(Native Method)
  61. at javax.security.auth.Subject.doAs(Subject.java:396)
  62. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:955)
  63.  
  64. .
  65. .
  66. .
  67.  
  68. 2014-02-21 13:41:00,206 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* NameSystem.addStoredBlock: blockMap updated: 10.0.0.151:50010 is added to blk_-6695300470410774365_837638 size 50386058
  69. 2014-02-21 13:41:00,207 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* NameSystem.addStoredBlock: blockMap updated: 10.0.0.108:50010 is added to blk_-6695300470410774365_837638 size 50386058
  70. 2014-02-21 13:41:00,209 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: commitBlockSynchronization(lastblock=blk_-6695300470410774365_837638, newgenerationstamp=837802, newlength=50386058, newtargets=[10.0.0.151:50010, 10.0.0.108:50010, 10.0.0.96:50010], closeFile=true, deleteBlock=false)
  71.  
  72. .
  73. .
  74. .
  75.  
  76. 2014-02-21 13:41:34,711 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: commitBlockSynchronization(lastblock=blk_-6695300470410774365_837638, newgenerationstamp=0, newlength=0, newtargets=[], closeFile=false, deleteBlock=true)
  77. 2014-02-21 13:41:34,712 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 54310, call commitBlockSynchronization(blk_-6695300470410774365_837638, 0, 0, false, true, [Lorg.apache.hadoop.hdfs.protocol.DatanodeID;@12103932) from 10.0.0.108:43309: error: java.io.IOException: Block (=blk_-6695300470410774365_837638) not found
  78. java.io.IOException: Block (=blk_-6695300470410774365_837638) not found
  79. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.commitBlockSynchronization(FSNamesystem.java:2025)
  80. at org.apache.hadoop.hdfs.server.namenode.NameNode.commitBlockSynchronization(NameNode.java:509)
  81. at sun.reflect.GeneratedMethodAccessor30.invoke(Unknown Source)
  82. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
  83. at java.lang.reflect.Method.invoke(Method.java:597)
  84. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
  85. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:961)
  86. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:957)
  87. at java.security.AccessController.doPrivileged(Native Method)
  88. at javax.security.auth.Subject.doAs(Subject.java:396)
  89. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:955)
  90. 2014-02-21 13:41:35,718 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: commitBlockSynchronization(lastblock=blk_-6695300470410774365_837638, newgenerationstamp=0, newlength=0, newtargets=[], closeFile=false, deleteBlock=true)
  91. 2014-02-21 13:41:35,718 INFO org.apache.hadoop.ipc.Server: IPC Server handler 3 on 54310, call commitBlockSynchronization(blk_-6695300470410774365_837638, 0, 0, false, true, [Lorg.apache.hadoop.hdfs.protocol.DatanodeID;@671b0ae1) from 10.0.0.108:43309: error: java.io.IOException: Block (=blk_-6695300470410774365_837638) not found
  92. java.io.IOException: Block (=blk_-6695300470410774365_837638) not found
  93. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.commitBlockSynchronization(FSNamesystem.java:2025)
  94. at org.apache.hadoop.hdfs.server.namenode.NameNode.commitBlockSynchronization(NameNode.java:509)
  95. at sun.reflect.GeneratedMethodAccessor30.invoke(Unknown Source)
  96. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
  97. at java.lang.reflect.Method.invoke(Method.java:597)
  98. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
  99. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:961)
  100. at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:957)
  101. at java.security.AccessController.doPrivileged(Native Method)
  102. at javax.security.auth.Subject.doAs(Subject.java:396)
  103. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:955)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement