Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- 2014-02-21 13:29:44,710 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* NameSystem.allocateBlock: /hbase/.logs/Tapi7,60020,1392967864961/Tapi7%2C60020%2C1392967864961.1393007384753. blk_-6695300470410774365_837638
- .
- .
- .
- 2014-02-21 13:38:25,730 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* blk_-6695300470410774365_837638 recovery started, primary=10.0.0.151:50010
- 2014-02-21 13:38:25,730 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 54310, call recoverLease(/hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753, DFSClient_hb_rs_Tapi6,60020,1392967864513) from 10.0.0.108:57767: error: org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to create file /hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753 for DFSClient_hb_rs_Tapi6,60020,1392967864513 on client 10.0.0.108, because this file is already being created by NN_Recovery on 10.0.0.151
- org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to create file /hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753 for DFSClient_hb_rs_Tapi6,60020,1392967864513 on client 10.0.0.108, because this file is already being created by NN_Recovery on 10.0.0.151
- at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLeaseInternal(FSNamesystem.java:1202)
- at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLease(FSNamesystem.java:1157)
- at org.apache.hadoop.hdfs.server.namenode.NameNode.recoverLease(NameNode.java:404)
- at sun.reflect.GeneratedMethodAccessor28.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
- at java.lang.reflect.Method.invoke(Method.java:597)
- at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:961)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:957)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:396)
- at org.apache.hadoop.ipc.Server$Handler.run(Server.java:955)
- .
- .
- .
- 2014-02-21 13:39:26,304 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* blk_-6695300470410774365_837638 recovery started, primary=10.0.0.108:50010
- 2014-02-21 13:39:26,304 INFO org.apache.hadoop.ipc.Server: IPC Server handler 5 on 54310, call recoverLease(/hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753, DFSClient_hb_rs_Tapi4,60020,1392967864339) from 10.0.0.96:52778: error: org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to create file /hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753 for DFSClient_hb_rs_Tapi4,60020,1392967864339 on client 10.0.0.96, because this file is already being created by NN_Recovery on 10.0.0.151
- org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to create file /hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753 for DFSClient_hb_rs_Tapi4,60020,1392967864339 on client 10.0.0.96, because this file is already being created by NN_Recovery on 10.0.0.151
- at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLeaseInternal(FSNamesystem.java:1202)
- at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLease(FSNamesystem.java:1157)
- at org.apache.hadoop.hdfs.server.namenode.NameNode.recoverLease(NameNode.java:404)
- at sun.reflect.GeneratedMethodAccessor28.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
- at java.lang.reflect.Method.invoke(Method.java:597)
- at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:961)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:957)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:396)
- at org.apache.hadoop.ipc.Server$Handler.run(Server.java:955)
- .
- .
- .
- 2014-02-21 13:40:26,335 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* blk_-6695300470410774365_837638 recovery started, primary=10.0.0.96:50010
- 2014-02-21 13:40:26,335 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 54310, call recoverLease(/hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753, DFSClient_hb_rs_Tapi4,60020,1392967864339) from 10.0.0.96:52778: error: org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to create file /hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753 for DFSClient_hb_rs_Tapi4,60020,1392967864339 on client 10.0.0.96, because this file is already being created by NN_Recovery on 10.0.0.151
- org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to create file /hbase/.logs/Tapi7,60020,1392967864961-splitting/Tapi7%2C60020%2C1392967864961.1393007384753 for DFSClient_hb_rs_Tapi4,60020,1392967864339 on client 10.0.0.96, because this file is already being created by NN_Recovery on 10.0.0.151
- at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLeaseInternal(FSNamesystem.java:1202)
- at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLease(FSNamesystem.java:1157)
- at org.apache.hadoop.hdfs.server.namenode.NameNode.recoverLease(NameNode.java:404)
- at sun.reflect.GeneratedMethodAccessor28.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
- at java.lang.reflect.Method.invoke(Method.java:597)
- at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:961)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:957)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:396)
- at org.apache.hadoop.ipc.Server$Handler.run(Server.java:955)
- .
- .
- .
- 2014-02-21 13:41:00,206 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* NameSystem.addStoredBlock: blockMap updated: 10.0.0.151:50010 is added to blk_-6695300470410774365_837638 size 50386058
- 2014-02-21 13:41:00,207 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* NameSystem.addStoredBlock: blockMap updated: 10.0.0.108:50010 is added to blk_-6695300470410774365_837638 size 50386058
- 2014-02-21 13:41:00,209 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: commitBlockSynchronization(lastblock=blk_-6695300470410774365_837638, newgenerationstamp=837802, newlength=50386058, newtargets=[10.0.0.151:50010, 10.0.0.108:50010, 10.0.0.96:50010], closeFile=true, deleteBlock=false)
- .
- .
- .
- 2014-02-21 13:41:34,711 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: commitBlockSynchronization(lastblock=blk_-6695300470410774365_837638, newgenerationstamp=0, newlength=0, newtargets=[], closeFile=false, deleteBlock=true)
- 2014-02-21 13:41:34,712 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 54310, call commitBlockSynchronization(blk_-6695300470410774365_837638, 0, 0, false, true, [Lorg.apache.hadoop.hdfs.protocol.DatanodeID;@12103932) from 10.0.0.108:43309: error: java.io.IOException: Block (=blk_-6695300470410774365_837638) not found
- java.io.IOException: Block (=blk_-6695300470410774365_837638) not found
- at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.commitBlockSynchronization(FSNamesystem.java:2025)
- at org.apache.hadoop.hdfs.server.namenode.NameNode.commitBlockSynchronization(NameNode.java:509)
- at sun.reflect.GeneratedMethodAccessor30.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
- at java.lang.reflect.Method.invoke(Method.java:597)
- at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:961)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:957)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:396)
- at org.apache.hadoop.ipc.Server$Handler.run(Server.java:955)
- 2014-02-21 13:41:35,718 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: commitBlockSynchronization(lastblock=blk_-6695300470410774365_837638, newgenerationstamp=0, newlength=0, newtargets=[], closeFile=false, deleteBlock=true)
- 2014-02-21 13:41:35,718 INFO org.apache.hadoop.ipc.Server: IPC Server handler 3 on 54310, call commitBlockSynchronization(blk_-6695300470410774365_837638, 0, 0, false, true, [Lorg.apache.hadoop.hdfs.protocol.DatanodeID;@671b0ae1) from 10.0.0.108:43309: error: java.io.IOException: Block (=blk_-6695300470410774365_837638) not found
- java.io.IOException: Block (=blk_-6695300470410774365_837638) not found
- at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.commitBlockSynchronization(FSNamesystem.java:2025)
- at org.apache.hadoop.hdfs.server.namenode.NameNode.commitBlockSynchronization(NameNode.java:509)
- at sun.reflect.GeneratedMethodAccessor30.invoke(Unknown Source)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
- at java.lang.reflect.Method.invoke(Method.java:597)
- at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:961)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:957)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:396)
- at org.apache.hadoop.ipc.Server$Handler.run(Server.java:955)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement