2013-04-29 16:30:43,382 WARN mapreduce.Counters: Group org.apache.hadoop.mapred.Task$Counter is deprecated. Use org.apache.hadoop.mapreduce.TaskCounter instead 2013-04-29 16:30:43,752 WARN org.apache.hadoop.conf.Configuration: session.id is deprecated. Instead, use dfs.metrics.session-id 2013-04-29 16:30:43,753 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: Initializing JVM Metrics with processName=MAP, sessionId= 2013-04-29 16:30:43,792 WARN org.apache.hadoop.conf.Configuration: slave.host.name is deprecated. Instead, use dfs.datanode.hostname 2013-04-29 16:30:44,092 INFO org.apache.hadoop.util.ProcessTree: setsid exited with exit code 0 2013-04-29 16:30:44,096 INFO org.apache.hadoop.mapred.Task: Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@10987197 2013-04-29 16:30:44,311 INFO org.apache.hadoop.mapred.MapTask: Processing split: hdfs://namenode-primary:8020/tmp/hbase-root/exportSnapshot-1367253018375/export-files.1367253018377/export-503.seq:0+883 2013-04-29 16:30:44,348 INFO org.apache.hadoop.io.compress.zlib.ZlibFactory: Successfully loaded & initialized native-zlib library 2013-04-29 16:30:44,348 INFO org.apache.hadoop.io.compress.CodecPool: Got brand-new decompressor [.deflate] 2013-04-29 16:30:44,365 INFO org.apache.hadoop.hbase.snapshot.ExportSnapshot: copy file input=d/queries=991625ef6c2a3db259dc984c990e823d-29384f58e6964b1a9044590988a390d3 output=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/991625ef6c2a3db259dc984c990e823d/d/29384f58e6964b1a9044590988a390d3 2013-04-29 16:30:44,374 WARN org.apache.hadoop.hbase.snapshot.ExportSnapshot: Unable to get the status for file=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/991625ef6c2a3db259dc984c990e823d/d/29384f58e6964b1a9044590988a390d3 2013-04-29 16:40:38,059 ERROR org.apache.hadoop.hbase.snapshot.ExportSnapshot: Unable to set the owner/group for file=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/991625ef6c2a3db259dc984c990e823d/d/29384f58e6964b1a9044590988a390d3 org.apache.hadoop.security.AccessControlException: Non-super user cannot change owner. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setOwnerInt(FSNamesystem.java:1180) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setOwner(FSNamesystem.java:1155) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.setOwner(NameNodeRpcServer.java:461) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setOwner(ClientNamenodeProtocolServerSideTranslatorPB.java:267) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44076) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1695) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1691) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1689) at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) at java.lang.reflect.Constructor.newInstance(Constructor.java:513) at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:90) at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:57) at org.apache.hadoop.hdfs.DFSClient.setOwner(DFSClient.java:1905) at org.apache.hadoop.hdfs.DistributedFileSystem.setOwner(DistributedFileSystem.java:831) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.preserveAttributes(ExportSnapshot.java:238) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:205) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:145) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:94) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:140) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:673) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:331) at org.apache.hadoop.mapred.Child$4.run(Child.java:268) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) at org.apache.hadoop.mapred.Child.main(Child.java:262) Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): Non-super user cannot change owner. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setOwnerInt(FSNamesystem.java:1180) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setOwner(FSNamesystem.java:1155) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.setOwner(NameNodeRpcServer.java:461) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setOwner(ClientNamenodeProtocolServerSideTranslatorPB.java:267) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44076) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1695) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1691) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1689) at org.apache.hadoop.ipc.Client.call(Client.java:1225) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202) at $Proxy10.setOwner(Unknown Source) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83) at $Proxy10.setOwner(Unknown Source) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.setOwner(ClientNamenodeProtocolTranslatorPB.java:258) at org.apache.hadoop.hdfs.DFSClient.setOwner(DFSClient.java:1903) ... 13 more 2013-04-29 16:40:38,062 INFO org.apache.hadoop.hbase.snapshot.ExportSnapshot: copy file input=d/queries=5b3e8544cff2f045c6e89f422500cd6e-bedf8536eb484360992dfc47369b70a7 output=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/5b3e8544cff2f045c6e89f422500cd6e/d/bedf8536eb484360992dfc47369b70a7 2013-04-29 16:40:38,068 WARN org.apache.hadoop.hbase.snapshot.ExportSnapshot: Unable to get the status for file=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/5b3e8544cff2f045c6e89f422500cd6e/d/bedf8536eb484360992dfc47369b70a7 2013-04-29 16:40:40,729 INFO org.apache.hadoop.hbase.snapshot.ExportSnapshot: copy completed for input=d/queries=5b3e8544cff2f045c6e89f422500cd6e-bedf8536eb484360992dfc47369b70a7 output=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/5b3e8544cff2f045c6e89f422500cd6e/d/bedf8536eb484360992dfc47369b70a7 2013-04-29 16:40:40,730 INFO org.apache.hadoop.hbase.snapshot.ExportSnapshot: copy file input=d/queries=4689deb6f26e82091aa127587df4777a-f9bc7e65a180459da53712205b991905 output=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/4689deb6f26e82091aa127587df4777a/d/f9bc7e65a180459da53712205b991905 2013-04-29 16:40:40,734 WARN org.apache.hadoop.hbase.snapshot.ExportSnapshot: Unable to get the status for file=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/4689deb6f26e82091aa127587df4777a/d/f9bc7e65a180459da53712205b991905 2013-04-29 16:40:41,766 INFO org.apache.hadoop.hbase.snapshot.ExportSnapshot: copy completed for input=d/queries=4689deb6f26e82091aa127587df4777a-f9bc7e65a180459da53712205b991905 output=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/4689deb6f26e82091aa127587df4777a/d/f9bc7e65a180459da53712205b991905 2013-04-29 16:40:41,766 INFO org.apache.hadoop.hbase.snapshot.ExportSnapshot: copy file input=d/queries=03f6a36fef8c48f956272c75b1e45666-cb14e43f2f5a46ee96b2cbf389987d80 output=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/03f6a36fef8c48f956272c75b1e45666/d/cb14e43f2f5a46ee96b2cbf389987d80 2013-04-29 16:40:41,769 WARN org.apache.hadoop.hbase.snapshot.ExportSnapshot: Unable to get the status for file=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/03f6a36fef8c48f956272c75b1e45666/d/cb14e43f2f5a46ee96b2cbf389987d80 2013-04-29 16:40:43,627 INFO org.apache.hadoop.hbase.snapshot.ExportSnapshot: copy completed for input=d/queries=03f6a36fef8c48f956272c75b1e45666-cb14e43f2f5a46ee96b2cbf389987d80 output=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/03f6a36fef8c48f956272c75b1e45666/d/cb14e43f2f5a46ee96b2cbf389987d80 2013-04-29 16:40:43,628 INFO org.apache.hadoop.hbase.snapshot.ExportSnapshot: copy file input=d/queries=11d8de1728f7211a7cecb3f18531d317-367c5196865d4930aeb9f738ee5ea2c2 output=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/11d8de1728f7211a7cecb3f18531d317/d/367c5196865d4930aeb9f738ee5ea2c2 2013-04-29 16:40:43,636 WARN org.apache.hadoop.hbase.snapshot.ExportSnapshot: Unable to get the status for file=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/11d8de1728f7211a7cecb3f18531d317/d/367c5196865d4930aeb9f738ee5ea2c2 2013-04-29 16:40:44,819 INFO org.apache.hadoop.hbase.snapshot.ExportSnapshot: copy completed for input=d/queries=11d8de1728f7211a7cecb3f18531d317-367c5196865d4930aeb9f738ee5ea2c2 output=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/11d8de1728f7211a7cecb3f18531d317/d/367c5196865d4930aeb9f738ee5ea2c2 2013-04-29 16:40:44,819 INFO org.apache.hadoop.hbase.snapshot.ExportSnapshot: copy file input=d/queries=e54db7771e12e7103c45189aa9f6d621-a433ab2f363545c5909ed21d579e302c output=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/e54db7771e12e7103c45189aa9f6d621/d/a433ab2f363545c5909ed21d579e302c 2013-04-29 16:40:44,826 WARN org.apache.hadoop.hbase.snapshot.ExportSnapshot: Unable to get the status for file=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/e54db7771e12e7103c45189aa9f6d621/d/a433ab2f363545c5909ed21d579e302c 2013-04-29 16:40:45,967 INFO org.apache.hadoop.hbase.snapshot.ExportSnapshot: copy completed for input=d/queries=e54db7771e12e7103c45189aa9f6d621-a433ab2f363545c5909ed21d579e302c output=hdfs://namenode-backup:8020/users/sean/hbase_test/.archive/queries/e54db7771e12e7103c45189aa9f6d621/d/a433ab2f363545c5909ed21d579e302c 2013-04-29 16:40:45,967 INFO org.apache.hadoop.hbase.snapshot.ExportSnapshot: copy file input=d/queries=ae58ff4e7149b14efe6e5a0cc1d2f338-4b713d79670c4fd9aa309320adcb0047.e6dd75e2b917adcd657f5294093f0edd output=hdfs://namenode-backup:8020/users/sean/hbase_test/.oldlogs/queries=ae58ff4e7149b14efe6e5a0cc1d2f338-4b713d79670c4fd9aa309320adcb0047.e6dd75e2b917adcd657f5294093f0edd 2013-04-29 16:40:45,973 ERROR org.apache.hadoop.hbase.snapshot.ExportSnapshot: Unable to open source file=d/queries=ae58ff4e7149b14efe6e5a0cc1d2f338-4b713d79670c4fd9aa309320adcb0047.e6dd75e2b917adcd657f5294093f0edd java.io.FileNotFoundException: Unable to open link: org.apache.hadoop.hbase.io.HLogLink locations=[hdfs://namenode-primary:8020/hbase/.logs/d/queries=ae58ff4e7149b14efe6e5a0cc1d2f338-4b713d79670c4fd9aa309320adcb0047.e6dd75e2b917adcd657f5294093f0edd, hdfs://namenode-primary:8020/hbase/.oldlogs/queries=ae58ff4e7149b14efe6e5a0cc1d2f338-4b713d79670c4fd9aa309320adcb0047.e6dd75e2b917adcd657f5294093f0edd] at org.apache.hadoop.hbase.io.FileLink$FileLinkInputStream.tryOpen(FileLink.java:304) at org.apache.hadoop.hbase.io.FileLink$FileLinkInputStream.(FileLink.java:119) at org.apache.hadoop.hbase.io.FileLink$FileLinkInputStream.(FileLink.java:110) at org.apache.hadoop.hbase.io.FileLink.open(FileLink.java:389) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.openSourceFile(ExportSnapshot.java:303) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:174) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:145) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:94) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:140) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:673) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:331) at org.apache.hadoop.mapred.Child$4.run(Child.java:268) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) at org.apache.hadoop.mapred.Child.main(Child.java:262) 2013-04-29 16:40:45,975 INFO org.apache.hadoop.mapred.Task: Task:attempt_201304260021_3855_m_000007_0 is done. And is in the process of commiting 2013-04-29 16:40:45,999 INFO org.apache.hadoop.mapred.Task: Task 'attempt_201304260021_3855_m_000007_0' done.