Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- -rw-r--r-- 1 root hadoop 9126 2015-12-11 19:16 pig_1449861377791.log
- -rw-r--r-- 1 root hadoop 9126 2015-12-11 19:23 pig_1449861813663.log
- -rw-r--r-- 1 root hadoop 9126 2015-12-11 19:26 pig_1449861996576.log
- -rwxrwxrwx 1 root hadoop 3543194 2012-02-06 19:05 Pitching.csv
- -rwxrwxrwx 1 root hadoop 367382 2012-02-07 23:46 PitchingPost.csv
- -rwxrwxrwx 1 root hadoop 30363 2012-02-08 00:04 readme59.txt
- -rwxrwxrwx 1 root hadoop 842057 2011-11-15 22:11 Salaries.csv
- -rwxrwxrwx 1 root hadoop 50365 2011-11-15 22:11 Schools.csv
- -rwxrwxrwx 1 root hadoop 205354 2011-11-15 22:11 SchoolsPlayers.csv
- -rwxrwxrwx 1 root hadoop 8088 2012-02-07 23:34 SeriesPost.csv
- -rwxrwxrwx 1 root hadoop 524526 2012-02-03 23:50 Teams.csv
- -rwxrwxrwx 1 root hadoop 4111 2011-11-15 22:11 TeamsFranchises.csv
- -rwxrwxrwx 1 root hadoop 2149 2011-11-15 22:11 TeamsHalf.csv
- [root@sandbox lahman591-csv]# hdfs dfs -ls /user/guest
- Found 1 items
- -rwxrwxrwx 3 root sandbox 6398886 2015-12-11 00:53 /user/guest/Batting.csv
- [root@sandbox lahman591-csv]# su hdfs; dfs -chown -R root:hadoop /user/guest
- [hdfs@sandbox lahman591-csv]$ hdfs dfs -chown -R root:hadoop /user/guest
- [hdfs@sandbox lahman591-csv]$ exit
- exit
- -bash: dfs: command not found
- [root@sandbox lahman591-csv]# hdfs dfs -ls /user/guest
- Found 1 items
- -rwxrwxrwx 3 root hadoop 6398886 2015-12-11 00:53 /user/guest/Batting.csv
- [root@sandbox lahman591-csv]# pig 1.pig
- WARNING: Use "yarn jar" to launch YARN applications.
- 15/12/11 19:31:44 INFO pig.ExecTypeProvider: Trying ExecType : LOCAL
- 15/12/11 19:31:44 INFO pig.ExecTypeProvider: Trying ExecType : MAPREDUCE
- 15/12/11 19:31:44 INFO pig.ExecTypeProvider: Picked MAPREDUCE as the ExecType
- 2015-12-11 19:31:44,356 [main] INFO org.apache.pig.Main - Apache Pig version 0.15.0.2.3.2.0-2950 (rexported) compiled Sep 30 2015, 19:39:20
- 2015-12-11 19:31:44,356 [main] INFO org.apache.pig.Main - Logging error messages to: /root/lahman591-csv/pig_1449862304354.log
- 2015-12-11 19:31:45,137 [main] INFO org.apache.pig.impl.util.Utils - Default bootup file /root/.pigbootup not found
- 2015-12-11 19:31:45,257 [main] INFO org.apache.pig.backend.hadoop.executionengine.HExecutionEngine - Connecting to hadoop file system at: hdfs://sandbox.hortonworks.com:8020
- 2015-12-11 19:31:46,516 [main] WARN org.apache.pig.newplan.BaseOperatorPlan - Encountered Warning IMPLICIT_CAST_TO_INT 1 time(s).
- 2015-12-11 19:31:46,516 [main] WARN org.apache.pig.newplan.BaseOperatorPlan - Encountered Warning IMPLICIT_CAST_TO_DOUBLE 1 time(s).
- 2015-12-11 19:31:46,545 [main] INFO org.apache.pig.tools.pigstats.ScriptState - Pig features used in the script: HASH_JOIN,GROUP_BY,FILTER
- 2015-12-11 19:31:46,593 [main] INFO org.apache.pig.data.SchemaTupleBackend - Key [pig.schematuple] was not set... will not generate code.
- 2015-12-11 19:31:46,641 [main] INFO org.apache.pig.newplan.logical.optimizer.LogicalPlanOptimizer - {RULES_ENABLED=[AddForEach, ColumnMapKeyPrune, ConstantCalculator, GroupByConstParallelSetter, LimitOptimizer, LoadTypeCastInserter, MergeFilter, MergeForEach, PartitionFilterOptimizer, PredicatePushdownOptimizer, PushDownForEachFlatten, PushUpFilter, SplitFilter, StreamTypeCastInserter]}
- 2015-12-11 19:31:46,785 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MRCompiler - File concatenation threshold: 100 optimistic? false
- 2015-12-11 19:31:46,819 [main] INFO org.apache.pig.backend.hadoop.executionengine.util.CombinerOptimizerUtil - Choosing to move algebraic foreach to combiner
- 2015-12-11 19:31:46,860 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MRCompiler$LastInputStreamingOptimizer - Rewrite: POPackage->POForEach to POPackage(JoinPackager)
- 2015-12-11 19:31:46,873 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MultiQueryOptimizer - MR plan size before optimization: 3
- 2015-12-11 19:31:46,874 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MultiQueryOptimizer - Merged 1 map-reduce splittees.
- 2015-12-11 19:31:46,874 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MultiQueryOptimizer - Merged 1 out of total 3 MR operators.
- 2015-12-11 19:31:46,874 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MultiQueryOptimizer - MR plan size after optimization: 2
- 2015-12-11 19:31:47,388 [main] INFO org.apache.hadoop.yarn.client.api.impl.TimelineClientImpl - Timeline service address: http://sandbox.hortonworks.com:8188/ws/v1/timeline/
- 2015-12-11 19:31:47,555 [main] INFO org.apache.hadoop.yarn.client.RMProxy - Connecting to ResourceManager at sandbox.hortonworks.com/10.0.2.15:8050
- 2015-12-11 19:31:47,752 [main] INFO org.apache.pig.tools.pigstats.mapreduce.MRScriptState - Pig script settings are added to the job
- 2015-12-11 19:31:47,758 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - mapred.job.reduce.markreset.buffer.percent is not set, set to default 0.3
- 2015-12-11 19:31:47,760 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Reduce phase detected, estimating # of required reducers.
- 2015-12-11 19:31:47,762 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Using reducer estimator: org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.InputSizeReducerEstimator
- 2015-12-11 19:31:47,777 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.InputSizeReducerEstimator - BytesPerReducer=1000000000 maxReducers=999 totalInputFileSize=6398886
- 2015-12-11 19:31:47,777 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Setting Parallelism to 1
- 2015-12-11 19:31:47,777 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - This job cannot be converted run in-process
- 2015-12-11 19:31:48,160 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Added jar file:/usr/hdp/2.3.2.0-2950/pig/pig-0.15.0.2.3.2.0-2950-core-h2.jar to DistributedCache through /tmp/temp-1331804897/tmp-86214663/pig-0.15.0.2.3.2.0-2950-core-h2.jar
- 2015-12-11 19:31:48,195 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Added jar file:/usr/hdp/2.3.2.0-2950/pig/lib/automaton-1.11-8.jar to DistributedCache through /tmp/temp-1331804897/tmp507004575/automaton-1.11-8.jar
- 2015-12-11 19:31:48,230 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Added jar file:/usr/hdp/2.3.2.0-2950/pig/lib/antlr-runtime-3.4.jar to DistributedCache through /tmp/temp-1331804897/tmp-2114352180/antlr-runtime-3.4.jar
- 2015-12-11 19:31:48,280 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Added jar file:/usr/hdp/2.3.2.0-2950/hadoop-mapreduce/joda-time-2.8.2.jar to DistributedCache through /tmp/temp-1331804897/tmp1114773941/joda-time-2.8.2.jar
- 2015-12-11 19:31:48,349 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Setting up multi store job
- 2015-12-11 19:31:48,360 [main] INFO org.apache.pig.data.SchemaTupleFrontend - Key [pig.schematuple] is false, will not generate code.
- 2015-12-11 19:31:48,360 [main] INFO org.apache.pig.data.SchemaTupleFrontend - Starting process to move generated code to distributed cacche
- 2015-12-11 19:31:48,360 [main] INFO org.apache.pig.data.SchemaTupleFrontend - Setting key [pig.schematuple.classes] with classes to deserialize []
- 2015-12-11 19:31:48,471 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - 1 map-reduce job(s) waiting for submission.
- 2015-12-11 19:31:48,717 [JobControl] INFO org.apache.hadoop.yarn.client.api.impl.TimelineClientImpl - Timeline service address: http://sandbox.hortonworks.com:8188/ws/v1/timeline/
- 2015-12-11 19:31:48,717 [JobControl] INFO org.apache.hadoop.yarn.client.RMProxy - Connecting to ResourceManager at sandbox.hortonworks.com/10.0.2.15:8050
- 2015-12-11 19:31:48,800 [JobControl] INFO org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob - PigLatin:1.pig got an error while submitting
- org.apache.hadoop.security.AccessControlException: Permission denied: user=root, access=WRITE, inode="/user/root/.staging":hdfs:hdfs:drwxr-xr-x
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:319)
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:292)
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:213)
- at org.apache.ranger.authorization.hadoop.RangerHdfsAuthorizer$RangerAccessControlEnforcer.checkPermission(RangerHdfsAuthorizer.java:300)
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:190)
- at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1771)
- at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1755)
- at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAncestorAccess(FSDirectory.java:1738)
- at org.apache.hadoop.hdfs.server.namenode.FSDirMkdirOp.mkdirs(FSDirMkdirOp.java:71)
- at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3896)
- at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:984)
- at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:622)
- at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
- at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2137)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2133)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:415)
- at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
- at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2131)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
- at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
- at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)
- at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:3010)
- at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2978)
- at org.apache.hadoop.hdfs.DistributedFileSystem$21.doCall(DistributedFileSystem.java:1047)
- at org.apache.hadoop.hdfs.DistributedFileSystem$21.doCall(DistributedFileSystem.java:1043)
- at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
- at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:1043)
- at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:1036)
- at org.apache.hadoop.mapreduce.JobSubmissionFiles.getStagingDir(JobSubmissionFiles.java:133)
- at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:144)
- at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
- at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:415)
- at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
- at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
- at org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:335)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:606)
- at org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:128)
- at org.apache.pig.backend.hadoop23.PigJobControl.run(PigJobControl.java:194)
- at java.lang.Thread.run(Thread.java:745)
- at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher$1.run(MapReduceLauncher.java:276)
- Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): Permission denied: user=root, access=WRITE, inode="/user/root/.staging":hdfs:hdfs:drwxr-xr-x
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:319)
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:292)
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:213)
- at org.apache.ranger.authorization.hadoop.RangerHdfsAuthorizer$RangerAccessControlEnforcer.checkPermission(RangerHdfsAuthorizer.java:300)
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:190)
- at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1771)
- at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1755)
- at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAncestorAccess(FSDirectory.java:1738)
- at org.apache.hadoop.hdfs.server.namenode.FSDirMkdirOp.mkdirs(FSDirMkdirOp.java:71)
- at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3896)
- at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:984)
- at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:622)
- at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
- at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2137)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2133)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:415)
- at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
- at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2131)
- at org.apache.hadoop.ipc.Client.call(Client.java:1427)
- at org.apache.hadoop.ipc.Client.call(Client.java:1358)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
- at com.sun.proxy.$Proxy11.mkdirs(Unknown Source)
- at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:558)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:606)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
- at com.sun.proxy.$Proxy12.mkdirs(Unknown Source)
- at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:3008)
- ... 23 more
- 2015-12-11 19:31:48,975 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - 0% complete
- 2015-12-11 19:31:53,993 [main] WARN org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - Ooops! Some job has failed! Specify -stop_on_failure if you want Pig to stop immediately on failure.
- 2015-12-11 19:31:53,993 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - job null has failed! Stop running all dependent jobs
- 2015-12-11 19:31:53,993 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - 100% complete
- 2015-12-11 19:31:54,044 [main] ERROR org.apache.pig.tools.pigstats.mapreduce.MRPigStatsUtil - 1 map reduce job(s) failed!
- 2015-12-11 19:31:54,047 [main] INFO org.apache.pig.tools.pigstats.mapreduce.SimplePigStats - Script Statistics:
- HadoopVersion PigVersion UserId StartedAt FinishedAt Features
- 2.7.1.2.3.2.0-2950 0.15.0.2.3.2.0-2950 root 2015-12-11 19:31:47 2015-12-11 19:31:54 HASH_JOIN,GROUP_BY,FILTER
- Failed!
- Failed Jobs:
- JobId Alias Feature Message Outputs
- N/A batting,grp_data,max_runs,raw_runs,runs MULTI_QUERY,COMBINER Message: org.apache.hadoop.security.AccessControlException: Permission denied: user=root, access=WRITE, inode="/user/root/.staging":hdfs:hdfs:drwxr-xr-x
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:319)
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:292)
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:213)
- at org.apache.ranger.authorization.hadoop.RangerHdfsAuthorizer$RangerAccessControlEnforcer.checkPermission(RangerHdfsAuthorizer.java:300)
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:190)
- at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1771)
- at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1755)
- at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAncestorAccess(FSDirectory.java:1738)
- at org.apache.hadoop.hdfs.server.namenode.FSDirMkdirOp.mkdirs(FSDirMkdirOp.java:71)
- at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3896)
- at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:984)
- at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:622)
- at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
- at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2137)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2133)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:415)
- at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
- at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2131)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
- at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
- at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)
- at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:3010)
- at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2978)
- at org.apache.hadoop.hdfs.DistributedFileSystem$21.doCall(DistributedFileSystem.java:1047)
- at org.apache.hadoop.hdfs.DistributedFileSystem$21.doCall(DistributedFileSystem.java:1043)
- at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
- at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:1043)
- at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:1036)
- at org.apache.hadoop.mapreduce.JobSubmissionFiles.getStagingDir(JobSubmissionFiles.java:133)
- at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:144)
- at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
- at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:415)
- at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
- at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
- at org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:335)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:606)
- at org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:128)
- at org.apache.pig.backend.hadoop23.PigJobControl.run(PigJobControl.java:194)
- at java.lang.Thread.run(Thread.java:745)
- at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher$1.run(MapReduceLauncher.java:276)
- Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): Permission denied: user=root, access=WRITE, inode="/user/root/.staging":hdfs:hdfs:drwxr-xr-x
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:319)
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:292)
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:213)
- at org.apache.ranger.authorization.hadoop.RangerHdfsAuthorizer$RangerAccessControlEnforcer.checkPermission(RangerHdfsAuthorizer.java:300)
- at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:190)
- at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1771)
- at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1755)
- at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAncestorAccess(FSDirectory.java:1738)
- at org.apache.hadoop.hdfs.server.namenode.FSDirMkdirOp.mkdirs(FSDirMkdirOp.java:71)
- at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3896)
- at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:984)
- at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:622)
- at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
- at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2137)
- at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2133)
- at java.security.AccessController.doPrivileged(Native Method)
- at javax.security.auth.Subject.doAs(Subject.java:415)
- at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
- at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2131)
- at org.apache.hadoop.ipc.Client.call(Client.java:1427)
- at org.apache.hadoop.ipc.Client.call(Client.java:1358)
- at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
- at com.sun.proxy.$Proxy11.mkdirs(Unknown Source)
- at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:558)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:606)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
- at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
- at com.sun.proxy.$Proxy12.mkdirs(Unknown Source)
- at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:3008)
- ... 23 more
- Input(s):
- Failed to read data from "/user/guest/Batting.csv"
- Output(s):
- Counters:
- Total records written : 0
- Total bytes written : 0
- Spillable Memory Manager spill count : 0
- Total bags proactively spilled: 0
- Total records proactively spilled: 0
- Job DAG:
- null -> null,
- null
- 2015-12-11 19:31:54,049 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - Failed!
- 2015-12-11 19:31:54,052 [main] ERROR org.apache.pig.tools.grunt.Grunt - ERROR 1066: Unable to open iterator for alias join_data
- Details at logfile: /root/lahman591-csv/pig_1449862304354.log
- 2015-12-11 19:31:54,097 [main] INFO org.apache.pig.Main - Pig script completed in 9 seconds and 968 milliseconds (9968 ms)
- [root@sandbox lahman591-csv]# hdfs dfs -ls /user/guest
- Found 1 items
- -rwxrwxrwx 3 root hadoop 6398886 2015-12-11 00:53 /user/guest/Batting.csv
- [root@sandbox lahman591-csv]# ls -l
- total 30496
- -rwxrwxrwx 1 root hadoop 454 2015-12-11 03:44 1.pig
- -rwxrwxrwx 1 root hadoop 195488 2011-11-29 04:55 AllstarFull.csv
- -rwxrwxrwx 1 root hadoop 5651119 2012-02-07 23:56 Appearances.csv
- -rwxrwxrwx 1 root hadoop 2273 2011-11-28 23:57 AwardsManagers.csv
- -rwxrwxrwx 1 root hadoop 97304 2011-11-29 00:15 AwardsPlayers.csv
- -rwxrwxrwx 1 root hadoop 16134 2011-11-29 05:25 AwardsShareManagers.csv
- -rwxrwxrwx 1 root hadoop 216987 2011-11-29 05:44 AwardsSharePlayers.csv
- -rwxrwxrwx 1 root hadoop 6398886 2014-09-05 00:05 Batting.csv
- -rwxrwxrwx 1 root hadoop 621765 2012-02-07 23:34 BattingPost.csv
- -rwxrwxrwx 1 root hadoop 8063747 2011-11-28 23:49 Fielding.csv
- -rwxrwxrwx 1 root hadoop 322538 2011-11-15 22:11 FieldingOF.csv
- -rwxrwxrwx 1 root hadoop 552230 2012-02-07 23:34 FieldingPost.csv
- -rwxrwxrwx 1 root hadoop 172984 2012-02-07 23:23 HallOfFame.csv
- -rwxrwxrwx 1 root hadoop 133114 2011-11-29 18:59 Managers.csv
- -rwxrwxrwx 1 root hadoop 4240 2011-11-15 22:11 ManagersHalf.csv
- -rwxrwxrwx 1 root hadoop 3024713 2012-02-07 23:34 Master.csv
- -rw-r--r-- 1 root hadoop 9126 2015-12-11 01:08 pig_1449796077165.log
- -rw-r--r-- 1 root hadoop 9126 2015-12-11 01:10 pig_1449796244442.log
- -rw-r--r-- 1 root hadoop 9126 2015-12-11 01:15 pig_1449796523062.log
- -rw-r--r-- 1 root hadoop 9126 2015-12-11 01:27 pig_1449797245424.log
- -rw-r--r-- 1 root hadoop 9126 2015-12-11 01:33 pig_1449797575651.log
- -rw-r--r-- 1 root hadoop 9126 2015-12-11 03:44 pig_1449805474981.log
- -rw-r--r-- 1 root hadoop 0 2015-12-11 07:03 pig_1449817401972.log
- -rw-r--r-- 1 root hadoop 9126 2015-12-11 19:16 pig_1449861377791.log
- -rw-r--r-- 1 root hadoop 9126 2015-12-11 19:23 pig_1449861813663.log
- -rw-r--r-- 1 root hadoop 9126 2015-12-11 19:26 pig_1449861996576.log
- -rw-r--r-- 1 root root 9126 2015-12-11 19:31 pig_1449862304354.log
- -rwxrwxrwx 1 root hadoop 3543194 2012-02-06 19:05 Pitching.csv
- -rwxrwxrwx 1 root hadoop 367382 2012-02-07 23:46 PitchingPost.csv
- -rwxrwxrwx 1 root hadoop 30363 2012-02-08 00:04 readme59.txt
- -rwxrwxrwx 1 root hadoop 842057 2011-11-15 22:11 Salaries.csv
- -rwxrwxrwx 1 root hadoop 50365 2011-11-15 22:11 Schools.csv
- -rwxrwxrwx 1 root hadoop 205354 2011-11-15 22:11 SchoolsPlayers.csv
- -rwxrwxrwx 1 root hadoop 8088 2012-02-07 23:34 SeriesPost.csv
- -rwxrwxrwx 1 root hadoop 524526 2012-02-03 23:50 Teams.csv
- -rwxrwxrwx 1 root hadoop 4111 2011-11-15 22:11 TeamsFranchises.csv
- -rwxrwxrwx 1 root hadoop 2149 2011-11-15 22:11 TeamsHalf.csv
- [root@sandbox lahman591-csv]#
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement