Advertisement
Guest User

Untitled

a guest
Jan 18th, 2018
99
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 9.24 KB | None | 0 0
  1. 18/01/18 14:49:33 INFO mapreduce.Job: Task Id : attempt_1516255624131_0003_m_000017_0, Status : FAILED
  2. Error: java.nio.file.AccessDeniedException: benchmarks/TestDFSIO/io_data/test_io_117: Multi-part upload with id '2~3ysCY3O6rmNu9bjekXpYfyar4DDt6K3' to benchmarks/TestDFSIO/io_data/test_io_117 on benchmarks/TestDFSIO/io_data/test_io_117: com.amazonaws.services.s3.model.AmazonS3Exception: null (Service: Amazon S3; Status Code: 403; Error Code: SignatureDoesNotMatch; Request ID: tx000000000000000752570-005a6042fa-1598-default; S3 Extended Request ID: 1598-default-default), S3 Extended Request ID: 1598-default-default
  3. at org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:174)
  4. at org.apache.hadoop.fs.s3a.S3AUtils.extractException(S3AUtils.java:216)
  5. at org.apache.hadoop.fs.s3a.S3ABlockOutputStream$MultiPartUpload.waitForAllPartUploads(S3ABlockOutputStream.java:549)
  6. at org.apache.hadoop.fs.s3a.S3ABlockOutputStream$MultiPartUpload.access$100(S3ABlockOutputStream.java:465)
  7. at org.apache.hadoop.fs.s3a.S3ABlockOutputStream.close(S3ABlockOutputStream.java:355)
  8. at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:72)
  9. at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:106)
  10. at org.apache.hadoop.fs.IOMapperBase.map(IOMapperBase.java:136)
  11. at org.apache.hadoop.fs.IOMapperBase.map(IOMapperBase.java:37)
  12. at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
  13. at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:459)
  14. at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
  15. at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:177)
  16. at java.security.AccessController.doPrivileged(Native Method)
  17. at javax.security.auth.Subject.doAs(Subject.java:422)
  18. at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1886)
  19. at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:171)
  20. Caused by: com.amazonaws.services.s3.model.AmazonS3Exception: null (Service: Amazon S3; Status Code: 403; Error Code: SignatureDoesNotMatch; Request ID: tx000000000000000752570-005a6042fa-1598-default; S3 Extended Request ID: 1598-default-default), S3 Extended Request ID: 1598-default-default
  21. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.handleErrorResponse(AmazonHttpClient.java:1638)
  22. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeOneRequest(AmazonHttpClient.java:1303)
  23. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeHelper(AmazonHttpClient.java:1055)
  24. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.doExecute(AmazonHttpClient.java:743)
  25. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeWithTimer(AmazonHttpClient.java:717)
  26. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.execute(AmazonHttpClient.java:699)
  27. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.access$500(AmazonHttpClient.java:667)
  28. at com.amazonaws.http.AmazonHttpClient$RequestExecutionBuilderImpl.execute(AmazonHttpClient.java:649)
  29. at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:513)
  30. at com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:4229)
  31. at com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:4176)
  32. at com.amazonaws.services.s3.AmazonS3Client.doUploadPart(AmazonS3Client.java:3213)
  33. at com.amazonaws.services.s3.AmazonS3Client.uploadPart(AmazonS3Client.java:3198)
  34. at org.apache.hadoop.fs.s3a.S3AFileSystem.uploadPart(S3AFileSystem.java:1301)
  35. at org.apache.hadoop.fs.s3a.S3ABlockOutputStream$MultiPartUpload$1.call(S3ABlockOutputStream.java:512)
  36. at org.apache.hadoop.fs.s3a.S3ABlockOutputStream$MultiPartUpload$1.call(S3ABlockOutputStream.java:503)
  37. at org.apache.hadoop.fs.s3a.SemaphoredDelegatingExecutor$CallableWithPermitRelease.call(SemaphoredDelegatingExecutor.java:222)
  38. at org.apache.hadoop.fs.s3a.SemaphoredDelegatingExecutor$CallableWithPermitRelease.call(SemaphoredDelegatingExecutor.java:222)
  39. at java.util.concurrent.FutureTask.run(FutureTask.java:266)
  40. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  41. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  42. at java.lang.Thread.run(Thread.java:748)
  43.  
  44.  
  45.  
  46.  
  47. 2018-01-18 14:34:29,800 WARN [main] org.apache.hadoop.mapred.YarnChild: Exception running child : org.apache.hadoop.fs.s3a.AWSS3IOException: Multi-part upload with id '2~05AR4KU-_CheO6MA5gFsGccdhWSydZK' to 10000/_temporary/1/_temporary/attempt_1516255624131_0001_m_000052_2/web_sales/data-m-00052 on 10000/_temporary/1/_temporary/attempt_1516255624131_0001_m_000052_2/web_sales/data-m-00052: com.amazonaws.services.s3.model.AmazonS3Exception: Bad Request (Service: Amazon S3; Status Code: 400; Error Code: 400 Bad Request; Request ID: null; S3 Extended Request ID: null), S3 Extended Request ID: null: Bad Request (Service: Amazon S3; Status Code: 400; Error Code: 400 Bad Request; Request ID: null; S3 Extended Request ID: null)
  48. at org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:194)
  49. at org.apache.hadoop.fs.s3a.S3AUtils.extractException(S3AUtils.java:216)
  50. at org.apache.hadoop.fs.s3a.S3ABlockOutputStream$MultiPartUpload.waitForAllPartUploads(S3ABlockOutputStream.java:549)
  51. at org.apache.hadoop.fs.s3a.S3ABlockOutputStream$MultiPartUpload.access$100(S3ABlockOutputStream.java:465)
  52. at org.apache.hadoop.fs.s3a.S3ABlockOutputStream.close(S3ABlockOutputStream.java:355)
  53. at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:72)
  54. at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:106)
  55. at org.apache.hadoop.mapreduce.lib.output.TextOutputFormat$LineRecordWriter.close(TextOutputFormat.java:101)
  56. at org.apache.hadoop.mapreduce.lib.output.MultipleOutputs.close(MultipleOutputs.java:574)
  57. at org.notmysock.tpcds.GenTable$DSDGen.cleanup(GenTable.java:169)
  58. at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149)
  59. at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:793)
  60. at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
  61. at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:177)
  62. at java.security.AccessController.doPrivileged(Native Method)
  63. at javax.security.auth.Subject.doAs(Subject.java:422)
  64. at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1886)
  65. at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:171)
  66. Caused by: com.amazonaws.services.s3.model.AmazonS3Exception: Bad Request (Service: Amazon S3; Status Code: 400; Error Code: 400 Bad Request; Request ID: null; S3 Extended Request ID: null), S3 Extended Request ID: null
  67. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.handleErrorResponse(AmazonHttpClient.java:1638)
  68. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeOneRequest(AmazonHttpClient.java:1303)
  69. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeHelper(AmazonHttpClient.java:1055)
  70. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.doExecute(AmazonHttpClient.java:743)
  71. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeWithTimer(AmazonHttpClient.java:717)
  72. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.execute(AmazonHttpClient.java:699)
  73. at com.amazonaws.http.AmazonHttpClient$RequestExecutor.access$500(AmazonHttpClient.java:667)
  74. at com.amazonaws.http.AmazonHttpClient$RequestExecutionBuilderImpl.execute(AmazonHttpClient.java:649)
  75. at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:513)
  76. at com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:4229)
  77. at com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:4176)
  78. at com.amazonaws.services.s3.AmazonS3Client.doUploadPart(AmazonS3Client.java:3213)
  79. at com.amazonaws.services.s3.AmazonS3Client.uploadPart(AmazonS3Client.java:3198)
  80. at org.apache.hadoop.fs.s3a.S3AFileSystem.uploadPart(S3AFileSystem.java:1301)
  81. at org.apache.hadoop.fs.s3a.S3ABlockOutputStream$MultiPartUpload$1.call(S3ABlockOutputStream.java:512)
  82. at org.apache.hadoop.fs.s3a.S3ABlockOutputStream$MultiPartUpload$1.call(S3ABlockOutputStream.java:503)
  83. at org.apache.hadoop.fs.s3a.SemaphoredDelegatingExecutor$CallableWithPermitRelease.call(SemaphoredDelegatingExecutor.java:222)
  84. at org.apache.hadoop.fs.s3a.SemaphoredDelegatingExecutor$CallableWithPermitRelease.call(SemaphoredDelegatingExecutor.java:222)
  85. at java.util.concurrent.FutureTask.run(FutureTask.java:266)
  86. at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  87. at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  88. at java.lang.Thread.run(Thread.java:748)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement