Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- george@...:~/adam$ bin/adam-submit --jars=$GOOGLE_CLOUD_NIO_JAR -- reads2coverage gs://genomics-public-data/test-data/dna/wgs/hiseq2500/NA12878/H06JUADXX130110.1.ATCACGAT.20k_reads.bam /users/george/out.coverage4.bam
- Using ADAM_MAIN=org.bdgenomics.adam.cli.ADAMMain
- Using spark-submit=/usr/lib/spark/bin/spark-submit
- 17/12/05 07:48:46 INFO org.bdgenomics.adam.cli.ADAMMain: ADAM invoked with args: "reads2coverage" "gs://genomics-public-data/test-data/dna/wgs/hiseq2500/NA12878/H06JUADXX130110.1.ATCACGAT.20k_reads.bam" "/users/george/out.coverage4.bam"
- 17/12/05 07:48:49 INFO org.spark_project.jetty.util.log: Logging initialized @4390ms
- 17/12/05 07:48:49 INFO org.spark_project.jetty.server.Server: jetty-9.3.z-SNAPSHOT
- 17/12/05 07:48:49 INFO org.spark_project.jetty.server.Server: Started @4496ms
- 17/12/05 07:48:49 INFO org.spark_project.jetty.server.AbstractConnector: Started ServerConnector@340fbf35{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
- 17/12/05 07:48:50 INFO com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystemBase: GHFS version: 1.6.1-hadoop2
- 17/12/05 07:48:52 INFO org.apache.hadoop.yarn.client.RMProxy: Connecting to ResourceManager at mango-2-m/10.142.0.7:8032
- 17/12/05 07:48:56 INFO org.apache.hadoop.yarn.client.api.impl.YarnClientImpl: Submitted application application_1512456264283_0011
- 17/12/05 07:49:03 INFO org.bdgenomics.adam.rdd.ADAMContext: Loading gs://genomics-public-data/test-data/dna/wgs/hiseq2500/NA12878/H06JUADXX130110.1.ATCACGAT.20k_reads.bam as BAM/CRAM/SAM and converting to AlignmentRecords.
- 17/12/05 07:49:05 INFO org.bdgenomics.adam.rdd.ADAMContext: Loaded header from gs://genomics-public-data/test-data/dna/wgs/hiseq2500/NA12878/H06JUADXX130110.1.ATCACGAT.20k_reads.bam
- 17/12/05 07:49:07 INFO org.apache.hadoop.mapreduce.lib.input.FileInputFormat: Total input files to process : 1
- 17/12/05 07:49:08 INFO org.bdgenomics.adam.rdd.feature.RDDBoundFeatureRDD: Saving data in ADAM format
- 17/12/05 07:49:26 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, mango-2-w-1.c.mango-bdgenomics.internal, executor 1): com.google.cloud.storage.StorageException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.translate(HttpStorageRpc.java:189)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:340)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:197)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:194)
- at shaded.cloud_nio.com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:89)
- at com.google.cloud.RetryHelper.run(RetryHelper.java:74)
- at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:51)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:194)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:208)
- at com.google.cloud.storage.contrib.nio.CloudStorageFileSystemProvider.readAttributes(CloudStorageFileSystemProvider.java:647)
- at java.nio.file.Files.readAttributes(Files.java:1737)
- at java.nio.file.Files.isRegularFile(Files.java:2229)
- at htsjdk.samtools.SamFiles.lookForIndex(SamFiles.java:72)
- at htsjdk.samtools.SamFiles.findIndex(SamFiles.java:39)
- at org.seqdoop.hadoop_bam.BAMRecordReader.initialize(BAMRecordReader.java:143)
- at org.seqdoop.hadoop_bam.BAMInputFormat.createRecordReader(BAMInputFormat.java:226)
- at org.seqdoop.hadoop_bam.AnySAMInputFormat.createRecordReader(AnySAMInputFormat.java:190)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:179)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.io.IOException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at shaded.cloud_nio.com.google.auth.oauth2.ComputeEngineCredentials.refreshAccessToken(ComputeEngineCredentials.java:152)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.refresh(OAuth2Credentials.java:175)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.getRequestMetadata(OAuth2Credentials.java:161)
- at shaded.cloud_nio.com.google.auth.http.HttpCredentialsAdapter.initialize(HttpCredentialsAdapter.java:96)
- at com.google.cloud.http.HttpTransportOptions$1.initialize(HttpTransportOptions.java:157)
- at shaded.cloud_nio.com.google.api.client.http.HttpRequestFactory.buildRequest(HttpRequestFactory.java:93)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.buildHttpRequest(AbstractGoogleClientRequest.java:300)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:419)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:338)
- ... 37 more
- 17/12/05 07:49:27 ERROR org.apache.spark.scheduler.TaskSetManager: Task 0 in stage 0.0 failed 4 times; aborting job
- 17/12/05 07:49:27 ERROR org.apache.spark.internal.io.SparkHadoopMapReduceWriter: Aborting job job_20171205074908_0007.
- org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 3, mango-2-w-1.c.mango-bdgenomics.internal, executor 1): com.google.cloud.storage.StorageException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.translate(HttpStorageRpc.java:189)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:340)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:197)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:194)
- at shaded.cloud_nio.com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:89)
- at com.google.cloud.RetryHelper.run(RetryHelper.java:74)
- at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:51)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:194)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:208)
- at com.google.cloud.storage.contrib.nio.CloudStorageFileSystemProvider.readAttributes(CloudStorageFileSystemProvider.java:647)
- at java.nio.file.Files.readAttributes(Files.java:1737)
- at java.nio.file.Files.isRegularFile(Files.java:2229)
- at htsjdk.samtools.SamFiles.lookForIndex(SamFiles.java:72)
- at htsjdk.samtools.SamFiles.findIndex(SamFiles.java:39)
- at org.seqdoop.hadoop_bam.BAMRecordReader.initialize(BAMRecordReader.java:143)
- at org.seqdoop.hadoop_bam.BAMInputFormat.createRecordReader(BAMInputFormat.java:226)
- at org.seqdoop.hadoop_bam.AnySAMInputFormat.createRecordReader(AnySAMInputFormat.java:190)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:179)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.io.IOException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at shaded.cloud_nio.com.google.auth.oauth2.ComputeEngineCredentials.refreshAccessToken(ComputeEngineCredentials.java:152)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.refresh(OAuth2Credentials.java:175)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.getRequestMetadata(OAuth2Credentials.java:161)
- at shaded.cloud_nio.com.google.auth.http.HttpCredentialsAdapter.initialize(HttpCredentialsAdapter.java:96)
- at com.google.cloud.http.HttpTransportOptions$1.initialize(HttpTransportOptions.java:157)
- at shaded.cloud_nio.com.google.api.client.http.HttpRequestFactory.buildRequest(HttpRequestFactory.java:93)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.buildHttpRequest(AbstractGoogleClientRequest.java:300)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:419)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:338)
- ... 37 more
- Driver stacktrace:
- at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1499)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1487)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1486)
- at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
- at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1486)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)
- at scala.Option.foreach(Option.scala:257)
- at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:814)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1714)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1669)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1658)
- at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:630)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2022)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2043)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2075)
- at org.apache.spark.internal.io.SparkHadoopMapReduceWriter$.write(SparkHadoopMapReduceWriter.scala:88)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1085)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1085)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1085)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopDataset(PairRDDFunctions.scala:1084)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply$mcV$sp(PairRDDFunctions.scala:1003)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply(PairRDDFunctions.scala:994)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply(PairRDDFunctions.scala:994)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopFile(PairRDDFunctions.scala:994)
- at org.apache.spark.rdd.InstrumentedPairRDDFunctions.saveAsNewAPIHadoopFile(InstrumentedPairRDDFunctions.scala:477)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions$$anonfun$saveRddAsParquet$1.apply$mcV$sp(ADAMRDDFunctions.scala:165)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions$$anonfun$saveRddAsParquet$1.apply(ADAMRDDFunctions.scala:149)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions$$anonfun$saveRddAsParquet$1.apply(ADAMRDDFunctions.scala:149)
- at scala.Option.fold(Option.scala:158)
- at org.apache.spark.rdd.Timer.time(Timer.scala:48)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions.saveRddAsParquet(ADAMRDDFunctions.scala:149)
- at org.bdgenomics.adam.rdd.AvroGenomicRDD.saveAsParquet(GenomicRDD.scala:2481)
- at org.bdgenomics.adam.rdd.AvroGenomicRDD.saveAsParquet(GenomicRDD.scala:2456)
- at org.bdgenomics.adam.rdd.feature.FeatureRDD.save(FeatureRDD.scala:411)
- at org.bdgenomics.adam.rdd.feature.CoverageRDD.save(CoverageRDD.scala:209)
- at org.bdgenomics.adam.cli.Reads2Coverage.run(Reads2Coverage.scala:102)
- at org.bdgenomics.utils.cli.BDGSparkCommand$class.run(BDGCommand.scala:55)
- at org.bdgenomics.adam.cli.Reads2Coverage.run(Reads2Coverage.scala:65)
- at org.bdgenomics.adam.cli.ADAMMain.apply(ADAMMain.scala:126)
- at org.bdgenomics.adam.cli.ADAMMain$.main(ADAMMain.scala:65)
- at org.bdgenomics.adam.cli.ADAMMain.main(ADAMMain.scala)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:755)
- at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
- at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
- at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:119)
- at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
- Caused by: com.google.cloud.storage.StorageException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.translate(HttpStorageRpc.java:189)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:340)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:197)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:194)
- at shaded.cloud_nio.com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:89)
- at com.google.cloud.RetryHelper.run(RetryHelper.java:74)
- at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:51)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:194)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:208)
- at com.google.cloud.storage.contrib.nio.CloudStorageFileSystemProvider.readAttributes(CloudStorageFileSystemProvider.java:647)
- at java.nio.file.Files.readAttributes(Files.java:1737)
- at java.nio.file.Files.isRegularFile(Files.java:2229)
- at htsjdk.samtools.SamFiles.lookForIndex(SamFiles.java:72)
- at htsjdk.samtools.SamFiles.findIndex(SamFiles.java:39)
- at org.seqdoop.hadoop_bam.BAMRecordReader.initialize(BAMRecordReader.java:143)
- at org.seqdoop.hadoop_bam.BAMInputFormat.createRecordReader(BAMInputFormat.java:226)
- at org.seqdoop.hadoop_bam.AnySAMInputFormat.createRecordReader(AnySAMInputFormat.java:190)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:179)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.io.IOException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at shaded.cloud_nio.com.google.auth.oauth2.ComputeEngineCredentials.refreshAccessToken(ComputeEngineCredentials.java:152)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.refresh(OAuth2Credentials.java:175)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.getRequestMetadata(OAuth2Credentials.java:161)
- at shaded.cloud_nio.com.google.auth.http.HttpCredentialsAdapter.initialize(HttpCredentialsAdapter.java:96)
- at com.google.cloud.http.HttpTransportOptions$1.initialize(HttpTransportOptions.java:157)
- at shaded.cloud_nio.com.google.api.client.http.HttpRequestFactory.buildRequest(HttpRequestFactory.java:93)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.buildHttpRequest(AbstractGoogleClientRequest.java:300)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:419)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:338)
- ... 37 more
- Command body threw exception:
- org.apache.spark.SparkException: Job aborted.
- 17/12/05 07:49:27 INFO org.bdgenomics.adam.cli.Reads2Coverage: Overall Duration: 40.56 secs
- Exception in thread "main" org.apache.spark.SparkException: Job aborted.
- at org.apache.spark.internal.io.SparkHadoopMapReduceWriter$.write(SparkHadoopMapReduceWriter.scala:107)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1085)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1085)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1085)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopDataset(PairRDDFunctions.scala:1084)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply$mcV$sp(PairRDDFunctions.scala:1003)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply(PairRDDFunctions.scala:994)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply(PairRDDFunctions.scala:994)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopFile(PairRDDFunctions.scala:994)
- at org.apache.spark.rdd.InstrumentedPairRDDFunctions.saveAsNewAPIHadoopFile(InstrumentedPairRDDFunctions.scala:477)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions$$anonfun$saveRddAsParquet$1.apply$mcV$sp(ADAMRDDFunctions.scala:165)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions$$anonfun$saveRddAsParquet$1.apply(ADAMRDDFunctions.scala:149)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions$$anonfun$saveRddAsParquet$1.apply(ADAMRDDFunctions.scala:149)
- at scala.Option.fold(Option.scala:158)
- at org.apache.spark.rdd.Timer.time(Timer.scala:48)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions.saveRddAsParquet(ADAMRDDFunctions.scala:149)
- at org.bdgenomics.adam.rdd.AvroGenomicRDD.saveAsParquet(GenomicRDD.scala:2481)
- at org.bdgenomics.adam.rdd.AvroGenomicRDD.saveAsParquet(GenomicRDD.scala:2456)
- at org.bdgenomics.adam.rdd.feature.FeatureRDD.save(FeatureRDD.scala:411)
- at org.bdgenomics.adam.rdd.feature.CoverageRDD.save(CoverageRDD.scala:209)
- at org.bdgenomics.adam.cli.Reads2Coverage.run(Reads2Coverage.scala:102)
- at org.bdgenomics.utils.cli.BDGSparkCommand$class.run(BDGCommand.scala:55)
- at org.bdgenomics.adam.cli.Reads2Coverage.run(Reads2Coverage.scala:65)
- at org.bdgenomics.adam.cli.ADAMMain.apply(ADAMMain.scala:126)
- at org.bdgenomics.adam.cli.ADAMMain$.main(ADAMMain.scala:65)
- at org.bdgenomics.adam.cli.ADAMMain.main(ADAMMain.scala)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:755)
- at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
- at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
- at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:119)
- at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
- Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 3, mango-2-w-1.c.mango-bdgenomics.internal, executor 1): com.google.cloud.storage.StorageException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.translate(HttpStorageRpc.java:189)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:340)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:197)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:194)
- at shaded.cloud_nio.com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:89)
- at com.google.cloud.RetryHelper.run(RetryHelper.java:74)
- at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:51)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:194)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:208)
- at com.google.cloud.storage.contrib.nio.CloudStorageFileSystemProvider.readAttributes(CloudStorageFileSystemProvider.java:647)
- at java.nio.file.Files.readAttributes(Files.java:1737)
- at java.nio.file.Files.isRegularFile(Files.java:2229)
- at htsjdk.samtools.SamFiles.lookForIndex(SamFiles.java:72)
- at htsjdk.samtools.SamFiles.findIndex(SamFiles.java:39)
- at org.seqdoop.hadoop_bam.BAMRecordReader.initialize(BAMRecordReader.java:143)
- at org.seqdoop.hadoop_bam.BAMInputFormat.createRecordReader(BAMInputFormat.java:226)
- at org.seqdoop.hadoop_bam.AnySAMInputFormat.createRecordReader(AnySAMInputFormat.java:190)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:179)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.io.IOException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at shaded.cloud_nio.com.google.auth.oauth2.ComputeEngineCredentials.refreshAccessToken(ComputeEngineCredentials.java:152)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.refresh(OAuth2Credentials.java:175)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.getRequestMetadata(OAuth2Credentials.java:161)
- at shaded.cloud_nio.com.google.auth.http.HttpCredentialsAdapter.initialize(HttpCredentialsAdapter.java:96)
- at com.google.cloud.http.HttpTransportOptions$1.initialize(HttpTransportOptions.java:157)
- at shaded.cloud_nio.com.google.api.client.http.HttpRequestFactory.buildRequest(HttpRequestFactory.java:93)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.buildHttpRequest(AbstractGoogleClientRequest.java:300)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:419)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:338)
- ... 37 more
- Driver stacktrace:
- at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1499)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1487)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1486)
- at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
- at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1486)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)
- at scala.Option.foreach(Option.scala:257)
- at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:814)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1714)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1669)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1658)
- at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:630)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2022)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2043)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2075)
- at org.apache.spark.internal.io.SparkHadoopMapReduceWriter$.write(SparkHadoopMapReduceWriter.scala:88)
- ... 40 more
- Caused by: com.google.cloud.storage.StorageException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.translate(HttpStorageRpc.java:189)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:340)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:197)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:194)
- at shaded.cloud_nio.com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:89)
- at com.google.cloud.RetryHelper.run(RetryHelper.java:74)
- at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:51)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:194)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:208)
- at com.google.cloud.storage.contrib.nio.CloudStorageFileSystemProvider.readAttributes(CloudStorageFileSystemProvider.java:647)
- at java.nio.file.Files.readAttributes(Files.java:1737)
- at java.nio.file.Files.isRegularFile(Files.java:2229)
- at htsjdk.samtools.SamFiles.lookForIndex(SamFiles.java:72)
- at htsjdk.samtools.SamFiles.findIndex(SamFiles.java:39)
- at org.seqdoop.hadoop_bam.BAMRecordReader.initialize(BAMRecordReader.java:143)
- at org.seqdoop.hadoop_bam.BAMInputFormat.createRecordReader(BAMInputFormat.java:226)
- at org.seqdoop.hadoop_bam.AnySAMInputFormat.createRecordReader(AnySAMInputFormat.java:190)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:179)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.io.IOException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at shaded.cloud_nio.com.google.auth.oauth2.ComputeEngineCredentials.refreshAccessToken(ComputeEngineCredentials.java:152)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.refresh(OAuth2Credentials.java:175)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.getRequestMetadata(OAuth2Credentials.java:161)
- at shaded.cloud_nio.com.google.auth.http.HttpCredentialsAdapter.initialize(HttpCredentialsAdapter.java:96)
- at com.google.cloud.http.HttpTransportOptions$1.initialize(HttpTransportOptions.java:157)
- at shaded.cloud_nio.com.google.api.client.http.HttpRequestFactory.buildRequest(HttpRequestFactory.java:93)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.buildHttpRequest(AbstractGoogleClientRequest.java:300)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:419)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:338)
- ... 37 more
- 17/12/05 07:49:27 INFO org.spark_project.jetty.server.AbstractConnector: Stopped Spark@340fbf35{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
- george@mango-2-m:~/adam$ bin/adam-submit --jars=$GOOGLE_CLOUD_NIO_JAR -- reads2coverage gs://gen^Cics-public-data/test-data/dna/wgs/hiseq2500/NA12878/H06JUADXX130110.1.ATCACGAT.20k_reads.bam /users/george/out.coverage4.bam
- george@mango-2-m:~/adam$ sudo bin/adam-submit --jars=$GOOGLE_CLOUD_NIO_JAR -- reads2coverage gs://genomics-public-data/test-data/dna/wgs/hiseq2500/NA12878/H06JUADXX130110.1.ATCACGAT.20k_reads.bam /users/george/out.coverage5.bam
- Using ADAM_MAIN=org.bdgenomics.adam.cli.ADAMMain
- Using spark-submit=/usr/bin/spark-submit
- 17/12/05 07:54:43 INFO org.bdgenomics.adam.cli.ADAMMain: ADAM invoked with args: "reads2coverage" "gs://genomics-public-data/test-data/dna/wgs/hiseq2500/NA12878/H06JUADXX130110.1.ATCACGAT.20k_reads.bam" "/users/george/out.coverage5.bam"
- 17/12/05 07:54:46 INFO org.spark_project.jetty.util.log: Logging initialized @4802ms
- 17/12/05 07:54:46 INFO org.spark_project.jetty.server.Server: jetty-9.3.z-SNAPSHOT
- 17/12/05 07:54:46 INFO org.spark_project.jetty.server.Server: Started @5031ms
- 17/12/05 07:54:46 INFO org.spark_project.jetty.server.AbstractConnector: Started ServerConnector@2dc74087{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
- 17/12/05 07:54:47 INFO com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystemBase: GHFS version: 1.6.1-hadoop2
- 17/12/05 07:54:49 INFO org.apache.hadoop.yarn.client.RMProxy: Connecting to ResourceManager at mango-2-m/10.142.0.7:8032
- 17/12/05 07:54:54 INFO org.apache.hadoop.yarn.client.api.impl.YarnClientImpl: Submitted application application_1512456264283_0012
- 17/12/05 07:55:03 INFO org.bdgenomics.adam.rdd.ADAMContext: Loading gs://genomics-public-data/test-data/dna/wgs/hiseq2500/NA12878/H06JUADXX130110.1.ATCACGAT.20k_reads.bam as BAM/CRAM/SAM and converting to AlignmentRecords.
- 17/12/05 07:55:05 INFO org.bdgenomics.adam.rdd.ADAMContext: Loaded header from gs://genomics-public-data/test-data/dna/wgs/hiseq2500/NA12878/H06JUADXX130110.1.ATCACGAT.20k_reads.bam
- 17/12/05 07:55:07 INFO org.apache.hadoop.mapreduce.lib.input.FileInputFormat: Total input files to process : 1
- 17/12/05 07:55:08 INFO org.bdgenomics.adam.rdd.feature.RDDBoundFeatureRDD: Saving data in ADAM format
- 17/12/05 07:55:26 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, mango-2-w-1.c.mango-bdgenomics.internal, executor 1): com.google.cloud.storage.StorageException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.translate(HttpStorageRpc.java:189)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:340)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:197)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:194)
- at shaded.cloud_nio.com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:89)
- at com.google.cloud.RetryHelper.run(RetryHelper.java:74)
- at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:51)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:194)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:208)
- at com.google.cloud.storage.contrib.nio.CloudStorageFileSystemProvider.readAttributes(CloudStorageFileSystemProvider.java:647)
- at java.nio.file.Files.readAttributes(Files.java:1737)
- at java.nio.file.Files.isRegularFile(Files.java:2229)
- at htsjdk.samtools.SamFiles.lookForIndex(SamFiles.java:72)
- at htsjdk.samtools.SamFiles.findIndex(SamFiles.java:39)
- at org.seqdoop.hadoop_bam.BAMRecordReader.initialize(BAMRecordReader.java:143)
- at org.seqdoop.hadoop_bam.BAMInputFormat.createRecordReader(BAMInputFormat.java:226)
- at org.seqdoop.hadoop_bam.AnySAMInputFormat.createRecordReader(AnySAMInputFormat.java:190)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:179)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.io.IOException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at shaded.cloud_nio.com.google.auth.oauth2.ComputeEngineCredentials.refreshAccessToken(ComputeEngineCredentials.java:152)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.refresh(OAuth2Credentials.java:175)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.getRequestMetadata(OAuth2Credentials.java:161)
- at shaded.cloud_nio.com.google.auth.http.HttpCredentialsAdapter.initialize(HttpCredentialsAdapter.java:96)
- at com.google.cloud.http.HttpTransportOptions$1.initialize(HttpTransportOptions.java:157)
- at shaded.cloud_nio.com.google.api.client.http.HttpRequestFactory.buildRequest(HttpRequestFactory.java:93)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.buildHttpRequest(AbstractGoogleClientRequest.java:300)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:419)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:338)
- ... 37 more
- 17/12/05 07:55:27 ERROR org.apache.spark.scheduler.TaskSetManager: Task 0 in stage 0.0 failed 4 times; aborting job
- 17/12/05 07:55:27 ERROR org.apache.spark.internal.io.SparkHadoopMapReduceWriter: Aborting job job_20171205075508_0007.
- org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 3, mango-2-w-1.c.mango-bdgenomics.internal, executor 1): com.google.cloud.storage.StorageException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.translate(HttpStorageRpc.java:189)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:340)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:197)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:194)
- at shaded.cloud_nio.com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:89)
- at com.google.cloud.RetryHelper.run(RetryHelper.java:74)
- at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:51)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:194)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:208)
- at com.google.cloud.storage.contrib.nio.CloudStorageFileSystemProvider.readAttributes(CloudStorageFileSystemProvider.java:647)
- at java.nio.file.Files.readAttributes(Files.java:1737)
- at java.nio.file.Files.isRegularFile(Files.java:2229)
- at htsjdk.samtools.SamFiles.lookForIndex(SamFiles.java:72)
- at htsjdk.samtools.SamFiles.findIndex(SamFiles.java:39)
- at org.seqdoop.hadoop_bam.BAMRecordReader.initialize(BAMRecordReader.java:143)
- at org.seqdoop.hadoop_bam.BAMInputFormat.createRecordReader(BAMInputFormat.java:226)
- at org.seqdoop.hadoop_bam.AnySAMInputFormat.createRecordReader(AnySAMInputFormat.java:190)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:179)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.io.IOException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at shaded.cloud_nio.com.google.auth.oauth2.ComputeEngineCredentials.refreshAccessToken(ComputeEngineCredentials.java:152)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.refresh(OAuth2Credentials.java:175)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.getRequestMetadata(OAuth2Credentials.java:161)
- at shaded.cloud_nio.com.google.auth.http.HttpCredentialsAdapter.initialize(HttpCredentialsAdapter.java:96)
- at com.google.cloud.http.HttpTransportOptions$1.initialize(HttpTransportOptions.java:157)
- at shaded.cloud_nio.com.google.api.client.http.HttpRequestFactory.buildRequest(HttpRequestFactory.java:93)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.buildHttpRequest(AbstractGoogleClientRequest.java:300)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:419)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:338)
- ... 37 more
- Driver stacktrace:
- at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1499)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1487)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1486)
- at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
- at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1486)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)
- at scala.Option.foreach(Option.scala:257)
- at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:814)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1714)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1669)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1658)
- at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:630)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2022)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2043)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2075)
- at org.apache.spark.internal.io.SparkHadoopMapReduceWriter$.write(SparkHadoopMapReduceWriter.scala:88)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1085)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1085)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1085)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopDataset(PairRDDFunctions.scala:1084)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply$mcV$sp(PairRDDFunctions.scala:1003)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply(PairRDDFunctions.scala:994)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply(PairRDDFunctions.scala:994)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopFile(PairRDDFunctions.scala:994)
- at org.apache.spark.rdd.InstrumentedPairRDDFunctions.saveAsNewAPIHadoopFile(InstrumentedPairRDDFunctions.scala:477)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions$$anonfun$saveRddAsParquet$1.apply$mcV$sp(ADAMRDDFunctions.scala:165)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions$$anonfun$saveRddAsParquet$1.apply(ADAMRDDFunctions.scala:149)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions$$anonfun$saveRddAsParquet$1.apply(ADAMRDDFunctions.scala:149)
- at scala.Option.fold(Option.scala:158)
- at org.apache.spark.rdd.Timer.time(Timer.scala:48)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions.saveRddAsParquet(ADAMRDDFunctions.scala:149)
- at org.bdgenomics.adam.rdd.AvroGenomicRDD.saveAsParquet(GenomicRDD.scala:2481)
- at org.bdgenomics.adam.rdd.AvroGenomicRDD.saveAsParquet(GenomicRDD.scala:2456)
- at org.bdgenomics.adam.rdd.feature.FeatureRDD.save(FeatureRDD.scala:411)
- at org.bdgenomics.adam.rdd.feature.CoverageRDD.save(CoverageRDD.scala:209)
- at org.bdgenomics.adam.cli.Reads2Coverage.run(Reads2Coverage.scala:102)
- at org.bdgenomics.utils.cli.BDGSparkCommand$class.run(BDGCommand.scala:55)
- at org.bdgenomics.adam.cli.Reads2Coverage.run(Reads2Coverage.scala:65)
- at org.bdgenomics.adam.cli.ADAMMain.apply(ADAMMain.scala:126)
- at org.bdgenomics.adam.cli.ADAMMain$.main(ADAMMain.scala:65)
- at org.bdgenomics.adam.cli.ADAMMain.main(ADAMMain.scala)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:755)
- at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
- at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
- at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:119)
- at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
- Caused by: com.google.cloud.storage.StorageException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.translate(HttpStorageRpc.java:189)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:340)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:197)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:194)
- at shaded.cloud_nio.com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:89)
- at com.google.cloud.RetryHelper.run(RetryHelper.java:74)
- at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:51)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:194)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:208)
- at com.google.cloud.storage.contrib.nio.CloudStorageFileSystemProvider.readAttributes(CloudStorageFileSystemProvider.java:647)
- at java.nio.file.Files.readAttributes(Files.java:1737)
- at java.nio.file.Files.isRegularFile(Files.java:2229)
- at htsjdk.samtools.SamFiles.lookForIndex(SamFiles.java:72)
- at htsjdk.samtools.SamFiles.findIndex(SamFiles.java:39)
- at org.seqdoop.hadoop_bam.BAMRecordReader.initialize(BAMRecordReader.java:143)
- at org.seqdoop.hadoop_bam.BAMInputFormat.createRecordReader(BAMInputFormat.java:226)
- at org.seqdoop.hadoop_bam.AnySAMInputFormat.createRecordReader(AnySAMInputFormat.java:190)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:179)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.io.IOException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at shaded.cloud_nio.com.google.auth.oauth2.ComputeEngineCredentials.refreshAccessToken(ComputeEngineCredentials.java:152)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.refresh(OAuth2Credentials.java:175)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.getRequestMetadata(OAuth2Credentials.java:161)
- at shaded.cloud_nio.com.google.auth.http.HttpCredentialsAdapter.initialize(HttpCredentialsAdapter.java:96)
- at com.google.cloud.http.HttpTransportOptions$1.initialize(HttpTransportOptions.java:157)
- at shaded.cloud_nio.com.google.api.client.http.HttpRequestFactory.buildRequest(HttpRequestFactory.java:93)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.buildHttpRequest(AbstractGoogleClientRequest.java:300)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:419)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:338)
- ... 37 more
- Command body threw exception:
- org.apache.spark.SparkException: Job aborted.
- 17/12/05 07:55:27 INFO org.bdgenomics.adam.cli.Reads2Coverage: Overall Duration: 43.69 secs
- Exception in thread "main" org.apache.spark.SparkException: Job aborted.
- at org.apache.spark.internal.io.SparkHadoopMapReduceWriter$.write(SparkHadoopMapReduceWriter.scala:107)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1085)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1085)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1085)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopDataset(PairRDDFunctions.scala:1084)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply$mcV$sp(PairRDDFunctions.scala:1003)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply(PairRDDFunctions.scala:994)
- at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopFile$2.apply(PairRDDFunctions.scala:994)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopFile(PairRDDFunctions.scala:994)
- at org.apache.spark.rdd.InstrumentedPairRDDFunctions.saveAsNewAPIHadoopFile(InstrumentedPairRDDFunctions.scala:477)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions$$anonfun$saveRddAsParquet$1.apply$mcV$sp(ADAMRDDFunctions.scala:165)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions$$anonfun$saveRddAsParquet$1.apply(ADAMRDDFunctions.scala:149)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions$$anonfun$saveRddAsParquet$1.apply(ADAMRDDFunctions.scala:149)
- at scala.Option.fold(Option.scala:158)
- at org.apache.spark.rdd.Timer.time(Timer.scala:48)
- at org.bdgenomics.adam.rdd.ADAMRDDFunctions.saveRddAsParquet(ADAMRDDFunctions.scala:149)
- at org.bdgenomics.adam.rdd.AvroGenomicRDD.saveAsParquet(GenomicRDD.scala:2481)
- at org.bdgenomics.adam.rdd.AvroGenomicRDD.saveAsParquet(GenomicRDD.scala:2456)
- at org.bdgenomics.adam.rdd.feature.FeatureRDD.save(FeatureRDD.scala:411)
- at org.bdgenomics.adam.rdd.feature.CoverageRDD.save(CoverageRDD.scala:209)
- at org.bdgenomics.adam.cli.Reads2Coverage.run(Reads2Coverage.scala:102)
- at org.bdgenomics.utils.cli.BDGSparkCommand$class.run(BDGCommand.scala:55)
- at org.bdgenomics.adam.cli.Reads2Coverage.run(Reads2Coverage.scala:65)
- at org.bdgenomics.adam.cli.ADAMMain.apply(ADAMMain.scala:126)
- at org.bdgenomics.adam.cli.ADAMMain$.main(ADAMMain.scala:65)
- at org.bdgenomics.adam.cli.ADAMMain.main(ADAMMain.scala)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:755)
- at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
- at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
- at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:119)
- at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
- Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 3, mango-2-w-1.c.mango-bdgenomics.internal, executor 1): com.google.cloud.storage.StorageException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.translate(HttpStorageRpc.java:189)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:340)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:197)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:194)
- at shaded.cloud_nio.com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:89)
- at com.google.cloud.RetryHelper.run(RetryHelper.java:74)
- at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:51)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:194)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:208)
- at com.google.cloud.storage.contrib.nio.CloudStorageFileSystemProvider.readAttributes(CloudStorageFileSystemProvider.java:647)
- at java.nio.file.Files.readAttributes(Files.java:1737)
- at java.nio.file.Files.isRegularFile(Files.java:2229)
- at htsjdk.samtools.SamFiles.lookForIndex(SamFiles.java:72)
- at htsjdk.samtools.SamFiles.findIndex(SamFiles.java:39)
- at org.seqdoop.hadoop_bam.BAMRecordReader.initialize(BAMRecordReader.java:143)
- at org.seqdoop.hadoop_bam.BAMInputFormat.createRecordReader(BAMInputFormat.java:226)
- at org.seqdoop.hadoop_bam.AnySAMInputFormat.createRecordReader(AnySAMInputFormat.java:190)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:179)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.io.IOException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at shaded.cloud_nio.com.google.auth.oauth2.ComputeEngineCredentials.refreshAccessToken(ComputeEngineCredentials.java:152)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.refresh(OAuth2Credentials.java:175)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.getRequestMetadata(OAuth2Credentials.java:161)
- at shaded.cloud_nio.com.google.auth.http.HttpCredentialsAdapter.initialize(HttpCredentialsAdapter.java:96)
- at com.google.cloud.http.HttpTransportOptions$1.initialize(HttpTransportOptions.java:157)
- at shaded.cloud_nio.com.google.api.client.http.HttpRequestFactory.buildRequest(HttpRequestFactory.java:93)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.buildHttpRequest(AbstractGoogleClientRequest.java:300)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:419)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:338)
- ... 37 more
- Driver stacktrace:
- at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1499)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1487)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1486)
- at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
- at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1486)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)
- at scala.Option.foreach(Option.scala:257)
- at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:814)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1714)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1669)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1658)
- at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:630)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2022)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2043)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2075)
- at org.apache.spark.internal.io.SparkHadoopMapReduceWriter$.write(SparkHadoopMapReduceWriter.scala:88)
- ... 40 more
- Caused by: com.google.cloud.storage.StorageException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.translate(HttpStorageRpc.java:189)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:340)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:197)
- at com.google.cloud.storage.StorageImpl$5.call(StorageImpl.java:194)
- at shaded.cloud_nio.com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:89)
- at com.google.cloud.RetryHelper.run(RetryHelper.java:74)
- at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:51)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:194)
- at com.google.cloud.storage.StorageImpl.get(StorageImpl.java:208)
- at com.google.cloud.storage.contrib.nio.CloudStorageFileSystemProvider.readAttributes(CloudStorageFileSystemProvider.java:647)
- at java.nio.file.Files.readAttributes(Files.java:1737)
- at java.nio.file.Files.isRegularFile(Files.java:2229)
- at htsjdk.samtools.SamFiles.lookForIndex(SamFiles.java:72)
- at htsjdk.samtools.SamFiles.findIndex(SamFiles.java:39)
- at org.seqdoop.hadoop_bam.BAMRecordReader.initialize(BAMRecordReader.java:143)
- at org.seqdoop.hadoop_bam.BAMInputFormat.createRecordReader(BAMInputFormat.java:226)
- at org.seqdoop.hadoop_bam.AnySAMInputFormat.createRecordReader(AnySAMInputFormat.java:190)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)
- at org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:179)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)
- at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
- at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
- at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
- at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:748)
- Caused by: java.io.IOException: Error code 404 trying to get security access token from Compute Engine metadata for the default service account. This may be because the virtual machine instance does not have permission scopes specified.
- at shaded.cloud_nio.com.google.auth.oauth2.ComputeEngineCredentials.refreshAccessToken(ComputeEngineCredentials.java:152)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.refresh(OAuth2Credentials.java:175)
- at shaded.cloud_nio.com.google.auth.oauth2.OAuth2Credentials.getRequestMetadata(OAuth2Credentials.java:161)
- at shaded.cloud_nio.com.google.auth.http.HttpCredentialsAdapter.initialize(HttpCredentialsAdapter.java:96)
- at com.google.cloud.http.HttpTransportOptions$1.initialize(HttpTransportOptions.java:157)
- at shaded.cloud_nio.com.google.api.client.http.HttpRequestFactory.buildRequest(HttpRequestFactory.java:93)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.buildHttpRequest(AbstractGoogleClientRequest.java:300)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:419)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
- at shaded.cloud_nio.com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
- at com.google.cloud.storage.spi.v1.HttpStorageRpc.get(HttpStorageRpc.java:338)
- ... 37 more
- 17/12/05 07:55:27 INFO org.spark_project.jetty.server.AbstractConnector: Stopped Spark@2dc74087{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement