Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- "Executor task launch worker-29" #42075 daemon prio=5 os_prio=0 tid=0x00007fcdfd32a000 nid=0x5782 runnable [0x00007fcf38a0c000]
- java.lang.Thread.State: RUNNABLE
- at org.apache.hadoop.io.compress.zlib.ZlibCompressor.deflateBytesDirect(Native Method)
- at org.apache.hadoop.io.compress.zlib.ZlibCompressor.compress(ZlibCompressor.java:368)
- - locked <0x00007fd766772b30> (a org.apache.hadoop.io.compress.GzipCodec$GzipZlibCompressor)
- at org.apache.hadoop.io.compress.CompressorStream.compress(CompressorStream.java:81)
- at org.apache.hadoop.io.compress.CompressorStream.write(CompressorStream.java:76)
- at parquet.bytes.CapacityByteArrayOutputStream.writeTo(CapacityByteArrayOutputStream.java:146)
- at parquet.bytes.BytesInput$CapacityBAOSBytesInput.writeAllTo(BytesInput.java:308)
- at parquet.bytes.BytesInput$SequenceBytesIn.writeAllTo(BytesInput.java:233)
- at parquet.hadoop.CodecFactory$BytesCompressor.compress(CodecFactory.java:108)
- at parquet.hadoop.ColumnChunkPageWriteStore$ColumnChunkPageWriter.writePage(ColumnChunkPageWriteStore.java:125)
- at parquet.column.impl.ColumnWriterImpl.writePage(ColumnWriterImpl.java:147)
- at parquet.column.impl.ColumnWriterImpl.accountForValueWritten(ColumnWriterImpl.java:108)
- at parquet.column.impl.ColumnWriterImpl.write(ColumnWriterImpl.java:200)
- at parquet.io.MessageColumnIO$MessageColumnIORecordConsumer.addBinary(MessageColumnIO.java:340)
- at org.apache.spark.sql.parquet.MutableRowWriteSupport.consumeType(ParquetTableSupport.scala:365)
- at org.apache.spark.sql.parquet.MutableRowWriteSupport.write(ParquetTableSupport.scala:347)
- at org.apache.spark.sql.parquet.MutableRowWriteSupport.write(ParquetTableSupport.scala:333)
- at parquet.hadoop.InternalParquetRecordWriter.write(InternalParquetRecordWriter.java:120)
- at parquet.hadoop.ParquetRecordWriter.write(ParquetRecordWriter.java:81)
- at parquet.hadoop.ParquetRecordWriter.write(ParquetRecordWriter.java:37)
- at org.apache.spark.sql.parquet.InsertIntoParquetTable.org$apache$spark$sql$parquet$InsertIntoParquetTable$$writeShard$1(ParquetTableOperations.scala:500)
- at org.apache.spark.sql.parquet.InsertIntoParquetTable$$anonfun$saveAsHadoopFile$1.apply(ParquetTableOperations.scala:517)
- at org.apache.spark.sql.parquet.InsertIntoParquetTable$$anonfun$saveAsHadoopFile$1.apply(ParquetTableOperations.scala:517)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61)
- at org.apache.spark.scheduler.Task.run(Task.scala:56)
- at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:200)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
- at java.lang.Thread.run(Unknown Source)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement