Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- at java.io.ByteArrayOutputStream.hugeCapacity(ByteArrayOutputStream.java:123)
- at java.io.ByteArrayOutputStream.grow(ByteArrayOutputStream.java:117)
- at java.io.ByteArrayOutputStream.ensureCapacity(ByteArrayOutputStream.java:93)
- at java.io.ByteArrayOutputStream.write(ByteArrayOutputStream.java:153)
- at org.apache.spark.util.ByteBufferOutputStream.write(ByteBufferOutputStream.scala:41)
- at java.io.ObjectOutputStream$BlockDataOutputStream.drain(ObjectOutputStream.java:1877)
- at java.io.ObjectOutputStream$BlockDataOutputStream.setBlockDataMode(ObjectOutputStream.java:1786)
- at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1189)
- at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:348)
- at org.apache.spark.serializer.JavaSerializationStream.writeObject(JavaSerializer.scala:43)
- at org.apache.spark.serializer.JavaSerializerInstance.serialize(JavaSerializer.scala:100)
- at org.apache.spark.util.ClosureCleaner$.ensureSerializable(ClosureCleaner.scala:295)
- at org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:288)
- at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:108)
- at org.apache.spark.SparkContext.clean(SparkContext.scala:2288)
- at org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1.apply(RDD.scala:794)
- at org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1.apply(RDD.scala:793)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.RDD.mapPartitions(RDD.scala:793)
- at org.apache.spark.rdd.RDD$$anonfun$treeAggregate$1.apply(RDD.scala:1137)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.RDD.treeAggregate(RDD.scala:1128)
- at org.apache.spark.mllib.linalg.distributed.RowMatrix.computeGramianMatrix(RowMatrix.scala:122)
- at org.apache.spark.mllib.linalg.distributed.RowMatrix.computeCovariance(RowMatrix.scala:344)
- at org.apache.spark.mllib.stat.correlation.PearsonCorrelation$.computeCorrelationMatrix(PearsonCorrelation.scala:49)
- at org.apache.spark.mllib.stat.correlation.Correlations$.corrMatrix(Correlation.scala:66)
- at org.apache.spark.mllib.stat.Statistics$.corr(Statistics.scala:74)
- at com.mitchell.connections.Practice$.createCorrMatrix$1(Practice.scala:141)
- at com.mitchell.connections.Practice$.com$mitchell$connections$Practice$$creatingCorrmatrix$1(Practice.scala:147)
- at com.mitchell.connections.Practice$$anonfun$1.apply(Practice.scala:79)
- at com.mitchell.connections.Practice$$anonfun$1.apply(Practice.scala:79)
- at scala.collection.parallel.AugmentedIterableIterator$class.map2combiner(RemainsIterator.scala:115)
- at scala.collection.parallel.immutable.ParVector$ParVectorIterator.map2combiner(ParVector.scala:62)
- at scala.collection.parallel.ParIterableLike$Map.leaf(ParIterableLike.scala:1054)
- at scala.collection.parallel.Task$$anonfun$tryLeaf$1.apply$mcV$sp(Tasks.scala:49)
- at scala.collection.parallel.Task$$anonfun$tryLeaf$1.apply(Tasks.scala:48)
- at scala.collection.parallel.Task$$anonfun$tryLeaf$1.apply(Tasks.scala:48)
- at scala.collection.parallel.Task$class.tryLeaf(Tasks.scala:51)
- at scala.collection.parallel.ParIterableLike$Map.tryLeaf(ParIterableLike.scala:1051)
- at scala.collection.parallel.AdaptiveWorkStealingTasks$WrappedTask$class.compute(Tasks.scala:152)
- at scala.collection.parallel.AdaptiveWorkStealingForkJoinTasks$WrappedTask.compute(Tasks.scala:443)
- at scala.concurrent.forkjoin.RecursiveAction.exec(RecursiveAction.java:160)
- at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
- at scala.concurrent.forkjoin.ForkJoinTask.doJoin(ForkJoinTask.java:341)
- at scala.concurrent.forkjoin.ForkJoinTask.join(ForkJoinTask.java:673)
- at scala.collection.parallel.ForkJoinTasks$WrappedTask$class.sync(Tasks.scala:378)
- at scala.collection.parallel.AdaptiveWorkStealingForkJoinTasks$WrappedTask.sync(Tasks.scala:443)
- at scala.collection.parallel.ForkJoinTasks$class.executeAndWaitResult(Tasks.scala:426)
- at scala.collection.parallel.ForkJoinTaskSupport.executeAndWaitResult(TaskSupport.scala:56)
- at scala.collection.parallel.ExecutionContextTasks$class.executeAndWaitResult(Tasks.scala:558)
- at scala.collection.parallel.ExecutionContextTaskSupport.executeAndWaitResult(TaskSupport.scala:80)
- at scala.collection.parallel.ParIterableLike$ResultMapping.leaf(ParIterableLike.scala:958)
- at scala.collection.parallel.Task$$anonfun$tryLeaf$1.apply$mcV$sp(Tasks.scala:49)
- at scala.collection.parallel.Task$$anonfun$tryLeaf$1.apply(Tasks.scala:48)
- at scala.collection.parallel.Task$$anonfun$tryLeaf$1.apply(Tasks.scala:48)
- at scala.collection.parallel.Task$class.tryLeaf(Tasks.scala:51)
- at scala.collection.parallel.ParIterableLike$ResultMapping.tryLeaf(ParIterableLike.scala:953)
- at scala.collection.parallel.AdaptiveWorkStealingTasks$WrappedTask$class.compute(Tasks.scala:152)
- at scala.collection.parallel.AdaptiveWorkStealingForkJoinTasks$WrappedTask.compute(Tasks.scala:443)
- at scala.concurrent.forkjoin.RecursiveAction.exec(RecursiveAction.java:160)
- at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
- at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
- at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
- at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Add Comment
Please, Sign In to add comment