Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- 2019-04-24T12:17:55.363+0000 SCALA 20190424121725g1lCs xcsf-pro processing xcsf-gl-krb-inr-ingresosefectivo-01 gl INFO org.apache.spark.internal.Logging$class: ResultStage 0 (zipWithIndex at OffsetTransformation.scala:56) failed in 4.744 s due to Job aborted due to stage failure: Task 5 in stage 0.0 failed 4 times, most recent failure: Lost task 5.3 in stage 0.0 (TID 9, 192.168.192.107, e 1): java.lang.RuntimeException: Malformed CSV record
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser.org$apache$spark$sql$e n$datasources$csv$UnivocityParser$$convert(UnivocityParser.scala:213)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser.parse(UnivocityParser.scala:191)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:308)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:308)
- at org.apache.spark.sql.e n.datasources.FailureSafeParser.parse(FailureSafeParser.scala:60)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$parseIterator$1.apply(UnivocityParser.scala:312)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$parseIterator$1.apply(UnivocityParser.scala:312)
- at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
- at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at org.apache.spark.sql.e n.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:105)
- at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
- at org.apache.spark.sql.e n.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
- at org.apache.spark.sql.e n.WholeStageCodegenE onfun$8$$anon$1.hasNext(WholeStageCodegenE la:395)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1793)
- at org.apache.spark.rdd.ZippedWithIndexRDD$$anonfun$2.apply(ZippedWithIndexRDD.scala:52)
- at org.apache.spark.rdd.ZippedWithIndexRDD$$anonfun$2.apply(ZippedWithIndexRDD.scala:52)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2070)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2070)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.e .E $TaskRunner.run(E .scala:338)
- at java.util.concurrent.ThreadPoolE .runWorker(ThreadPoolE .java:1149)
- at java.util.concurrent.ThreadPoolE $Worker.run(ThreadPoolE .java:624)
- at java.lang.Thread.run(Thread.java:748)
- Driver stacktrace:
- 2019-04-24T12:17:55.368+0000 SCALA 20190424121725g1lCs xcsf-pro processing xcsf-gl-krb-inr-ingresosefectivo-01 gl INFO org.apache.spark.internal.Logging$class: Job 0 failed: zipWithIndex at OffsetTransformation.scala:56, took 4.906124 s
- 2019-04-24T12:17:55.370+0000 SCALA 20190424121725g1lCs xcsf-pro processing xcsf-gl-krb-inr-ingresosefectivo-01 gl ERROR com.datio.kirby.CheckFlow$class: Exception applying transformations: org.apache.spark.SparkException: Job aborted due to stage failure: Task 5 in stage 0.0 failed 4 times, most recent failure: Lost task 5.3 in stage 0.0 (TID 9, 192.168.192.107, e 1): java.lang.RuntimeException: Malformed CSV record
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser.org$apache$spark$sql$e n$datasources$csv$UnivocityParser$$convert(UnivocityParser.scala:213)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser.parse(UnivocityParser.scala:191)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:308)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:308)
- at org.apache.spark.sql.e n.datasources.FailureSafeParser.parse(FailureSafeParser.scala:60)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$parseIterator$1.apply(UnivocityParser.scala:312)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$parseIterator$1.apply(UnivocityParser.scala:312)
- at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
- at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at org.apache.spark.sql.e n.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:105)
- at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
- at org.apache.spark.sql.e n.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
- at org.apache.spark.sql.e n.WholeStageCodegenE onfun$8$$anon$1.hasNext(WholeStageCodegenE la:395)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1793)
- at org.apache.spark.rdd.ZippedWithIndexRDD$$anonfun$2.apply(ZippedWithIndexRDD.scala:52)
- at org.apache.spark.rdd.ZippedWithIndexRDD$$anonfun$2.apply(ZippedWithIndexRDD.scala:52)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2070)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2070)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.e .E $TaskRunner.run(E .scala:338)
- at java.util.concurrent.ThreadPoolE .runWorker(ThreadPoolE .java:1149)
- at java.util.concurrent.ThreadPoolE $Worker.run(ThreadPoolE .java:624)
- at java.lang.Thread.run(Thread.java:748)
- Driver stacktrace:
- at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1517)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1505)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1504)
- at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
- at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1504)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)
- at scala.Option.foreach(Option.scala:257)
- at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:814)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1732)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1687)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1676)
- at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:630)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2030)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2051)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2070)
- at org.apache.spark.rdd.ZippedWithIndexRDD.(ZippedWithIndexRDD.scala:50)
- at org.apache.spark.rdd.RDD$$anonfun$zipWithIndex$1.apply(RDD.scala:1294)
- at org.apache.spark.rdd.RDD$$anonfun$zipWithIndex$1.apply(RDD.scala:1294)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.RDD.zipWithIndex(RDD.scala:1293)
- at com.bbva.cib.core.transformations.row.offset.OffsetTransformation.transform(OffsetTransformation.scala:56)
- at com.datio.kirby.CheckFlow$$anonfun$applyTransformations$1.apply(CheckFlow.scala:48)
- at com.datio.kirby.CheckFlow$$anonfun$applyTransformations$1.apply(CheckFlow.scala:42)
- at scala.collection.Inde ptimized$class.foldl(Inde ptimized.scala:57)
- at scala.collection.Inde ptimized$class.foldLeft(Inde ptimized.scala:66)
- at scala.collection.mutable.ArrayBuffer.foldLeft(ArrayBuffer.scala:48)
- at com.datio.kirby.CheckFlow$class.applyTransformations(CheckFlow.scala:42)
- at com.datio.kirby.Launcher$.applyTransformations(Launcher.scala:23)
- at com.datio.kirby.Launcher$$anonfun$1.apply$mcV$sp(Launcher.scala:73)
- at com.datio.kirby.Launcher$$anonfun$1.apply(Launcher.scala:60)
- at com.datio.kirby.Launcher$$anonfun$1.apply(Launcher.scala:60)
- at scala.util.Try$.apply(Try.scala:192)
- at com.datio.kirby.Launcher$.runProcess(Launcher.scala:60)
- at com.bbva.cib.cirby.CIBLauncherTrait$class.runProcess(CIBLauncher.scala:30)
- at com.bbva.cib.cirby.CIBLauncher$.runProcess(CIBLauncher.scala:44)
- at com.datio.spark.SparkLauncher$class.runTask(SparkLauncher.scala:65)
- at com.bbva.cib.cirby.CIBLauncher$.runTask(CIBLauncher.scala:44)
- at com.datio.spark.InitSpark$class.main(InitSpark.scala:19)
- at com.bbva.cib.cirby.CIBLauncher$.main(CIBLauncher.scala:44)
- at com.bbva.cib.cirby.CIBLauncher.main(CIBLauncher.scala)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:809)
- at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:182)
- at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:207)
- at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
- at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
- Caused by: java.lang.RuntimeException: Malformed CSV record
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser.org$apache$spark$sql$e n$datasources$csv$UnivocityParser$$convert(UnivocityParser.scala:213)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser.parse(UnivocityParser.scala:191)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:308)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:308)
- at org.apache.spark.sql.e n.datasources.FailureSafeParser.parse(FailureSafeParser.scala:60)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$parseIterator$1.apply(UnivocityParser.scala:312)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$parseIterator$1.apply(UnivocityParser.scala:312)
- at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
- at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at org.apache.spark.sql.e n.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:105)
- at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
- at org.apache.spark.sql.e n.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
- at org.apache.spark.sql.e n.WholeStageCodegenE onfun$8$$anon$1.hasNext(WholeStageCodegenE la:395)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1793)
- at org.apache.spark.rdd.ZippedWithIndexRDD$$anonfun$2.apply(ZippedWithIndexRDD.scala:52)
- at org.apache.spark.rdd.ZippedWithIndexRDD$$anonfun$2.apply(ZippedWithIndexRDD.scala:52)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2070)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2070)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.e .E $TaskRunner.run(E .scala:338)
- at java.util.concurrent.ThreadPoolE .runWorker(ThreadPoolE .java:1149)
- at java.util.concurrent.ThreadPoolE $Worker.run(ThreadPoolE .java:624)
- at java.lang.Thread.run(Thread.java:748)
- 2019-04-24T12:17:55.378+0000 SCALA 20190424121725g1lCs xcsf-pro processing xcsf-gl-krb-inr-ingresosefectivo-01 gl WARN com.datio.kirby.Launcher$: Input Args: Config(SimpleConfigObject({"InputCsvRegexReplaced":{},"kirby":{"input":{"class":"com.bbva.cib.core.input.csvregexreplaced.CsvRegexReplacedInput","encoding":"UTF-16","options":{"delimiter":";","header":true},"paths":["/in/staging/ratransmit/ptcib_es/endt/ENDT_D02_20190423_INGRESOSEFECTIVO.dat"],"regex":[{"regex2Find":"\\.","regex2Replace":""}],"schema":{"path":"/data/raw/xcsf/endt/schemas/current/t_endt_ingresos_en_efectivo.json"},"tmpPath":"/in/staging/ratransmit/ptcib_es/endt/tmp/","type":"custom","withoutSpark":true},"output":{"mode":"append","partition":["odate_date"],"path":"/data/raw/xcsf/endt/data/t_endt_cash_receipts","repartition":{"partitions":10},"schema":{"path":"/data/raw/xcsf/endt/schemas/current/t_endt_ingresos_en_efectivo.json"},"type":"avro"},"transformations":[{"class":"com.bbva.cib.core.transformations.row.offset.OffsetTransformation","type":"custom"},{"auditIdValue":"XCSF5000_GTB_NBO_T_ENDT_INGRESOS_EN_EFECTIVO_0022D","class":"com.bbva.cib.core.transformation.row.audit.AuditRawTransformation","type":"custom"},{"default":"20190423","defaultType":"string","field":"odate_date","type":"literal"},{"field":"personal_id","trimType":"right","type":"trim"},{"class":"com.bbva.cib.core.transformations.row.replace.ReplaceTransformation","columnPattern":".*_(per|amount)","findPattern":",","replacePattern":"\\.","type":"custom"}]},"sparkMetrics":{"listeners":["default"],"output":[{"auth":true,"disable-hostname-verification":true, ,"type":"rest","url":"https://monitoring-kibana-1.live01.daas.gl.igrupobbva:443/monitoring/argos-api/metrics", type":"console"}],"streaming_listeners":["default"],"streaming_output":[{"auth":true,"disable-hostname-verification":true, ,"type":"rest","url":"https://monitoring-kibana-1.live01.daas.gl.igrupobbva:443/monitoring/argos-api/metrics", type":"console"}]}}))
- 2019-04-24T12:17:55.380+0000 SCALA 20190424121725g1lCs xcsf-pro processing xcsf-gl-krb-inr-ingresosefectivo-01 gl ERROR com.datio.kirby.Launcher$: Exception: com.datio.kirby.api.exceptions.KirbyException: 116 - Transformation Error: Fatal error in the transformations
- at com.datio.kirby.CheckFlow$class.applyTransformations(CheckFlow.scala:56)
- at com.datio.kirby.Launcher$.applyTransformations(Launcher.scala:23)
- at com.datio.kirby.Launcher$$anonfun$1.apply$mcV$sp(Launcher.scala:73)
- at com.datio.kirby.Launcher$$anonfun$1.apply(Launcher.scala:60)
- at com.datio.kirby.Launcher$$anonfun$1.apply(Launcher.scala:60)
- at scala.util.Try$.apply(Try.scala:192)
- at com.datio.kirby.Launcher$.runProcess(Launcher.scala:60)
- at com.bbva.cib.cirby.CIBLauncherTrait$class.runProcess(CIBLauncher.scala:30)
- at com.bbva.cib.cirby.CIBLauncher$.runProcess(CIBLauncher.scala:44)
- at com.datio.spark.SparkLauncher$class.runTask(SparkLauncher.scala:65)
- at com.bbva.cib.cirby.CIBLauncher$.runTask(CIBLauncher.scala:44)
- at com.datio.spark.InitSpark$class.main(InitSpark.scala:19)
- at com.bbva.cib.cirby.CIBLauncher$.main(CIBLauncher.scala:44)
- at com.bbva.cib.cirby.CIBLauncher.main(CIBLauncher.scala)
- at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
- at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
- at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
- at java.lang.reflect.Method.invoke(Method.java:498)
- at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:809)
- at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:182)
- at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:207)
- at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
- at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
- Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 5 in stage 0.0 failed 4 times, most recent failure: Lost task 5.3 in stage 0.0 (TID 9, 192.168.192.107, e 1): java.lang.RuntimeException: Malformed CSV record
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser.org$apache$spark$sql$e n$datasources$csv$UnivocityParser$$convert(UnivocityParser.scala:213)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser.parse(UnivocityParser.scala:191)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:308)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:308)
- at org.apache.spark.sql.e n.datasources.FailureSafeParser.parse(FailureSafeParser.scala:60)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$parseIterator$1.apply(UnivocityParser.scala:312)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$parseIterator$1.apply(UnivocityParser.scala:312)
- at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
- at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at org.apache.spark.sql.e n.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:105)
- at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
- at org.apache.spark.sql.e n.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
- at org.apache.spark.sql.e n.WholeStageCodegenE onfun$8$$anon$1.hasNext(WholeStageCodegenE la:395)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1793)
- at org.apache.spark.rdd.ZippedWithIndexRDD$$anonfun$2.apply(ZippedWithIndexRDD.scala:52)
- at org.apache.spark.rdd.ZippedWithIndexRDD$$anonfun$2.apply(ZippedWithIndexRDD.scala:52)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2070)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2070)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.e .E $TaskRunner.run(E .scala:338)
- at java.util.concurrent.ThreadPoolE .runWorker(ThreadPoolE .java:1149)
- at java.util.concurrent.ThreadPoolE $Worker.run(ThreadPoolE .java:624)
- at java.lang.Thread.run(Thread.java:748)
- Driver stacktrace:
- at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1517)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1505)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1504)
- at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
- at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1504)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)
- at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)
- at scala.Option.foreach(Option.scala:257)
- at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:814)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1732)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1687)
- at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1676)
- at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
- at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:630)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2030)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2051)
- at org.apache.spark.SparkContext.runJob(SparkContext.scala:2070)
- at org.apache.spark.rdd.ZippedWithIndexRDD.(ZippedWithIndexRDD.scala:50)
- at org.apache.spark.rdd.RDD$$anonfun$zipWithIndex$1.apply(RDD.scala:1294)
- at org.apache.spark.rdd.RDD$$anonfun$zipWithIndex$1.apply(RDD.scala:1294)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
- at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
- at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
- at org.apache.spark.rdd.RDD.zipWithIndex(RDD.scala:1293)
- at com.bbva.cib.core.transformations.row.offset.OffsetTransformation.transform(OffsetTransformation.scala:56)
- at com.datio.kirby.CheckFlow$$anonfun$applyTransformations$1.apply(CheckFlow.scala:48)
- at com.datio.kirby.CheckFlow$$anonfun$applyTransformations$1.apply(CheckFlow.scala:42)
- at scala.collection.Inde ptimized$class.foldl(Inde ptimized.scala:57)
- at scala.collection.Inde ptimized$class.foldLeft(Inde ptimized.scala:66)
- at scala.collection.mutable.ArrayBuffer.foldLeft(ArrayBuffer.scala:48)
- at com.datio.kirby.CheckFlow$class.applyTransformations(CheckFlow.scala:42)
- ... 22 more
- Caused by: java.lang.RuntimeException: Malformed CSV record
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser.org$apache$spark$sql$e n$datasources$csv$UnivocityParser$$convert(UnivocityParser.scala:213)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser.parse(UnivocityParser.scala:191)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:308)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:308)
- at org.apache.spark.sql.e n.datasources.FailureSafeParser.parse(FailureSafeParser.scala:60)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$parseIterator$1.apply(UnivocityParser.scala:312)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$parseIterator$1.apply(UnivocityParser.scala:312)
- at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
- at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at org.apache.spark.sql.e n.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:105)
- at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
- at org.apache.spark.sql.e n.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
- at org.apache.spark.sql.e n.WholeStageCodegenE onfun$8$$anon$1.hasNext(WholeStageCodegenE la:395)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1793)
- at org.apache.spark.rdd.ZippedWithIndexRDD$$anonfun$2.apply(ZippedWithIndexRDD.scala:52)
- at org.apache.spark.rdd.ZippedWithIndexRDD$$anonfun$2.apply(ZippedWithIndexRDD.scala:52)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2070)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2070)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.e .E $TaskRunner.run(E .scala:338)
- at java.util.concurrent.ThreadPoolE .runWorker(ThreadPoolE .java:1149)
- at java.util.concurrent.ThreadPoolE $Worker.run(ThreadPoolE .java:624)
- at java.lang.Thread.run(Thread.java:748)
- 2019-04-24T12:17:55.380+0000 SCALA 20190424121725g1lCs xcsf-pro processing xcsf-gl-krb-inr-ingresosefectivo-01 gl ERROR com.datio.spark.SparkLauncher$class: Exception message: 116 - Transformation Error: Fatal error in the transformations
- 2019-04-24T12:17:55.381+0000 SCALA 20190424121725g1lCs xcsf-pro processing xcsf-gl-krb-inr-ingresosefectivo-01 gl ERROR com.datio.spark.SparkLauncher$class: Exception cause: Job aborted due to stage failure: Task 5 in stage 0.0 failed 4 times, most recent failure: Lost task 5.3 in stage 0.0 (TID 9, 192.168.192.107, e 1): java.lang.RuntimeException: Malformed CSV record
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser.org$apache$spark$sql$e n$datasources$csv$UnivocityParser$$convert(UnivocityParser.scala:213)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser.parse(UnivocityParser.scala:191)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:308)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:308)
- at org.apache.spark.sql.e n.datasources.FailureSafeParser.parse(FailureSafeParser.scala:60)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$parseIterator$1.apply(UnivocityParser.scala:312)
- at org.apache.spark.sql.e n.datasources.csv.UnivocityParser$$anonfun$parseIterator$1.apply(UnivocityParser.scala:312)
- at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
- at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at org.apache.spark.sql.e n.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:105)
- at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
- at org.apache.spark.sql.e n.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
- at org.apache.spark.sql.e n.WholeStageCodegenE onfun$8$$anon$1.hasNext(WholeStageCodegenE la:395)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
- at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1793)
- at org.apache.spark.rdd.ZippedWithIndexRDD$$anonfun$2.apply(ZippedWithIndexRDD.scala:52)
- at org.apache.spark.rdd.ZippedWithIndexRDD$$anonfun$2.apply(ZippedWithIndexRDD.scala:52)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2070)
- at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2070)
- at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
- at org.apache.spark.scheduler.Task.run(Task.scala:108)
- at org.apache.spark.e .E $TaskRunner.run(E .scala:338)
- at java.util.concurrent.ThreadPoolE .runWorker(ThreadPoolE .java:1149)
- at java.util.concurrent.ThreadPoolE $Worker.run(ThreadPoolE .java:624)
- at java.lang.Thread.run(Thread.java:748)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement