See <https://builds.apache.org/job/carbondata-master-spark-2.1/org.apache.carbondata$carbondata-examples-spark2/2779/display/redirect>
------------------------------------------ [...truncated 504.92 KB...] at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1526) at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186) at org.scalatest.Suite$class.runNestedSuites(Suite.scala:1526) at org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:29) at org.scalatest.Suite$class.run(Suite.scala:1421) at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:29) at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55) at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2563) at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2557) at scala.collection.immutable.List.foreach(List.scala:381) at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:2557) at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1044) at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1043) at org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:2722) at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043) at org.scalatest.tools.Runner$.main(Runner.scala:860) at org.scalatest.tools.Runner.main(Runner.scala) 18/08/07 08:07:02 AUDIT CarbonLoadDataCommand: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Dataload failure for default.timeseriestable_agg0_hour. Please check the logs 18/08/07 08:07:02 ERROR CarbonLoadDataCommand: ScalaTest-main-running-RunExamples java.lang.Exception: DataLoad failure: /tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/04/temp_shuffle_6e450b30-f165-480a-9134-3dce366217ce (No such file or directory) at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:485) at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.loadData(CarbonLoadDataCommand.scala:585) at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:310) at org.apache.spark.sql.execution.command.preaaggregate.PreAggregateUtil$.startDataLoadForDataMap(PreAggregateUtil.scala:604) at org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$$anonfun$onEvent$10.apply(PreAggregateListeners.scala:488) at org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$$anonfun$onEvent$10.apply(PreAggregateListeners.scala:463) at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732) at org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$.onEvent(PreAggregateListeners.scala:463) at org.apache.carbondata.events.OperationListenerBus.fireEvent(OperationListenerBus.java:83) at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:535) at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.loadData(CarbonLoadDataCommand.scala:585) at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:310) at org.apache.spark.sql.execution.command.AtomicRunnableCommand.run(package.scala:92) at org.apache.spark.sql.CarbonDataFrameWriter.loadDataFrame(CarbonDataFrameWriter.scala:61) at org.apache.spark.sql.CarbonDataFrameWriter.writeToCarbonFile(CarbonDataFrameWriter.scala:45) at org.apache.spark.sql.CarbonDataFrameWriter.appendToCarbonFile(CarbonDataFrameWriter.scala:40) at org.apache.spark.sql.CarbonSource.createRelation(CarbonSource.scala:115) at org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:426) at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:215) at org.apache.carbondata.examples.TimeSeriesPreAggregateTableExample$.exampleBody(TimeSeriesPreAggregateTableExample.scala:85) at org.apache.carbondata.examplesCI.RunExamples$$anonfun$15.apply$mcV$sp(RunExamples.scala:106) at org.apache.carbondata.examplesCI.RunExamples$$anonfun$15.apply(RunExamples.scala:106) at org.apache.carbondata.examplesCI.RunExamples$$anonfun$15.apply(RunExamples.scala:106) at org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22) at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85) at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104) at org.scalatest.Transformer.apply(Transformer.scala:22) at org.scalatest.Transformer.apply(Transformer.scala:20) at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:166) at org.apache.spark.sql.test.util.CarbonFunSuite.withFixture(CarbonFunSuite.scala:41) at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:163) at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175) at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175) at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306) at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:175) at org.scalatest.FunSuite.runTest(FunSuite.scala:1555) at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208) at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208) at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413) at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401) at scala.collection.immutable.List.foreach(List.scala:381) at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401) at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396) at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483) at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:208) at org.scalatest.FunSuite.runTests(FunSuite.scala:1555) at org.scalatest.Suite$class.run(Suite.scala:1424) at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1555) at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212) at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212) at org.scalatest.SuperEngine.runImpl(Engine.scala:545) at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:212) at org.apache.carbondata.examplesCI.RunExamples.org$scalatest$BeforeAndAfterAll$$super$run(RunExamples.scala:31) at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:257) at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:256) at org.apache.carbondata.examplesCI.RunExamples.run(RunExamples.scala:31) at org.scalatest.Suite$class.callExecuteOnSuite$1(Suite.scala:1492) at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1528) at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1526) at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186) at org.scalatest.Suite$class.runNestedSuites(Suite.scala:1526) at org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:29) at org.scalatest.Suite$class.run(Suite.scala:1421) at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:29) at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55) at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2563) at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2557) at scala.collection.immutable.List.foreach(List.scala:381) at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:2557) at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1044) at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1043) at org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:2722) at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043) at org.scalatest.tools.Runner$.main(Runner.scala:860) at org.scalatest.tools.Runner.main(Runner.scala) 18/08/07 08:07:02 AUDIT CarbonLoadDataCommand: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Dataload failure for default.timeSeriesTable. Please check the logs 18/08/07 08:07:02 ERROR CarbonLoadDataCommand: ScalaTest-main-running-RunExamples Got exception java.lang.Exception: DataLoad failure: /tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/04/temp_shuffle_6e450b30-f165-480a-9134-3dce366217ce (No such file or directory) when processing data. But this command does not support undo yet, skipping the undo part. [31m- TimeSeriesPreAggregateTableExample *** FAILED ***[0m [31m java.lang.Exception: DataLoad failure: /tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/04/temp_shuffle_6e450b30-f165-480a-9134-3dce366217ce (No such file or directory)[0m [31m at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:485)[0m [31m at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.loadData(CarbonLoadDataCommand.scala:585)[0m [31m at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:310)[0m [31m at org.apache.spark.sql.execution.command.preaaggregate.PreAggregateUtil$.startDataLoadForDataMap(PreAggregateUtil.scala:604)[0m [31m at org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$$anonfun$onEvent$10.apply(PreAggregateListeners.scala:488)[0m [31m at org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$$anonfun$onEvent$10.apply(PreAggregateListeners.scala:463)[0m [31m at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)[0m [31m at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)[0m [31m at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)[0m [31m at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)[0m [31m ...[0m 18/08/07 08:07:02 AUDIT CarbonCreateTableCommand: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Creating Table with Database name [default] and Table name [persontable] 18/08/07 08:07:02 AUDIT CarbonCreateTableCommand: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Table created with Database name [default] and Table name [persontable] 18/08/07 08:07:02 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.persontable 18/08/07 08:07:04 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.persontable 18/08/07 08:07:04 AUDIT MergeIndexEventListener: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index 18/08/07 08:07:04 AUDIT CarbonCreateDataMapCommand: [asf928.gq1.ygridcore.net][jenkins][Thread-1]DataMap dm successfully added 18/08/07 08:07:07 ERROR DiskBlockObjectWriter: Uncaught exception while reverting partial writes to file /tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/28/temp_shuffle_1ea1096d-3ab8-49a2-a6b8-58694202b001 java.io.FileNotFoundException: /tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/28/temp_shuffle_1ea1096d-3ab8-49a2-a6b8-58694202b001 (No such file or directory) at java.io.FileOutputStream.open0(Native Method) at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) at org.apache.spark.storage.DiskBlockObjectWriter.revertPartialWritesAndClose(DiskBlockObjectWriter.scala:210) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.stop(BypassMergeSortShuffleWriter.java:238) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:102) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) at org.apache.spark.scheduler.Task.run(Task.scala:99) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 18/08/07 08:07:07 ERROR BypassMergeSortShuffleWriter: Error while deleting file /tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/28/temp_shuffle_1ea1096d-3ab8-49a2-a6b8-58694202b001 18/08/07 08:07:07 ERROR Executor: Exception in task 0.0 in stage 364.0 (TID 2185) java.io.FileNotFoundException: /tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/28/temp_shuffle_1ea1096d-3ab8-49a2-a6b8-58694202b001 (No such file or directory) at java.io.FileOutputStream.open0(Native Method) at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:102) at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:115) at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:229) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:152) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) at org.apache.spark.scheduler.Task.run(Task.scala:99) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 18/08/07 08:07:07 ERROR TaskSetManager: Task 0 in stage 364.0 failed 1 times; aborting job [31m- LuceneDataMapExample *** FAILED ***[0m [31m org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 364.0 failed 1 times, most recent failure: Lost task 0.0 in stage 364.0 (TID 2185, localhost, executor driver): java.io.FileNotFoundException: /tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/28/temp_shuffle_1ea1096d-3ab8-49a2-a6b8-58694202b001 (No such file or directory)[0m [31m at java.io.FileOutputStream.open0(Native Method)[0m [31m at java.io.FileOutputStream.open(FileOutputStream.java:270)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:213)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:102)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:115)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:229)[0m [31m at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:152)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)[0m [31m at org.apache.spark.scheduler.Task.run(Task.scala:99)[0m [31m at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)[0m [31m at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)[0m [31m at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)[0m [31m at java.lang.Thread.run(Thread.java:748)[0m [31m[0m [31mDriver stacktrace:[0m [31m at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1435)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1423)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1422)[0m [31m at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)[0m [31m at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)[0m [31m at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1422)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:802)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:802)[0m [31m at scala.Option.foreach(Option.scala:257)[0m [31m at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:802)[0m [31m ...[0m [31m Cause: java.io.FileNotFoundException: /tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/28/temp_shuffle_1ea1096d-3ab8-49a2-a6b8-58694202b001 (No such file or directory)[0m [31m at java.io.FileOutputStream.open0(Native Method)[0m [31m at java.io.FileOutputStream.open(FileOutputStream.java:270)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:213)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:102)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:115)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:229)[0m [31m at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:152)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)[0m [31m at org.apache.spark.scheduler.Task.run(Task.scala:99)[0m [31m ...[0m 18/08/07 08:07:07 AUDIT CarbonCreateTableCommand: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Creating Table with Database name [default] and Table name [origin_table] 18/08/07 08:07:07 AUDIT CarbonCreateTableCommand: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Table created with Database name [default] and Table name [origin_table] 18/08/07 08:07:07 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.origin_table 18/08/07 08:07:08 ERROR DataLoadExecutor: [Executor task launch worker-2][partitionID:table;queryID:2530439306783283] Data Load is partially success for table origin_table 18/08/07 08:07:08 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.origin_table 18/08/07 08:07:08 AUDIT MergeIndexEventListener: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index 18/08/07 08:07:08 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.origin_table 18/08/07 08:07:08 ERROR DataLoadExecutor: [Executor task launch worker-2][partitionID:table;queryID:2530439686041893] Data Load is partially success for table origin_table 18/08/07 08:07:08 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.origin_table 18/08/07 08:07:08 AUDIT MergeIndexEventListener: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index 18/08/07 08:07:08 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.origin_table 18/08/07 08:07:08 ERROR DataLoadExecutor: [Executor task launch worker-2][partitionID:table;queryID:2530440047447048] Data Load is partially success for table origin_table 18/08/07 08:07:08 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.origin_table 18/08/07 08:07:08 AUDIT MergeIndexEventListener: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index 18/08/07 08:07:08 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.origin_table 18/08/07 08:07:09 ERROR DataLoadExecutor: [Executor task launch worker-2][partitionID:table;queryID:2530440390529754] Data Load is partially success for table origin_table 18/08/07 08:07:09 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.origin_table 18/08/07 08:07:09 AUDIT MergeIndexEventListener: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index +--------+ |count(1)| +--------+ | 40| +--------+ 18/08/07 08:07:09 AUDIT CarbonCreateTableCommand: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Creating Table with Database name [default] and Table name [external_table] 18/08/07 08:07:09 AUDIT CarbonCreateTableCommand: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Table created with Database name [default] and Table name [external_table] +--------+ |count(1)| +--------+ | 40| +--------+ 18/08/07 08:07:09 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.origin_table 18/08/07 08:07:09 ERROR DataLoadExecutor: [Executor task launch worker-2][partitionID:table;queryID:2530440996284400] Data Load is partially success for table origin_table 18/08/07 08:07:09 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.origin_table 18/08/07 08:07:09 AUDIT MergeIndexEventListener: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index 18/08/07 08:07:09 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.origin_table 18/08/07 08:07:10 ERROR DataLoadExecutor: [Executor task launch worker-2][partitionID:table;queryID:2530441369443194] Data Load is partially success for table origin_table 18/08/07 08:07:10 AUDIT CarbonDataRDDFactory$: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.origin_table 18/08/07 08:07:10 AUDIT MergeIndexEventListener: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index +--------+ |count(1)| +--------+ | 60| +--------+ 18/08/07 08:07:10 AUDIT CarbonDropTableCommand: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Deleting table [origin_table] under database [default] 18/08/07 08:07:10 AUDIT CarbonDropTableCommand: [asf928.gq1.ygridcore.net][jenkins][Thread-1]Deleted table [origin_table] under database [default] [32m- ExternalTableExample[0m [36mRun completed in 2 minutes, 40 seconds.[0m [36mTotal number of tests run: 17[0m [36mSuites: completed 2, aborted 0[0m [36mTests: succeeded 15, failed 2, canceled 0, ignored 0, pending 0[0m [31m*** 2 TESTS FAILED ***[0m [JENKINS] Recording test results