See <https://builds.apache.org/job/carbondata-master-spark-2.2/org.apache.carbondata$carbondata-examples-spark2/871/display/redirect>
------------------------------------------ [...truncated 902.19 KB...] at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043) at org.scalatest.tools.Runner$.main(Runner.scala:860) at org.scalatest.tools.Runner.main(Runner.scala) 18/08/07 08:07:36 AUDIT CarbonLoadDataCommand: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Dataload failure for default.timeseriestable_agg0_hour. Please check the logs 18/08/07 08:07:36 ERROR CarbonLoadDataCommand: ScalaTest-main-running-RunExamples java.lang.Exception: DataLoad failure: /tmp/blockmgr-f9d21bb2-92d4-40c7-be53-b1d72b277171/0e/temp_shuffle_f5d70a1c-8017-430a-a394-68946c2d3cf1 (No such file or directory) at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:485) at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.loadData(CarbonLoadDataCommand.scala:585) at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:310) at org.apache.spark.sql.execution.command.preaaggregate.PreAggregateUtil$.startDataLoadForDataMap(PreAggregateUtil.scala:604) at org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$$anonfun$onEvent$10.apply(PreAggregateListeners.scala:488) at org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$$anonfun$onEvent$10.apply(PreAggregateListeners.scala:463) at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732) at org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$.onEvent(PreAggregateListeners.scala:463) at org.apache.carbondata.events.OperationListenerBus.fireEvent(OperationListenerBus.java:83) at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:535) at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.loadData(CarbonLoadDataCommand.scala:585) at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:310) at org.apache.spark.sql.execution.command.AtomicRunnableCommand.run(package.scala:92) at org.apache.spark.sql.CarbonDataFrameWriter.loadDataFrame(CarbonDataFrameWriter.scala:61) at org.apache.spark.sql.CarbonDataFrameWriter.writeToCarbonFile(CarbonDataFrameWriter.scala:45) at org.apache.spark.sql.CarbonDataFrameWriter.appendToCarbonFile(CarbonDataFrameWriter.scala:40) at org.apache.spark.sql.CarbonSource.createRelation(CarbonSource.scala:115) at org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:469) at org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:50) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56) at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:117) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:117) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:138) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:135) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:116) at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:92) at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:92) at org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:609) at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:233) at org.apache.carbondata.examples.TimeSeriesPreAggregateTableExample$.exampleBody(TimeSeriesPreAggregateTableExample.scala:85) at org.apache.carbondata.examplesCI.RunExamples$$anonfun$15.apply$mcV$sp(RunExamples.scala:106) at org.apache.carbondata.examplesCI.RunExamples$$anonfun$15.apply(RunExamples.scala:106) at org.apache.carbondata.examplesCI.RunExamples$$anonfun$15.apply(RunExamples.scala:106) at org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22) at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85) at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104) at org.scalatest.Transformer.apply(Transformer.scala:22) at org.scalatest.Transformer.apply(Transformer.scala:20) at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:166) at org.apache.spark.sql.test.util.CarbonFunSuite.withFixture(CarbonFunSuite.scala:41) at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:163) at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175) at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175) at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306) at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:175) at org.scalatest.FunSuite.runTest(FunSuite.scala:1555) at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208) at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208) at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413) at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401) at scala.collection.immutable.List.foreach(List.scala:381) at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401) at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396) at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483) at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:208) at org.scalatest.FunSuite.runTests(FunSuite.scala:1555) at org.scalatest.Suite$class.run(Suite.scala:1424) at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1555) at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212) at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212) at org.scalatest.SuperEngine.runImpl(Engine.scala:545) at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:212) at org.apache.carbondata.examplesCI.RunExamples.org$scalatest$BeforeAndAfterAll$$super$run(RunExamples.scala:31) at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:257) at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:256) at org.apache.carbondata.examplesCI.RunExamples.run(RunExamples.scala:31) at org.scalatest.Suite$class.callExecuteOnSuite$1(Suite.scala:1492) at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1528) at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1526) at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186) at org.scalatest.Suite$class.runNestedSuites(Suite.scala:1526) at org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:29) at org.scalatest.Suite$class.run(Suite.scala:1421) at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:29) at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55) at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2563) at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2557) at scala.collection.immutable.List.foreach(List.scala:381) at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:2557) at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1044) at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1043) at org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:2722) at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043) at org.scalatest.tools.Runner$.main(Runner.scala:860) at org.scalatest.tools.Runner.main(Runner.scala) 18/08/07 08:07:36 AUDIT CarbonLoadDataCommand: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Dataload failure for default.timeSeriesTable. Please check the logs 18/08/07 08:07:36 ERROR CarbonLoadDataCommand: ScalaTest-main-running-RunExamples Got exception java.lang.Exception: DataLoad failure: /tmp/blockmgr-f9d21bb2-92d4-40c7-be53-b1d72b277171/0e/temp_shuffle_f5d70a1c-8017-430a-a394-68946c2d3cf1 (No such file or directory) when processing data. But this command does not support undo yet, skipping the undo part. [31m- TimeSeriesPreAggregateTableExample *** FAILED ***[0m [31m java.lang.Exception: DataLoad failure: /tmp/blockmgr-f9d21bb2-92d4-40c7-be53-b1d72b277171/0e/temp_shuffle_f5d70a1c-8017-430a-a394-68946c2d3cf1 (No such file or directory)[0m [31m at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:485)[0m [31m at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.loadData(CarbonLoadDataCommand.scala:585)[0m [31m at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:310)[0m [31m at org.apache.spark.sql.execution.command.preaaggregate.PreAggregateUtil$.startDataLoadForDataMap(PreAggregateUtil.scala:604)[0m [31m at org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$$anonfun$onEvent$10.apply(PreAggregateListeners.scala:488)[0m [31m at org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$$anonfun$onEvent$10.apply(PreAggregateListeners.scala:463)[0m [31m at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)[0m [31m at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)[0m [31m at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)[0m [31m at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)[0m [31m ...[0m 18/08/07 08:07:36 AUDIT CarbonCreateTableCommand: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Creating Table with Database name [default] and Table name [persontable] 18/08/07 08:07:36 AUDIT CarbonCreateTableCommand: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Table created with Database name [default] and Table name [persontable] 18/08/07 08:07:36 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.persontable 18/08/07 08:07:38 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.persontable 18/08/07 08:07:38 AUDIT MergeIndexEventListener: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index 18/08/07 08:07:38 AUDIT CarbonCreateDataMapCommand: [asf930.gq1.ygridcore.net][jenkins][Thread-1]DataMap dm successfully added 18/08/07 08:07:40 ERROR DiskBlockObjectWriter: Uncaught exception while reverting partial writes to file /tmp/blockmgr-f9d21bb2-92d4-40c7-be53-b1d72b277171/3f/temp_shuffle_e51dfb34-0e2c-4514-ac70-41eca1bade11 java.io.FileNotFoundException: /tmp/blockmgr-f9d21bb2-92d4-40c7-be53-b1d72b277171/3f/temp_shuffle_e51dfb34-0e2c-4514-ac70-41eca1bade11 (No such file or directory) at java.io.FileOutputStream.open0(Native Method) at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) at org.apache.spark.storage.DiskBlockObjectWriter$$anonfun$revertPartialWritesAndClose$2.apply$mcV$sp(DiskBlockObjectWriter.scala:217) at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1346) at org.apache.spark.storage.DiskBlockObjectWriter.revertPartialWritesAndClose(DiskBlockObjectWriter.scala:214) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.stop(BypassMergeSortShuffleWriter.java:237) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:102) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) at org.apache.spark.scheduler.Task.run(Task.scala:108) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 18/08/07 08:07:40 ERROR BypassMergeSortShuffleWriter: Error while deleting file /tmp/blockmgr-f9d21bb2-92d4-40c7-be53-b1d72b277171/3f/temp_shuffle_e51dfb34-0e2c-4514-ac70-41eca1bade11 18/08/07 08:07:40 ERROR Executor: Exception in task 0.0 in stage 170.0 (TID 388) java.io.FileNotFoundException: /tmp/blockmgr-f9d21bb2-92d4-40c7-be53-b1d72b277171/3f/temp_shuffle_e51dfb34-0e2c-4514-ac70-41eca1bade11 (No such file or directory) at java.io.FileOutputStream.open0(Native Method) at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:103) at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:116) at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:237) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:151) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) at org.apache.spark.scheduler.Task.run(Task.scala:108) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 18/08/07 08:07:40 ERROR TaskSetManager: Task 0 in stage 170.0 failed 1 times; aborting job [31m- LuceneDataMapExample *** FAILED ***[0m [31m org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 170.0 failed 1 times, most recent failure: Lost task 0.0 in stage 170.0 (TID 388, localhost, executor driver): java.io.FileNotFoundException: /tmp/blockmgr-f9d21bb2-92d4-40c7-be53-b1d72b277171/3f/temp_shuffle_e51dfb34-0e2c-4514-ac70-41eca1bade11 (No such file or directory)[0m [31m at java.io.FileOutputStream.open0(Native Method)[0m [31m at java.io.FileOutputStream.open(FileOutputStream.java:270)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:213)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:103)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:116)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:237)[0m [31m at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:151)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)[0m [31m at org.apache.spark.scheduler.Task.run(Task.scala:108)[0m [31m at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338)[0m [31m at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)[0m [31m at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)[0m [31m at java.lang.Thread.run(Thread.java:748)[0m [31m[0m [31mDriver stacktrace:[0m [31m at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1517)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1505)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1504)[0m [31m at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)[0m [31m at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)[0m [31m at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1504)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)[0m [31m at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:814)[0m [31m at scala.Option.foreach(Option.scala:257)[0m [31m at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:814)[0m [31m ...[0m [31m Cause: java.io.FileNotFoundException: /tmp/blockmgr-f9d21bb2-92d4-40c7-be53-b1d72b277171/3f/temp_shuffle_e51dfb34-0e2c-4514-ac70-41eca1bade11 (No such file or directory)[0m [31m at java.io.FileOutputStream.open0(Native Method)[0m [31m at java.io.FileOutputStream.open(FileOutputStream.java:270)[0m [31m at java.io.FileOutputStream.<init>(FileOutputStream.java:213)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:103)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:116)[0m [31m at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:237)[0m [31m at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:151)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)[0m [31m at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)[0m [31m at org.apache.spark.scheduler.Task.run(Task.scala:108)[0m [31m ...[0m 18/08/07 08:07:40 AUDIT CarbonCreateTableCommand: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Creating Table with Database name [default] and Table name [origin_table] 18/08/07 08:07:40 AUDIT CarbonCreateTableCommand: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Table created with Database name [default] and Table name [origin_table] 18/08/07 08:07:41 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.origin_table 18/08/07 08:07:41 ERROR DataLoadExecutor: [Executor task launch worker for task 389][partitionID:table;queryID:2530312221156261] Data Load is partially success for table origin_table 18/08/07 08:07:41 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.origin_table 18/08/07 08:07:41 AUDIT MergeIndexEventListener: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index 18/08/07 08:07:41 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.origin_table 18/08/07 08:07:41 ERROR DataLoadExecutor: [Executor task launch worker for task 391][partitionID:table;queryID:2530312536634430] Data Load is partially success for table origin_table 18/08/07 08:07:41 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.origin_table 18/08/07 08:07:41 AUDIT MergeIndexEventListener: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index 18/08/07 08:07:41 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.origin_table 18/08/07 08:07:42 ERROR DataLoadExecutor: [Executor task launch worker for task 393][partitionID:table;queryID:2530312913464497] Data Load is partially success for table origin_table 18/08/07 08:07:42 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.origin_table 18/08/07 08:07:42 AUDIT MergeIndexEventListener: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index 18/08/07 08:07:42 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.origin_table 18/08/07 08:07:42 ERROR DataLoadExecutor: [Executor task launch worker for task 395][partitionID:table;queryID:2530313268041752] Data Load is partially success for table origin_table 18/08/07 08:07:42 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.origin_table 18/08/07 08:07:42 AUDIT MergeIndexEventListener: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index +--------+ |count(1)| +--------+ | 40| +--------+ 18/08/07 08:07:42 AUDIT CarbonCreateTableCommand: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Creating Table with Database name [default] and Table name [external_table] 18/08/07 08:07:42 AUDIT CarbonCreateTableCommand: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Table created with Database name [default] and Table name [external_table] +--------+ |count(1)| +--------+ | 40| +--------+ 18/08/07 08:07:42 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.origin_table 18/08/07 08:07:43 ERROR DataLoadExecutor: [Executor task launch worker for task 401][partitionID:table;queryID:2530313949122406] Data Load is partially success for table origin_table 18/08/07 08:07:43 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.origin_table 18/08/07 08:07:43 AUDIT MergeIndexEventListener: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index 18/08/07 08:07:43 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been received for table default.origin_table 18/08/07 08:07:43 ERROR DataLoadExecutor: [Executor task launch worker for task 403][partitionID:table;queryID:2530314315634043] Data Load is partially success for table origin_table 18/08/07 08:07:43 AUDIT CarbonDataRDDFactory$: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for default.origin_table 18/08/07 08:07:43 AUDIT MergeIndexEventListener: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener called for merge index +--------+ |count(1)| +--------+ | 60| +--------+ 18/08/07 08:07:43 AUDIT CarbonDropTableCommand: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Deleting table [origin_table] under database [default] 18/08/07 08:07:43 AUDIT CarbonDropTableCommand: [asf930.gq1.ygridcore.net][jenkins][Thread-1]Deleted table [origin_table] under database [default] [32m- ExternalTableExample[0m [36mRun completed in 1 minute, 45 seconds.[0m [36mTotal number of tests run: 17[0m [36mSuites: completed 2, aborted 0[0m [36mTests: succeeded 11, failed 6, canceled 0, ignored 0, pending 0[0m [31m*** 6 TESTS FAILED ***[0m [JENKINS] Recording test results