See 
<https://builds.apache.org/job/carbondata-master-spark-2.1/org.apache.carbondata$carbondata-examples-spark2/2779/display/redirect>

------------------------------------------
[...truncated 504.92 KB...]
        at 
org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1526)
        at 
scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
        at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
        at org.scalatest.Suite$class.runNestedSuites(Suite.scala:1526)
        at 
org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:29)
        at org.scalatest.Suite$class.run(Suite.scala:1421)
        at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:29)
        at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55)
        at 
org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2563)
        at 
org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2557)
        at scala.collection.immutable.List.foreach(List.scala:381)
        at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:2557)
        at 
org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1044)
        at 
org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1043)
        at 
org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:2722)
        at 
org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043)
        at org.scalatest.tools.Runner$.main(Runner.scala:860)
        at org.scalatest.tools.Runner.main(Runner.scala)
18/08/07 08:07:02 AUDIT CarbonLoadDataCommand: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Dataload failure for 
default.timeseriestable_agg0_hour. Please check the logs
18/08/07 08:07:02 ERROR CarbonLoadDataCommand: 
ScalaTest-main-running-RunExamples 
java.lang.Exception: DataLoad failure: 
/tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/04/temp_shuffle_6e450b30-f165-480a-9134-3dce366217ce
 (No such file or directory)
        at 
org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:485)
        at 
org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.loadData(CarbonLoadDataCommand.scala:585)
        at 
org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:310)
        at 
org.apache.spark.sql.execution.command.preaaggregate.PreAggregateUtil$.startDataLoadForDataMap(PreAggregateUtil.scala:604)
        at 
org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$$anonfun$onEvent$10.apply(PreAggregateListeners.scala:488)
        at 
org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$$anonfun$onEvent$10.apply(PreAggregateListeners.scala:463)
        at 
scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
        at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
        at 
scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
        at 
org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$.onEvent(PreAggregateListeners.scala:463)
        at 
org.apache.carbondata.events.OperationListenerBus.fireEvent(OperationListenerBus.java:83)
        at 
org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:535)
        at 
org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.loadData(CarbonLoadDataCommand.scala:585)
        at 
org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:310)
        at 
org.apache.spark.sql.execution.command.AtomicRunnableCommand.run(package.scala:92)
        at 
org.apache.spark.sql.CarbonDataFrameWriter.loadDataFrame(CarbonDataFrameWriter.scala:61)
        at 
org.apache.spark.sql.CarbonDataFrameWriter.writeToCarbonFile(CarbonDataFrameWriter.scala:45)
        at 
org.apache.spark.sql.CarbonDataFrameWriter.appendToCarbonFile(CarbonDataFrameWriter.scala:40)
        at 
org.apache.spark.sql.CarbonSource.createRelation(CarbonSource.scala:115)
        at 
org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:426)
        at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:215)
        at 
org.apache.carbondata.examples.TimeSeriesPreAggregateTableExample$.exampleBody(TimeSeriesPreAggregateTableExample.scala:85)
        at 
org.apache.carbondata.examplesCI.RunExamples$$anonfun$15.apply$mcV$sp(RunExamples.scala:106)
        at 
org.apache.carbondata.examplesCI.RunExamples$$anonfun$15.apply(RunExamples.scala:106)
        at 
org.apache.carbondata.examplesCI.RunExamples$$anonfun$15.apply(RunExamples.scala:106)
        at 
org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22)
        at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
        at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
        at org.scalatest.Transformer.apply(Transformer.scala:22)
        at org.scalatest.Transformer.apply(Transformer.scala:20)
        at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:166)
        at 
org.apache.spark.sql.test.util.CarbonFunSuite.withFixture(CarbonFunSuite.scala:41)
        at 
org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:163)
        at 
org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
        at 
org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
        at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306)
        at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:175)
        at org.scalatest.FunSuite.runTest(FunSuite.scala:1555)
        at 
org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
        at 
org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
        at 
org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413)
        at 
org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401)
        at scala.collection.immutable.List.foreach(List.scala:381)
        at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
        at 
org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396)
        at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483)
        at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:208)
        at org.scalatest.FunSuite.runTests(FunSuite.scala:1555)
        at org.scalatest.Suite$class.run(Suite.scala:1424)
        at 
org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1555)
        at 
org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
        at 
org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
        at org.scalatest.SuperEngine.runImpl(Engine.scala:545)
        at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:212)
        at 
org.apache.carbondata.examplesCI.RunExamples.org$scalatest$BeforeAndAfterAll$$super$run(RunExamples.scala:31)
        at 
org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:257)
        at 
org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:256)
        at 
org.apache.carbondata.examplesCI.RunExamples.run(RunExamples.scala:31)
        at org.scalatest.Suite$class.callExecuteOnSuite$1(Suite.scala:1492)
        at 
org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1528)
        at 
org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1526)
        at 
scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
        at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
        at org.scalatest.Suite$class.runNestedSuites(Suite.scala:1526)
        at 
org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:29)
        at org.scalatest.Suite$class.run(Suite.scala:1421)
        at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:29)
        at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55)
        at 
org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2563)
        at 
org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2557)
        at scala.collection.immutable.List.foreach(List.scala:381)
        at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:2557)
        at 
org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1044)
        at 
org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1043)
        at 
org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:2722)
        at 
org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043)
        at org.scalatest.tools.Runner$.main(Runner.scala:860)
        at org.scalatest.tools.Runner.main(Runner.scala)
18/08/07 08:07:02 AUDIT CarbonLoadDataCommand: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Dataload failure for 
default.timeSeriesTable. Please check the logs
18/08/07 08:07:02 ERROR CarbonLoadDataCommand: 
ScalaTest-main-running-RunExamples Got exception java.lang.Exception: DataLoad 
failure: 
/tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/04/temp_shuffle_6e450b30-f165-480a-9134-3dce366217ce
 (No such file or directory) when processing data. But this command does not 
support undo yet, skipping the undo part.
- TimeSeriesPreAggregateTableExample *** FAILED ***
  java.lang.Exception: DataLoad failure: 
/tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/04/temp_shuffle_6e450b30-f165-480a-9134-3dce366217ce
 (No such file or directory)
  at 
org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:485)
  at 
org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.loadData(CarbonLoadDataCommand.scala:585)
  at 
org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:310)
  at 
org.apache.spark.sql.execution.command.preaaggregate.PreAggregateUtil$.startDataLoadForDataMap(PreAggregateUtil.scala:604)
  at 
org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$$anonfun$onEvent$10.apply(PreAggregateListeners.scala:488)
  at 
org.apache.spark.sql.execution.command.preaaggregate.LoadPostAggregateListener$$anonfun$onEvent$10.apply(PreAggregateListeners.scala:463)
  at 
scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
  at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
  at 
scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
  ...
18/08/07 08:07:02 AUDIT CarbonCreateTableCommand: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Creating Table with Database name 
[default] and Table name [persontable]
18/08/07 08:07:02 AUDIT CarbonCreateTableCommand: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Table created with Database name 
[default] and Table name [persontable]
18/08/07 08:07:02 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been 
received for table default.persontable
18/08/07 08:07:04 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for 
default.persontable
18/08/07 08:07:04 AUDIT MergeIndexEventListener: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener 
called for merge index
18/08/07 08:07:04 AUDIT CarbonCreateDataMapCommand: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]DataMap dm successfully added
18/08/07 08:07:07 ERROR DiskBlockObjectWriter: Uncaught exception while 
reverting partial writes to file 
/tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/28/temp_shuffle_1ea1096d-3ab8-49a2-a6b8-58694202b001
java.io.FileNotFoundException: 
/tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/28/temp_shuffle_1ea1096d-3ab8-49a2-a6b8-58694202b001
 (No such file or directory)
        at java.io.FileOutputStream.open0(Native Method)
        at java.io.FileOutputStream.open(FileOutputStream.java:270)
        at java.io.FileOutputStream.<init>(FileOutputStream.java:213)
        at 
org.apache.spark.storage.DiskBlockObjectWriter.revertPartialWritesAndClose(DiskBlockObjectWriter.scala:210)
        at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.stop(BypassMergeSortShuffleWriter.java:238)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:102)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
        at org.apache.spark.scheduler.Task.run(Task.scala:99)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
18/08/07 08:07:07 ERROR BypassMergeSortShuffleWriter: Error while deleting file 
/tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/28/temp_shuffle_1ea1096d-3ab8-49a2-a6b8-58694202b001
18/08/07 08:07:07 ERROR Executor: Exception in task 0.0 in stage 364.0 (TID 
2185)
java.io.FileNotFoundException: 
/tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/28/temp_shuffle_1ea1096d-3ab8-49a2-a6b8-58694202b001
 (No such file or directory)
        at java.io.FileOutputStream.open0(Native Method)
        at java.io.FileOutputStream.open(FileOutputStream.java:270)
        at java.io.FileOutputStream.<init>(FileOutputStream.java:213)
        at 
org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:102)
        at 
org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:115)
        at 
org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:229)
        at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:152)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
        at org.apache.spark.scheduler.Task.run(Task.scala:99)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
18/08/07 08:07:07 ERROR TaskSetManager: Task 0 in stage 364.0 failed 1 times; 
aborting job
- LuceneDataMapExample *** FAILED ***
  org.apache.spark.SparkException: Job aborted due to stage failure: Task 
0 in stage 364.0 failed 1 times, most recent failure: Lost task 0.0 in stage 
364.0 (TID 2185, localhost, executor driver): java.io.FileNotFoundException: 
/tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/28/temp_shuffle_1ea1096d-3ab8-49a2-a6b8-58694202b001
 (No such file or directory)
   at java.io.FileOutputStream.open0(Native Method)
   at java.io.FileOutputStream.open(FileOutputStream.java:270)
   at java.io.FileOutputStream.<init>(FileOutputStream.java:213)
   at 
org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:102)
   at 
org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:115)
   at 
org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:229)
   at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:152)
   at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
   at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
   at org.apache.spark.scheduler.Task.run(Task.scala:99)
   at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
   at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
   at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
   at java.lang.Thread.run(Thread.java:748)

Driver stacktrace:
  at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1435)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1423)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1422)
  at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
  at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1422)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:802)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:802)
  at scala.Option.foreach(Option.scala:257)
  at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:802)
  ...
  Cause: java.io.FileNotFoundException: 
/tmp/blockmgr-d312673e-22f6-4a22-bf02-f6e66e4587b8/28/temp_shuffle_1ea1096d-3ab8-49a2-a6b8-58694202b001
 (No such file or directory)
  at java.io.FileOutputStream.open0(Native Method)
  at java.io.FileOutputStream.open(FileOutputStream.java:270)
  at java.io.FileOutputStream.<init>(FileOutputStream.java:213)
  at 
org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:102)
  at 
org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:115)
  at 
org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:229)
  at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:152)
  at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
  at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
  at org.apache.spark.scheduler.Task.run(Task.scala:99)
  ...
18/08/07 08:07:07 AUDIT CarbonCreateTableCommand: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Creating Table with Database name 
[default] and Table name [origin_table]
18/08/07 08:07:07 AUDIT CarbonCreateTableCommand: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Table created with Database name 
[default] and Table name [origin_table]
18/08/07 08:07:07 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been 
received for table default.origin_table
18/08/07 08:07:08 ERROR DataLoadExecutor: [Executor task launch 
worker-2][partitionID:table;queryID:2530439306783283] Data Load is partially 
success for table origin_table
18/08/07 08:07:08 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for 
default.origin_table
18/08/07 08:07:08 AUDIT MergeIndexEventListener: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener 
called for merge index
18/08/07 08:07:08 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been 
received for table default.origin_table
18/08/07 08:07:08 ERROR DataLoadExecutor: [Executor task launch 
worker-2][partitionID:table;queryID:2530439686041893] Data Load is partially 
success for table origin_table
18/08/07 08:07:08 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for 
default.origin_table
18/08/07 08:07:08 AUDIT MergeIndexEventListener: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener 
called for merge index
18/08/07 08:07:08 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been 
received for table default.origin_table
18/08/07 08:07:08 ERROR DataLoadExecutor: [Executor task launch 
worker-2][partitionID:table;queryID:2530440047447048] Data Load is partially 
success for table origin_table
18/08/07 08:07:08 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for 
default.origin_table
18/08/07 08:07:08 AUDIT MergeIndexEventListener: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener 
called for merge index
18/08/07 08:07:08 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been 
received for table default.origin_table
18/08/07 08:07:09 ERROR DataLoadExecutor: [Executor task launch 
worker-2][partitionID:table;queryID:2530440390529754] Data Load is partially 
success for table origin_table
18/08/07 08:07:09 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for 
default.origin_table
18/08/07 08:07:09 AUDIT MergeIndexEventListener: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener 
called for merge index
+--------+
|count(1)|
+--------+
|      40|
+--------+

18/08/07 08:07:09 AUDIT CarbonCreateTableCommand: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Creating Table with Database name 
[default] and Table name [external_table]
18/08/07 08:07:09 AUDIT CarbonCreateTableCommand: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Table created with Database name 
[default] and Table name [external_table]
+--------+
|count(1)|
+--------+
|      40|
+--------+

18/08/07 08:07:09 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been 
received for table default.origin_table
18/08/07 08:07:09 ERROR DataLoadExecutor: [Executor task launch 
worker-2][partitionID:table;queryID:2530440996284400] Data Load is partially 
success for table origin_table
18/08/07 08:07:09 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for 
default.origin_table
18/08/07 08:07:09 AUDIT MergeIndexEventListener: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener 
called for merge index
18/08/07 08:07:09 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load request has been 
received for table default.origin_table
18/08/07 08:07:10 ERROR DataLoadExecutor: [Executor task launch 
worker-2][partitionID:table;queryID:2530441369443194] Data Load is partially 
success for table origin_table
18/08/07 08:07:10 AUDIT CarbonDataRDDFactory$: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Data load is successful for 
default.origin_table
18/08/07 08:07:10 AUDIT MergeIndexEventListener: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Load post status event-listener 
called for merge index
+--------+
|count(1)|
+--------+
|      60|
+--------+

18/08/07 08:07:10 AUDIT CarbonDropTableCommand: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Deleting table [origin_table] 
under database [default]
18/08/07 08:07:10 AUDIT CarbonDropTableCommand: 
[asf928.gq1.ygridcore.net][jenkins][Thread-1]Deleted table [origin_table] under 
database [default]
- ExternalTableExample
Run completed in 2 minutes, 40 seconds.
Total number of tests run: 17
Suites: completed 2, aborted 0
Tests: succeeded 15, failed 2, canceled 0, ignored 0, pending 0
*** 2 TESTS FAILED ***
[JENKINS] Recording test results

Reply via email to