[ https://issues.apache.org/jira/browse/CARBONDATA-4049?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Chetan Bhat closed CARBONDATA-4049. ----------------------------------- Resolution: Won't Fix Issue is identified as that of refresh which is handled by Spark. As issue is not a carbon issue its closed. > Sometimes refresh table fails with error "table not found in database" error > ---------------------------------------------------------------------------- > > Key: CARBONDATA-4049 > URL: https://issues.apache.org/jira/browse/CARBONDATA-4049 > Project: CarbonData > Issue Type: Bug > Components: data-query > Affects Versions: 2.1.0 > Environment: Spark 2.4.5 > Reporter: Chetan Bhat > Priority: Minor > > In Carbon 2.1 version user creates a database. > user copies a old version store such as 1.6.1 to HDFS folder of the database > in the In Carbon 2.1 version > In Spark-SQL or beeline the user accesses the database using the use db > command. > Refresh table command is executed on the old version store table and then the > subsequent operations on the table are performed. > Next refresh table command is tried to be executed on another old version > store table . > > Issue : Sometimes refresh table fails with error "table not found in > database" error. > spark-sql> refresh table brinjal_deleteseg; > *Error in query: Table or view 'brinjal_deleteseg' not found in database > '1_6_1';* > > **Log - > 2020-11-12 18:55:46,922 | INFO | [main] | Created broadcast 171 from > broadCastHadoopConf at CarbonRDD.scala:58 | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,922 | INFO | [main] | Created broadcast 171 from > broadCastHadoopConf at CarbonRDD.scala:58 | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,924 | INFO | [main] | Pushed Filters: | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,939 | INFO | [main] | Distributed Index server is enabled for > 1_6_1.brinjal_update | > org.apache.carbondata.core.util.CarbonProperties.isDistributedPruningEnabled(CarbonProperties.java:1742)2020-11-12 > 18:55:46,939 | INFO | [main] | Started block pruning ... | > org.apache.carbondata.hadoop.api.CarbonInputFormat.getPrunedBlocklets(CarbonInputFormat.java:526)2020-11-12 > 18:55:46,940 | INFO | [main] | Distributed Index server is enabled for > 1_6_1.brinjal_update | > org.apache.carbondata.core.util.CarbonProperties.isDistributedPruningEnabled(CarbonProperties.java:1742)2020-11-12 > 18:55:46,945 | INFO | [main] | Successfully Created directory: > hdfs://hacluster/tmp/indexservertmp/4b6353d4-65d7-4856-b3cd-b3bc11d15c55 | > org.apache.carbondata.core.util.CarbonUtil.createTempFolderForIndexServer(CarbonUtil.java:3273)2020-11-12 > 18:55:46,945 | INFO | [main] | Temp folder path for Query ID: > 4b6353d4-65d7-4856-b3cd-b3bc11d15c55 is > org.apache.carbondata.core.datastore.filesystem.HDFSCarbonFile@b8f2e1bf | > org.apache.carbondata.indexserver.DistributedIndexJob.execute(IndexJobs.scala:57)2020-11-12 > 18:55:46,946 | ERROR | [main] | Configured port for index server is not a > valid number | > org.apache.carbondata.core.util.CarbonProperties.getIndexServerPort(CarbonProperties.java:1779)java.lang.NumberFormatException: > null at java.lang.Integer.parseInt(Integer.java:542) at > java.lang.Integer.parseInt(Integer.java:615) at > org.apache.carbondata.core.util.CarbonProperties.getIndexServerPort(CarbonProperties.java:1777) > at > org.apache.carbondata.indexserver.IndexServer$.serverPort$lzycompute(IndexServer.scala:88) > at > org.apache.carbondata.indexserver.IndexServer$.serverPort(IndexServer.scala:88) > at > org.apache.carbondata.indexserver.IndexServer$.getClient(IndexServer.scala:312) > at > org.apache.carbondata.indexserver.IndexServer$.getClient(IndexServer.scala:301) > at > org.apache.carbondata.indexserver.DistributedIndexJob$$anonfun$1.apply(IndexJobs.scala:83) > at > org.apache.carbondata.indexserver.DistributedIndexJob$$anonfun$1.apply(IndexJobs.scala:59) > at > org.apache.carbondata.spark.util.CarbonScalaUtil$.logTime(CarbonScalaUtil.scala:769) > at > org.apache.carbondata.indexserver.DistributedIndexJob.execute(IndexJobs.scala:58) > at > org.apache.carbondata.core.index.IndexUtil.executeIndexJob(IndexUtil.java:304) > at > org.apache.carbondata.hadoop.api.CarbonInputFormat.getDistributedSplit(CarbonInputFormat.java:431) > at > org.apache.carbondata.hadoop.api.CarbonInputFormat.getPrunedBlocklets(CarbonInputFormat.java:532) > at > org.apache.carbondata.hadoop.api.CarbonInputFormat.getDataBlocksOfSegment(CarbonInputFormat.java:477) > at > org.apache.carbondata.hadoop.api.CarbonTableInputFormat.getSplits(CarbonTableInputFormat.java:356) > at > org.apache.carbondata.hadoop.api.CarbonTableInputFormat.getSplits(CarbonTableInputFormat.java:204) > at > org.apache.carbondata.spark.rdd.CarbonScanRDD.internalGetPartitions(CarbonScanRDD.scala:159) > at > org.apache.carbondata.spark.rdd.CarbonRDD.getPartitions(CarbonRDD.scala:68) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:273) at > org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:269) at > scala.Option.getOrElse(Option.scala:121) at > org.apache.spark.rdd.RDD.partitions(RDD.scala:269) at > org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:273) at > org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:269) at > scala.Option.getOrElse(Option.scala:121) at > org.apache.spark.rdd.RDD.partitions(RDD.scala:269) at > org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:273) at > org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:269) at > scala.Option.getOrElse(Option.scala:121) at > org.apache.spark.rdd.RDD.partitions(RDD.scala:269) at > org.apache.spark.SparkContext.runJob(SparkContext.scala:2126) at > org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:990) at > org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > at > org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) > at org.apache.spark.rdd.RDD.withScope(RDD.scala:385) at > org.apache.spark.rdd.RDD.collect(RDD.scala:989) at > org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:299) > at > org.apache.spark.sql.execution.SparkPlan.executeCollectPublic(SparkPlan.scala:326) > at > org.apache.spark.sql.execution.QueryExecution.hiveResultString(QueryExecution.scala:128) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLDriver$$anonfun$run$1.apply(SparkSQLDriver.scala:64) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLDriver$$anonfun$run$1.apply(SparkSQLDriver.scala:64) > at > org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:80) > at > org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:127) > at > org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:75) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:371) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376) at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:274) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) at > org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) > at > org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:845) > at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:161) at > org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:184) at > org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86) at > org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:920) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:929) at > org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)2020-11-12 > 18:55:46,949 | ERROR | [main] | Exception occurred while getting splits using > index server. Initiating Fall back to embedded mode | > org.apache.carbondata.hadoop.api.CarbonInputFormat.getDistributedSplit(CarbonInputFormat.java:438)java.lang.NumberFormatException: > null at java.lang.Integer.parseInt(Integer.java:542) at > java.lang.Integer.parseInt(Integer.java:615) at > org.apache.carbondata.core.util.CarbonProperties.getIndexServerPort(CarbonProperties.java:1777) > at > org.apache.carbondata.indexserver.IndexServer$.serverPort$lzycompute(IndexServer.scala:88) > at > org.apache.carbondata.indexserver.IndexServer$.serverPort(IndexServer.scala:88) > at > org.apache.carbondata.indexserver.IndexServer$.getClient(IndexServer.scala:312) > at > org.apache.carbondata.indexserver.IndexServer$.getClient(IndexServer.scala:301) > at > org.apache.carbondata.indexserver.DistributedIndexJob$$anonfun$1.apply(IndexJobs.scala:83) > at > org.apache.carbondata.indexserver.DistributedIndexJob$$anonfun$1.apply(IndexJobs.scala:59) > at > org.apache.carbondata.spark.util.CarbonScalaUtil$.logTime(CarbonScalaUtil.scala:769) > at > org.apache.carbondata.indexserver.DistributedIndexJob.execute(IndexJobs.scala:58) > at > org.apache.carbondata.core.index.IndexUtil.executeIndexJob(IndexUtil.java:304) > at > org.apache.carbondata.hadoop.api.CarbonInputFormat.getDistributedSplit(CarbonInputFormat.java:431) > at > org.apache.carbondata.hadoop.api.CarbonInputFormat.getPrunedBlocklets(CarbonInputFormat.java:532) > at > org.apache.carbondata.hadoop.api.CarbonInputFormat.getDataBlocksOfSegment(CarbonInputFormat.java:477) > at > org.apache.carbondata.hadoop.api.CarbonTableInputFormat.getSplits(CarbonTableInputFormat.java:356) > at > org.apache.carbondata.hadoop.api.CarbonTableInputFormat.getSplits(CarbonTableInputFormat.java:204) > at > org.apache.carbondata.spark.rdd.CarbonScanRDD.internalGetPartitions(CarbonScanRDD.scala:159) > at > org.apache.carbondata.spark.rdd.CarbonRDD.getPartitions(CarbonRDD.scala:68) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:273) at > org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:269) at > scala.Option.getOrElse(Option.scala:121) at > org.apache.spark.rdd.RDD.partitions(RDD.scala:269) at > org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:273) at > org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:269) at > scala.Option.getOrElse(Option.scala:121) at > org.apache.spark.rdd.RDD.partitions(RDD.scala:269) at > org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:49) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:273) at > org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:269) at > scala.Option.getOrElse(Option.scala:121) at > org.apache.spark.rdd.RDD.partitions(RDD.scala:269) at > org.apache.spark.SparkContext.runJob(SparkContext.scala:2126) at > org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:990) at > org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > at > org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) > at org.apache.spark.rdd.RDD.withScope(RDD.scala:385) at > org.apache.spark.rdd.RDD.collect(RDD.scala:989) at > org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:299) > at > org.apache.spark.sql.execution.SparkPlan.executeCollectPublic(SparkPlan.scala:326) > at > org.apache.spark.sql.execution.QueryExecution.hiveResultString(QueryExecution.scala:128) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLDriver$$anonfun$run$1.apply(SparkSQLDriver.scala:64) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLDriver$$anonfun$run$1.apply(SparkSQLDriver.scala:64) > at > org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:80) > at > org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:127) > at > org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:75) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:371) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376) at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:274) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) at > org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) > at > org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:845) > at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:161) at > org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:184) at > org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86) at > org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:920) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:929) at > org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)2020-11-12 > 18:55:46,951 | INFO | [main] | Block broadcast_172 stored as values in > memory (estimated size 370.9 KB, free 909.1 MB) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,960 | INFO | [main] | Block broadcast_172_piece0 stored as bytes in > memory (estimated size 29.4 KB, free 909.0 MB) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,961 | INFO | [dispatcher-event-loop-1] | Added broadcast_172_piece0 > in memory on vm1:43460 (size: 29.4 KB, free: 912.0 MB) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,961 | INFO | [main] | Created broadcast 172 from > broadCastHadoopConf at CarbonRDD.scala:58 | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,962 | INFO | [main] | Distributed Index server is enabled for > 1_6_1.brinjal_update | > org.apache.carbondata.core.util.CarbonProperties.isDistributedPruningEnabled(CarbonProperties.java:1742)2020-11-12 > 18:55:46,966 | INFO | [main] | Starting job: collect at > IndexServer.scala:178 | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,967 | INFO | [dag-scheduler-event-loop] | Got job 87 (collect at > IndexServer.scala:178) with 1 output partitions | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,967 | INFO | [dag-scheduler-event-loop] | Final stage: ResultStage > 83 (collect at IndexServer.scala:178) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,967 | INFO | [dag-scheduler-event-loop] | Parents of final stage: > List() | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,967 | INFO | [dag-scheduler-event-loop] | Missing parents: List() | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,967 | INFO | [dag-scheduler-event-loop] | Submitting ResultStage 83 > (DistributedPruneRDD[249] at RDD at CarbonRDD.scala:38), which has no missing > parents | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,968 | INFO | [dag-scheduler-event-loop] | Block broadcast_173 > stored as values in memory (estimated size 17.7 KB, free 909.0 MB) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,969 | INFO | [dag-scheduler-event-loop] | Block > broadcast_173_piece0 stored as bytes in memory (estimated size 8.7 KB, free > 909.0 MB) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,970 | INFO | [dispatcher-event-loop-2] | Added broadcast_173_piece0 > in memory on vm1:43460 (size: 8.7 KB, free: 912.0 MB) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,970 | INFO | [dag-scheduler-event-loop] | Created broadcast 173 > from broadcast at DAGScheduler.scala:1163 | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,970 | INFO | [dag-scheduler-event-loop] | Submitting 1 missing > tasks from ResultStage 83 (DistributedPruneRDD[249] at RDD at > CarbonRDD.scala:38) (first 15 tasks are for partitions Vector(0)) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,970 | INFO | [dag-scheduler-event-loop] | Adding task set 83.0 with > 1 tasks | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,971 | INFO | [dispatcher-event-loop-5] | Starting task 0.0 in stage > 83.0 (TID 464, localhost, executor driver, partition 0, PROCESS_LOCAL, 9449 > bytes) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,971 | INFO | [Executor task launch worker for task 464] | Running > task 0.0 in stage 83.0 (TID 464) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,974 | INFO | [Executor task launch worker for task 464] | Value for > carbon.max.executor.threads.for.block.pruning is 4 | > org.apache.carbondata.core.util.CarbonProperties.getNumOfThreadsForExecutorPruning(CarbonProperties.java:1809)2020-11-12 > 18:55:46,995 | INFO | [IndexPruningPool_1605178546974] | Constructing new > SegmentProperties for table: 1_6_1_brinjal_update. Current size of segment > properties holder list is: 2 | > org.apache.carbondata.core.datastore.block.SegmentPropertiesAndSchemaHolder.addSegmentProperties(SegmentPropertiesAndSchemaHolder.java:115)2020-11-12 > 18:55:46,997 | INFO | [IndexPruningPool_1605178546974] | Removed entry from > InMemory lru cache :: > hdfs://hacluster/user/sparkhive/warehouse/1_6_1.db/brinjal_update/Fact/Part0/Segment_5/5_1605178541880.carbonindexmerge > | > org.apache.carbondata.core.cache.CarbonLRUCache.removeKey(CarbonLRUCache.java:189)2020-11-12 > 18:55:46,997 | INFO | [Executor task launch worker for task 464] | Time > taken to collect 1 blocklets : 23 | > org.apache.carbondata.indexserver.DistributedPruneRDD.internalCompute(DistributedPruneRDD.scala:118)2020-11-12 > 18:55:46,998 | INFO | [Executor task launch worker for task 464] | Finished > task 0.0 in stage 83.0 (TID 464). 2475 bytes result sent to driver | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,999 | INFO | [task-result-getter-3] | Finished task 0.0 in stage > 83.0 (TID 464) in 27 ms on localhost (executor driver) (1/1) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,999 | INFO | [task-result-getter-3] | Removed TaskSet 83.0, whose > tasks have all completed, from pool | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,999 | INFO | [dag-scheduler-event-loop] | ResultStage 83 (collect > at IndexServer.scala:178) finished in 0.031 s | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:46,999 | INFO | [main] | Job 87 finished: collect at > IndexServer.scala:178, took 0.032684 s | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,001 | INFO | [main] | Block broadcast_174 stored as values in > memory (estimated size 370.9 KB, free 908.6 MB) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,010 | INFO | [main] | Block broadcast_174_piece0 stored as bytes in > memory (estimated size 29.4 KB, free 908.6 MB) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,011 | INFO | [dispatcher-event-loop-7] | Added broadcast_174_piece0 > in memory on vm1:43460 (size: 29.4 KB, free: 912.0 MB) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,011 | INFO | [main] | Created broadcast 174 from > broadCastHadoopConf at CarbonRDD.scala:58 | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,016 | INFO | [main] | Starting job: collect at > IndexServer.scala:205 | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,017 | INFO | [main] | Job 88 finished: collect at > IndexServer.scala:205, took 0.000032 s | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,017 | INFO | [main] | Finished block pruning ... | > org.apache.carbondata.hadoop.api.CarbonInputFormat.getPrunedBlocklets(CarbonInputFormat.java:622)2020-11-12 > 18:55:47,018 | INFO | [main] | Planning scan with bin packing, max size: > 4194304 bytes, open cost is considered as scanning 4194304 bytes. | > org.apache.carbondata.spark.rdd.CarbonScanRDD.distributeColumnarSplits(CarbonScanRDD.scala:366)2020-11-12 > 18:55:47,018 | INFO | [main] | Identified no.of.blocks: 5, no.of.tasks: 5, > no.of.nodes: 0, parallelism: 8 | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,022 | INFO | [main] | Starting job: processCmd at > CliDriver.java:376 | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,023 | INFO | [dag-scheduler-event-loop] | Got job 89 (processCmd at > CliDriver.java:376) with 5 output partitions | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,023 | INFO | [dag-scheduler-event-loop] | Final stage: ResultStage > 84 (processCmd at CliDriver.java:376) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,023 | INFO | [dag-scheduler-event-loop] | Parents of final stage: > List() | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,023 | INFO | [dag-scheduler-event-loop] | Missing parents: List() | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,023 | INFO | [dag-scheduler-event-loop] | Submitting ResultStage 84 > (MapPartitionsRDD[248] at processCmd at CliDriver.java:376), which has no > missing parents | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,025 | INFO | [dag-scheduler-event-loop] | Block broadcast_175 > stored as values in memory (estimated size 24.3 KB, free 908.6 MB) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,026 | INFO | [dag-scheduler-event-loop] | Block > broadcast_175_piece0 stored as bytes in memory (estimated size 11.6 KB, free > 908.6 MB) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,026 | INFO | [dispatcher-event-loop-6] | Added broadcast_175_piece0 > in memory on vm1:43460 (size: 11.6 KB, free: 912.0 MB) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,026 | INFO | [dag-scheduler-event-loop] | Created broadcast 175 > from broadcast at DAGScheduler.scala:1163 | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,027 | INFO | [dag-scheduler-event-loop] | Submitting 5 missing > tasks from ResultStage 84 (MapPartitionsRDD[248] at processCmd at > CliDriver.java:376) (first 15 tasks are for partitions Vector(0, 1, 2, 3, 4)) > | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,027 | INFO | [dag-scheduler-event-loop] | Adding task set 84.0 with > 5 tasks | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,028 | INFO | [dispatcher-event-loop-1] | Starting task 0.0 in stage > 84.0 (TID 465, localhost, executor driver, partition 0, ANY, 9185 bytes) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,028 | INFO | [dispatcher-event-loop-1] | Starting task 1.0 in stage > 84.0 (TID 466, localhost, executor driver, partition 1, ANY, 9185 bytes) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,028 | INFO | [dispatcher-event-loop-1] | Starting task 2.0 in stage > 84.0 (TID 467, localhost, executor driver, partition 2, ANY, 9185 bytes) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,029 | INFO | [dispatcher-event-loop-1] | Starting task 3.0 in stage > 84.0 (TID 468, localhost, executor driver, partition 3, ANY, 9185 bytes) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,029 | INFO | [dispatcher-event-loop-1] | Starting task 4.0 in stage > 84.0 (TID 469, localhost, executor driver, partition 4, ANY, 9185 bytes) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,029 | INFO | [Executor task launch worker for task 466] | Running > task 1.0 in stage 84.0 (TID 466) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,029 | INFO | [Executor task launch worker for task 467] | Running > task 2.0 in stage 84.0 (TID 467) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,029 | INFO | [Executor task launch worker for task 469] | Running > task 4.0 in stage 84.0 (TID 469) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,029 | INFO | [Executor task launch worker for task 468] | Running > task 3.0 in stage 84.0 (TID 468) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,029 | INFO | [Executor task launch worker for task 465] | Running > task 0.0 in stage 84.0 (TID 465) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,033 | INFO | [Executor task launch worker for task 469] | > Projection Columns: [imei, amsize, channelsid, activecountry, activecity, > productiondate, deliverydate, gamepointid, deviceinformationid, > deliverycharge] | > org.apache.carbondata.core.scan.model.QueryModelBuilder.projectColumns(QueryModelBuilder.java:94)2020-11-12 > 18:55:47,033 | INFO | [Executor task launch worker for task 467] | > Projection Columns: [imei, amsize, channelsid, activecountry, activecity, > productiondate, deliverydate, gamepointid, deviceinformationid, > deliverycharge] | > org.apache.carbondata.core.scan.model.QueryModelBuilder.projectColumns(QueryModelBuilder.java:94)2020-11-12 > 18:55:47,033 | INFO | [Executor task launch worker for task 466] | > Projection Columns: [imei, amsize, channelsid, activecountry, activecity, > productiondate, deliverydate, gamepointid, deviceinformationid, > deliverycharge] | > org.apache.carbondata.core.scan.model.QueryModelBuilder.projectColumns(QueryModelBuilder.java:94)2020-11-12 > 18:55:47,033 | INFO | [Executor task launch worker for task 468] | > Projection Columns: [imei, amsize, channelsid, activecountry, activecity, > productiondate, deliverydate, gamepointid, deviceinformationid, > deliverycharge] | > org.apache.carbondata.core.scan.model.QueryModelBuilder.projectColumns(QueryModelBuilder.java:94)2020-11-12 > 18:55:47,033 | INFO | [Executor task launch worker for task 465] | > Projection Columns: [imei, amsize, channelsid, activecountry, activecity, > productiondate, deliverydate, gamepointid, deviceinformationid, > deliverycharge] | > org.apache.carbondata.core.scan.model.QueryModelBuilder.projectColumns(QueryModelBuilder.java:94)2020-11-12 > 18:55:47,034 | INFO | [Executor task launch worker for task 466] | Query > will be executed on table: brinjal_update | > org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.initQuery(AbstractQueryExecutor.java:122)2020-11-12 > 18:55:47,034 | INFO | [Executor task launch worker for task 467] | Query > will be executed on table: brinjal_update | > org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.initQuery(AbstractQueryExecutor.java:122)2020-11-12 > 18:55:47,034 | INFO | [Executor task launch worker for task 469] | Query > will be executed on table: brinjal_update | > org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.initQuery(AbstractQueryExecutor.java:122)2020-11-12 > 18:55:47,034 | INFO | [Executor task launch worker for task 466] | Query > prefetch is: true | > org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.getBlockExecutionInfoForBlock(AbstractQueryExecutor.java:479)2020-11-12 > 18:55:47,034 | INFO | [Executor task launch worker for task 465] | Query > will be executed on table: brinjal_update | > org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.initQuery(AbstractQueryExecutor.java:122)2020-11-12 > 18:55:47,034 | INFO | [Executor task launch worker for task 469] | Query > prefetch is: true | > org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.getBlockExecutionInfoForBlock(AbstractQueryExecutor.java:479)2020-11-12 > 18:55:47,034 | INFO | [Executor task launch worker for task 465] | Query > prefetch is: true | > org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.getBlockExecutionInfoForBlock(AbstractQueryExecutor.java:479)2020-11-12 > 18:55:47,034 | INFO | [Executor task launch worker for task 468] | Query > will be executed on table: brinjal_update | > org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.initQuery(AbstractQueryExecutor.java:122)2020-11-12 > 18:55:47,034 | INFO | [Executor task launch worker for task 467] | Query > prefetch is: true | > org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.getBlockExecutionInfoForBlock(AbstractQueryExecutor.java:479)2020-11-12 > 18:55:47,035 | INFO | [Executor task launch worker for task 468] | Query > prefetch is: true | > org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.getBlockExecutionInfoForBlock(AbstractQueryExecutor.java:479)2020-11-12 > 18:55:47,036 | INFO | [Executor task launch worker for task 466] | Vector > based dictionary collector is used to scan and collect the data | > org.apache.carbondata.core.scan.collector.ResultCollectorFactory.getScannedResultCollector(ResultCollectorFactory.java:78)2020-11-12 > 18:55:47,036 | INFO | [Executor task launch worker for task 466] | Direct > page-wise vector fill collector is used to scan and collect the data | > org.apache.carbondata.core.scan.collector.impl.DictionaryBasedVectorResultCollector.<init>(DictionaryBasedVectorResultCollector.java:73)2020-11-12 > 18:55:47,037 | INFO | [Executor task launch worker for task 469] | Vector > based dictionary collector is used to scan and collect the data | > org.apache.carbondata.core.scan.collector.ResultCollectorFactory.getScannedResultCollector(ResultCollectorFactory.java:78)2020-11-12 > 18:55:47,037 | INFO | [Executor task launch worker for task 469] | Direct > page-wise vector fill collector is used to scan and collect the data | > org.apache.carbondata.core.scan.collector.impl.DictionaryBasedVectorResultCollector.<init>(DictionaryBasedVectorResultCollector.java:73)2020-11-12 > 18:55:47,037 | INFO | [Executor task launch worker for task 465] | Vector > based dictionary collector is used to scan and collect the data | > org.apache.carbondata.core.scan.collector.ResultCollectorFactory.getScannedResultCollector(ResultCollectorFactory.java:78)2020-11-12 > 18:55:47,037 | INFO | [Executor task launch worker for task 465] | Direct > page-wise vector fill collector is used to scan and collect the data | > org.apache.carbondata.core.scan.collector.impl.DictionaryBasedVectorResultCollector.<init>(DictionaryBasedVectorResultCollector.java:73)2020-11-12 > 18:55:47,037 | INFO | [Executor task launch worker for task 468] | Vector > based dictionary collector is used to scan and collect the data | > org.apache.carbondata.core.scan.collector.ResultCollectorFactory.getScannedResultCollector(ResultCollectorFactory.java:78)2020-11-12 > 18:55:47,037 | INFO | [Executor task launch worker for task 468] | Direct > page-wise vector fill collector is used to scan and collect the data | > org.apache.carbondata.core.scan.collector.impl.DictionaryBasedVectorResultCollector.<init>(DictionaryBasedVectorResultCollector.java:73)2020-11-12 > 18:55:47,037 | INFO | [Executor task launch worker for task 467] | Vector > based dictionary collector is used to scan and collect the data | > org.apache.carbondata.core.scan.collector.ResultCollectorFactory.getScannedResultCollector(ResultCollectorFactory.java:78)2020-11-12 > 18:55:47,038 | INFO | [Executor task launch worker for task 467] | Direct > page-wise vector fill collector is used to scan and collect the data | > org.apache.carbondata.core.scan.collector.impl.DictionaryBasedVectorResultCollector.<init>(DictionaryBasedVectorResultCollector.java:73)2020-11-12 > 18:55:47,050 | INFO | [Executor task launch worker for task 465] | Total > off-heap working memory used after task 645a72eb-8d06-489b-940d-f7ca1e901bc7 > is 128073. Current running tasks are 7a7a3012-c819-4e4c-9f08-54dd890850e4, > f251027c-707b-4df1-b35f-00eb74c8d77f, 5233fa5c-5173-4a9e-a8bf-8c024dac1509, > a05de0f7-398e-45a5-a91d-a4e1045e9a98 | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,050 | INFO | [Executor task launch worker for task 467] | Total > off-heap working memory used after task a05de0f7-398e-45a5-a91d-a4e1045e9a98 > is 128073. Current running tasks are 7a7a3012-c819-4e4c-9f08-54dd890850e4, > f251027c-707b-4df1-b35f-00eb74c8d77f, 5233fa5c-5173-4a9e-a8bf-8c024dac1509 | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,050 | INFO | [Executor task launch worker for task 469] | Total > off-heap working memory used after task f251027c-707b-4df1-b35f-00eb74c8d77f > is 128073. Current running tasks are 7a7a3012-c819-4e4c-9f08-54dd890850e4, > 5233fa5c-5173-4a9e-a8bf-8c024dac1509 | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,050 | INFO | [Executor task launch worker for task 467] | Total > off-heap working memory used after task 11861831-3ae3-4dd7-98ce-352f4416e4d5 > is 128073. Current running tasks are 7a7a3012-c819-4e4c-9f08-54dd890850e4, > 5233fa5c-5173-4a9e-a8bf-8c024dac1509 | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,050 | INFO | [Executor task launch worker for task 469] | Total > off-heap working memory used after task c4193df5-ceb8-4ff9-8844-0d9c42f989d8 > is 128045. Current running tasks are 7a7a3012-c819-4e4c-9f08-54dd890850e4, > 5233fa5c-5173-4a9e-a8bf-8c024dac1509 | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,051 | INFO | [Executor task launch worker for task 467] | Total > off-heap working memory used after task 88e4dbd6-c19e-43e6-bc9b-42dfefc8378c > is 128045. Current running tasks are 7a7a3012-c819-4e4c-9f08-54dd890850e4, > 5233fa5c-5173-4a9e-a8bf-8c024dac1509 | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,051 | INFO | [Executor task launch worker for task 469] | Total > off-heap working memory used after task f75746ea-783d-4098-91d5-1a93fda81f04 > is 128045. Current running tasks are 7a7a3012-c819-4e4c-9f08-54dd890850e4, > 5233fa5c-5173-4a9e-a8bf-8c024dac1509 | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,051 | INFO | [Executor task launch worker for task 465] | Total > off-heap working memory used after task ae4ac11d-7a43-426d-8578-ac77b6ddf74b > is 128090. Current running tasks are 7a7a3012-c819-4e4c-9f08-54dd890850e4, > 5233fa5c-5173-4a9e-a8bf-8c024dac1509 | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,051 | INFO | [Executor task launch worker for task 466] | Total > off-heap working memory used after task 7a7a3012-c819-4e4c-9f08-54dd890850e4 > is 128090. Current running tasks are 5233fa5c-5173-4a9e-a8bf-8c024dac1509 | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,051 | INFO | [Executor task launch worker for task 465] | Total > off-heap working memory used after task fb7aa505-1163-4fe2-936e-0de1e4f78271 > is 128090. Current running tasks are 5233fa5c-5173-4a9e-a8bf-8c024dac1509 | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,051 | INFO | [Executor task launch worker for task 466] | Total > off-heap working memory used after task 27cb623e-3b1c-4f49-adaa-c11c19d766af > is 128000. Current running tasks are 5233fa5c-5173-4a9e-a8bf-8c024dac1509 | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,051 | INFO | [Executor task launch worker for task 466] | Total > off-heap working memory used after task b2d72c4c-7d5d-4030-9f79-54219d56474d > is 0. Current running tasks are 5233fa5c-5173-4a9e-a8bf-8c024dac1509 | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,052 | INFO | [Executor task launch worker for task 465] | Finished > task 0.0 in stage 84.0 (TID 465). 2766 bytes result sent to driver | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,051 | INFO | [Executor task launch worker for task 469] | Finished > task 4.0 in stage 84.0 (TID 469). 2198 bytes result sent to driver | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,051 | INFO | [Executor task launch worker for task 467] | Finished > task 2.0 in stage 84.0 (TID 467). 2444 bytes result sent to driver | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,052 | INFO | [Executor task launch worker for task 466] | Finished > task 1.0 in stage 84.0 (TID 466). 2524 bytes result sent to driver | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,052 | INFO | [Executor task launch worker for task 468] | Total > off-heap working memory used after task 5233fa5c-5173-4a9e-a8bf-8c024dac1509 > is 0. Current running tasks are | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,052 | INFO | [Executor task launch worker for task 468] | Total > off-heap working memory used after task 4ea15806-df6e-4377-9cea-84f07586fe02 > is 0. Current running tasks are | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,052 | INFO | [task-result-getter-2] | Finished task 0.0 in stage > 84.0 (TID 465) in 25 ms on localhost (executor driver) (1/5) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,052 | INFO | [Executor task launch worker for task 468] | Total > off-heap working memory used after task 1086a9e4-a644-4340-8215-b8c3d936e32d > is 0. Current running tasks are | > org.apache.carbondata.core.memory.UnsafeMemoryManager.freeMemoryAll(UnsafeMemoryManager.java:179)2020-11-12 > 18:55:47,053 | INFO | [task-result-getter-1] | Finished task 4.0 in stage > 84.0 (TID 469) in 24 ms on localhost (executor driver) (2/5) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,053 | INFO | [Executor task launch worker for task 468] | Finished > task 3.0 in stage 84.0 (TID 468). 2380 bytes result sent to driver | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,053 | INFO | [task-result-getter-0] | Finished task 2.0 in stage > 84.0 (TID 467) in 25 ms on localhost (executor driver) (3/5) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,054 | INFO | [task-result-getter-3] | Finished task 1.0 in stage > 84.0 (TID 466) in 26 ms on localhost (executor driver) (4/5) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,054 | INFO | [task-result-getter-2] | Finished task 3.0 in stage > 84.0 (TID 468) in 26 ms on localhost (executor driver) (5/5) | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,054 | INFO | [task-result-getter-2] | Removed TaskSet 84.0, whose > tasks have all completed, from pool | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,054 | INFO | [dag-scheduler-event-loop] | ResultStage 84 > (processCmd at CliDriver.java:376) finished in 0.030 s | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,055 | INFO | [main] | Job 89 finished: processCmd at > CliDriver.java:376, took 0.032288 s | > org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)2020-11-12 > 18:55:47,071 | INFO | [main] | Time taken: 0.344 seconds, Fetched 534 row(s) > | > org.apache.hadoop.hive.ql.session.SessionState$LogHelper.printInfo(SessionState.java:951)2020-11-12 > 18:55:47,086 | AUDIT | [main] | \{"time":"November 12, 2020 6:55:47 PM > CST","username":"root","opName":"SET","opId":"3304241283944785","opStatus":"START"} > | > org.apache.carbondata.processing.util.Auditor.logOperationStart(Auditor.java:74)2020-11-12 > 18:55:47,086 | INFO | [main] | The key carbon.enable.index.server with > value false added in the session param | > org.apache.carbondata.core.util.SessionParams.addProperty(SessionParams.java:102)2020-11-12 > 18:55:47,087 | AUDIT | [main] | \{"time":"November 12, 2020 6:55:47 PM > CST","username":"root","opName":"SET","opId":"3304241283944785","opStatus":"SUCCESS","opTime":"1 > ms","table":"NA","extraInfo":{}} | > org.apache.carbondata.processing.util.Auditor.logOperationEnd(Auditor.java:97)2020-11-12 > 18:55:47,094 | INFO | [main] | Time taken: 0.016 seconds, Fetched 1 row(s) > | > org.apache.hadoop.hive.ql.session.SessionState$LogHelper.printInfo(SessionState.java:951)2020-11-12 > 18:58:04,451 | AUDIT | [main] | \{"time":"November 12, 2020 6:58:04 PM > CST","username":"root","opName":"REFRESH > TABLE","opId":"3304378649144145","opStatus":"START"} | > org.apache.carbondata.processing.util.Auditor.logOperationStart(Auditor.java:74)2020-11-12 > 18:58:04,452 | INFO | [main] | 0: get_database: 1_6_1 | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logInfo(HiveMetaStore.java:746)2020-11-12 > 18:58:04,452 | INFO | [main] | ugi=root ip=unknown-ip-addr > cmd=get_database: 1_6_1 | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logAuditEvent(HiveMetaStore.java:371)2020-11-12 > 18:58:04,455 | INFO | [main] | 0: get_table : db=1_6_1 > tbl=brinjal_deleteseg | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logInfo(HiveMetaStore.java:746)2020-11-12 > 18:58:04,455 | INFO | [main] | ugi=root ip=unknown-ip-addr cmd=get_table : > db=1_6_1 tbl=brinjal_deleteseg | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logAuditEvent(HiveMetaStore.java:371)2020-11-12 > 18:58:04,458 | INFO | [main] | 0: get_table : db=1_6_1 > tbl=brinjal_deleteseg | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logInfo(HiveMetaStore.java:746)2020-11-12 > 18:58:04,458 | INFO | [main] | ugi=root ip=unknown-ip-addr cmd=get_table : > db=1_6_1 tbl=brinjal_deleteseg | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logAuditEvent(HiveMetaStore.java:371)2020-11-12 > 18:58:04,460 | INFO | [main] | 0: get_database: 1_6_1 | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logInfo(HiveMetaStore.java:746)2020-11-12 > 18:58:04,460 | INFO | [main] | ugi=root ip=unknown-ip-addr > cmd=get_database: 1_6_1 | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logAuditEvent(HiveMetaStore.java:371)2020-11-12 > 18:58:04,462 | INFO | [main] | 0: get_database: 1_6_1 | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logInfo(HiveMetaStore.java:746)2020-11-12 > 18:58:04,462 | INFO | [main] | ugi=root ip=unknown-ip-addr > cmd=get_database: 1_6_1 | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logAuditEvent(HiveMetaStore.java:371)2020-11-12 > 18:58:04,465 | INFO | [main] | 0: get_database: 1_6_1 | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logInfo(HiveMetaStore.java:746)2020-11-12 > 18:58:04,465 | INFO | [main] | ugi=root ip=unknown-ip-addr > cmd=get_database: 1_6_1 | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logAuditEvent(HiveMetaStore.java:371)2020-11-12 > 18:58:04,467 | INFO | [main] | 0: get_table : db=1_6_1 > tbl=brinjal_deleteseg | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logInfo(HiveMetaStore.java:746)2020-11-12 > 18:58:04,467 | INFO | [main] | ugi=root ip=unknown-ip-addr cmd=get_table : > db=1_6_1 tbl=brinjal_deleteseg | > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.logAuditEvent(HiveMetaStore.java:371)2020-11-12 > 18:58:04,472 | AUDIT | [main] | \{"time":"November 12, 2020 6:58:04 PM > CST","username":"root","opName":"REFRESH > TABLE","opId":"3304378649144145","opStatus":"FAILED","opTime":"21 > ms","table":"1_6_1.brinjal_deleteseg","extraInfo":{"Exception":"org.apache.spark.sql.catalyst.analysis.NoSuchTableException","Message":"Table > or view 'brinjal_deleteseg' not found in database '1_6_1';"}} | > org.apache.carbondata.processing.util.Auditor.logOperationEnd(Auditor.java:97) -- This message was sent by Atlassian Jira (v8.3.4#803005)