OK.  Thanks.

________________________________
pengli0...@outlook.com

From: kumarvishal09 [via Apache CarbonData Mailing List 
archive]<mailto:ml-node+s1130556n6108...@n5.nabble.com>
Date: 2017-01-13 15:23
To: Li Peng<mailto:pengli0...@outlook.com>
Subject: Re: query err 'NullPointerException' but fine after table cached in 
memory
Hi Le Peng,
This is an issue which got fixed after 0.2 release. Please take latest  0.2
branch from git and compile and try.

-Regards
Kumar Vishal

On Thu, Jan 12, 2017 at 4:14 PM, Li Peng <[hidden 
email]</user/SendEmail.jtp?type=node&node=6108&i=0>> wrote:

> Hi,
> Executor log as follows:
>
> INFO  12-01 15:45:58,574 - Registered signal handlers for [TERM, HUP, INT]
> INFO  12-01 15:45:59,380 - Changing view acls to: yarn,spark
> INFO  12-01 15:45:59,387 - Changing modify acls to: yarn,spark
> INFO  12-01 15:45:59,387 - SecurityManager: authentication disabled; ui
> acls
> disabled; users with view permissions: Set(yarn, spark); users with modify
> permissions: Set(yarn, spark)
> INFO  12-01 15:46:00,422 - Changing view acls to: yarn,spark
> INFO  12-01 15:46:00,422 - Changing modify acls to: yarn,spark
> INFO  12-01 15:46:00,422 - SecurityManager: authentication disabled; ui
> acls
> disabled; users with view permissions: Set(yarn, spark); users with modify
> permissions: Set(yarn, spark)
> INFO  12-01 15:46:00,825 - Slf4jLogger started
> INFO  12-01 15:46:00,873 - Starting remoting
> INFO  12-01 15:46:01,046 - Remoting started; listening on addresses
> :[akka.tcp://sparkExecutorActorSystem@dpnode03:42230]
> INFO  12-01 15:46:01,051 - Successfully started service
> 'sparkExecutorActorSystem' on port 42230.
> INFO  12-01 15:46:01,117 - Created local directory at
> /data02/hadoop/yarn/local/usercache/spark/appcache/
> application_1484187732071_0003/blockmgr-f1549267-39fd-
> 4f92-a793-db3975c05272
> INFO  12-01 15:46:01,117 - Created local directory at
> /data03/hadoop/yarn/local/usercache/spark/appcache/
> application_1484187732071_0003/blockmgr-e1f964d4-877c-
> 45e2-88a0-3d9d14ebf47e
> INFO  12-01 15:46:01,117 - Created local directory at
> /data04/hadoop/yarn/local/usercache/spark/appcache/
> application_1484187732071_0003/blockmgr-63faa7b7-248b-
> 424a-98bb-ab85d9aef0c3
> INFO  12-01 15:46:01,118 - Created local directory at
> /data05/hadoop/yarn/local/usercache/spark/appcache/
> application_1484187732071_0003/blockmgr-e1ac73f4-54e0-
> 4239-933f-2f16f6997a49
> INFO  12-01 15:46:01,118 - Created local directory at
> /data06/hadoop/yarn/local/usercache/spark/appcache/
> application_1484187732071_0003/blockmgr-5b8bf328-594a-
> 419b-af4d-d34a8a350dce
> INFO  12-01 15:46:01,118 - Created local directory at
> /data07/hadoop/yarn/local/usercache/spark/appcache/
> application_1484187732071_0003/blockmgr-6eed5405-1b71-
> 4d70-931a-3c68a8bff14a
> INFO  12-01 15:46:01,119 - Created local directory at
> /data08/hadoop/yarn/local/usercache/spark/appcache/
> application_1484187732071_0003/blockmgr-5dfcf5a3-d06c-
> 4ae7-a2ec-1a41eb709a26
> INFO  12-01 15:46:01,119 - Created local directory at
> /data09/hadoop/yarn/local/usercache/spark/appcache/
> application_1484187732071_0003/blockmgr-8ed5f4f3-6c92-
> 4f46-adea-362324bed778
> INFO  12-01 15:46:01,119 - Created local directory at
> /data10/hadoop/yarn/local/usercache/spark/appcache/
> application_1484187732071_0003/blockmgr-c7367242-4a68-
> 4b45-9a13-cffd4793b46e
> INFO  12-01 15:46:01,119 - Created local directory at
> /data11/hadoop/yarn/local/usercache/spark/appcache/
> application_1484187732071_0003/blockmgr-2e25390e-835e-
> 4f1e-bca5-96b6ecdc7247
> INFO  12-01 15:46:01,120 - Created local directory at
> /data12/hadoop/yarn/local/usercache/spark/appcache/
> application_1484187732071_0003/blockmgr-c0911b1e-9c85-
> 4393-bdab-db35a772b4fb
> INFO  12-01 15:46:01,141 - MemoryStore started with capacity 5.5 GB
> INFO  12-01 15:46:01,632 - Connecting to driver:
> spark://CoarseGrainedScheduler@192.168.50.8:39206
> INFO  12-01 15:46:01,681 - Successfully registered with driver
> INFO  12-01 15:46:01,684 - Starting executor ID 2 on host dpnode03
> INFO  12-01 15:46:01,722 - Successfully started service
> 'org.apache.spark.network.netty.NettyBlockTransferService' on port 47403.
> INFO  12-01 15:46:01,723 - Server created on 47403
> INFO  12-01 15:46:01,728 - external shuffle service port = 7337
> INFO  12-01 15:46:01,728 - Trying to register BlockManager
> INFO  12-01 15:46:01,743 - Registered BlockManager
> INFO  12-01 15:46:01,747 - Registering executor with local external shuffle
> service.
> INFO  12-01 15:46:01,778 - Using REPL class URI: http://192.168.50.8:44119
> INFO  12-01 15:49:19,059 - Got assigned task 0
> INFO  12-01 15:49:19,066 - Running task 0.0 in stage 0.0 (TID 0)
> INFO  12-01 15:49:19,270 - Started reading broadcast variable 0
> INFO  12-01 15:49:19,442 - Block broadcast_0_piece0 stored as bytes in
> memory (estimated size 18.8 KB, free 18.8 KB)
> INFO  12-01 15:49:19,468 - Reading broadcast variable 0 took 198 ms
> INFO  12-01 15:49:19,616 - Block broadcast_0 stored as values in memory
> (estimated size 35.2 KB, free 54.0 KB)
> INFO  12-01 15:49:21,375 -
> *************************/usr/hdp/2.5.0.0-1245/spark/conf/
> carbon.properties
> INFO  12-01 15:49:21,376 - [Executor task launch
> worker-0][partitionID:sale;queryID:1457355617520266_0] Query will be
> executed on table: sale
> INFO  12-01 15:49:21,379 - [Executor task launch
> worker-0][partitionID:sale;queryID:1457355617520266_0] Property file path:
> /usr/hdp/2.5.0.0-1245/spark/conf/carbon.properties
> INFO  12-01 15:49:21,379 - [Executor task launch
> worker-0][partitionID:sale;queryID:1457355617520266_0] ------Using
> Carbon.properties --------
> INFO  12-01 15:49:21,379 - [Executor task launch
> worker-0][partitionID:sale;queryID:1457355617520266_0]
> {carbon.number.of.cores.while.loading=6,
> carbon.number.of.cores.while.compacting=4, carbon.sort.file.buffer.size=
> 20,
> carbon.inmemory.record.size=120000, carbon.sort.size=500000,
> carbon.graph.rowset.size=100000, carbon.ddl.base.hdfs.url=/user/spark,
> carbon.compaction.level.threshold=8,6, carbon.number.of.cores=4,
> carbon.kettle.home=/usr/hdp/2.5.0.0-1245/spark/carbonlib/carbonplugins,
> carbon.storelocation=hdfs://julong/carbondata/carbonstore,
> carbon.enable.auto.load.merge=true, carbon.enableXXHash=true,
> carbon.sort.intermediate.files.limit=100, carbon.major.compaction.size=
> 1024,
> carbon.badRecords.location=/opt/Carbon/Spark/badrecords,
> carbon.use.local.dir=true, carbon.enable.quick.filter=false}
> INFO  12-01 15:49:23,999 - pool-19-thread-1
> ****************************Total Number Rows In BTREE: 10380500
> INFO  12-01 15:49:24,005 - [Executor task launch
> worker-0][partitionID:sale;queryID:1457355617520266_0] Column cache size
> not
> configured. Therefore default behavior will be considered and no LRU based
> eviction of columns will be done
> ERROR 12-01 15:49:24,565 - [Executor task launch
> worker-0][partitionID:sale;queryID:1457355617520266_0]
> java.lang.NullPointerException
>         at
> org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultItera
> tor.intialiseInfos(AbstractDetailQueryResultIterator.java:117)
>         at
> org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultItera
> tor.<init>(AbstractDetailQueryResultIterator.java:107)
>         at
> org.apache.carbondata.scan.result.iterator.DetailQueryResultIterator.<
> init>(DetailQueryResultIterator.java:43)
>         at
> org.apache.carbondata.scan.executor.impl.DetailQueryExecutor.execute(
> DetailQueryExecutor.java:39)
>         at
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<
> init>(CarbonScanRDD.scala:216)
>         at
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(
> CarbonScanRDD.scala:192)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> scala:66)
>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
>         at org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:227)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:745)
> ERROR 12-01 15:49:24,577 - Exception in task 0.0 in stage 0.0 (TID 0)
> java.lang.RuntimeException: Exception occurred in query execution.Please
> check logs.
>         at scala.sys.package$.error(package.scala:27)
>         at
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<
> init>(CarbonScanRDD.scala:226)
>         at
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(
> CarbonScanRDD.scala:192)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> scala:66)
>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
>         at org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:227)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:745)
> INFO  12-01 15:49:34,727 - Got assigned task 3
> INFO  12-01 15:49:34,727 - Running task 0.3 in stage 0.0 (TID 3)
> INFO  12-01 15:49:34,799 -
> *************************/usr/hdp/2.5.0.0-1245/spark/conf/
> carbon.properties
> INFO  12-01 15:49:34,799 - [Executor task launch
> worker-0][partitionID:sale;queryID:1457355617520266_0] Query will be
> executed on table: sale
> ERROR 12-01 15:49:34,813 - [Executor task launch
> worker-0][partitionID:sale;queryID:1457355617520266_0]
> java.lang.NullPointerException
>         at
> org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultItera
> tor.intialiseInfos(AbstractDetailQueryResultIterator.java:117)
>         at
> org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultItera
> tor.<init>(AbstractDetailQueryResultIterator.java:107)
>         at
> org.apache.carbondata.scan.result.iterator.DetailQueryResultIterator.<
> init>(DetailQueryResultIterator.java:43)
>         at
> org.apache.carbondata.scan.executor.impl.DetailQueryExecutor.execute(
> DetailQueryExecutor.java:39)
>         at
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<
> init>(CarbonScanRDD.scala:216)
>         at
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(
> CarbonScanRDD.scala:192)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> scala:66)
>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
>         at org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:227)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:745)
> ERROR 12-01 15:49:34,816 - Exception in task 0.3 in stage 0.0 (TID 3)
> java.lang.RuntimeException: Exception occurred in query execution.Please
> check logs.
>         at scala.sys.package$.error(package.scala:27)
>         at
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<
> init>(CarbonScanRDD.scala:226)
>         at
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(
> CarbonScanRDD.scala:192)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> scala:66)
>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
>         at org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:227)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:745)
>
> Thanks.
>
>
>
>
>
> kumarvishal09 wrote
> > Hi,
> > Can u please share executor log.
> >
> > -Regards
> > Kumar Vishal
> >
> > On Thu, Jan 12, 2017 at 1:59 PM, Li Peng &lt;
>
> > pengli0606@
>
> > &gt; wrote:
> >
> >> Hello,
> >>
> >> use carbondata 0.2.0, following is the problem:
> >>
> >> Only one column 'store_id' throws NullPointerException  when query, but
> >> it
> >> works fine  when some value or table is cached in memory.
> >>
> >> store_id's  type is int, cardinality is 200 Thousand, is configured
> about
> >> dictionary and inverted index.
> >>
> >> sql:
> >> select
> >> order_code,saletype,checkout_date,cashier_code,item_cont,
> >> invoice_price,giveamt,saleamt
> >> from store.sale where store_id=299998
> >>
> >> error:
> >> ERROR 12-01 10:40:16,861 - [Executor task launch
> >> worker-0][partitionID:sale;queryID:1438806645368420_0]
> >> java.lang.NullPointerException
> >>         at
> >> org.apache.carbondata.scan.result.iterator.
> AbstractDetailQueryResultItera
> >> tor.intialiseInfos(AbstractDetailQueryResultIterator.java:117)
> >>         at
> >> org.apache.carbondata.scan.result.iterator.
> AbstractDetailQueryResultItera
> >> tor.
> > <init>
> > (AbstractDetailQueryResultIterator.java:107)
> >>         at
> >> org.apache.carbondata.scan.result.iterator.DetailQueryResultIterator.<
> >> init>(DetailQueryResultIterator.java:43)
> >>         at
> >> org.apache.carbondata.scan.executor.impl.DetailQueryExecutor.execute(
> >> DetailQueryExecutor.java:39)
> >>         at
> >> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<
> >> init>(CarbonScanRDD.scala:216)
> >>         at
> >> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(
> >> CarbonScanRDD.scala:192)
> >>         at
> >> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
> >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
> >>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> >> MapPartitionsRDD.scala:38)
> >>         at
> >> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
> >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
> >>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> >> MapPartitionsRDD.scala:38)
> >>         at
> >> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
> >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
> >>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> >> MapPartitionsRDD.scala:38)
> >>         at
> >> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
> >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
> >>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> >> MapPartitionsRDD.scala:38)
> >>         at
> >> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
> >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
> >>         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> >> scala:66)
> >>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
> >>         at org.apache.spark.executor.Executor$TaskRunner.run(
> >> Executor.scala:227)
> >>         at
> >> java.util.concurrent.ThreadPoolExecutor.runWorker(
> >> ThreadPoolExecutor.java:1145)
> >>         at
> >> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> >> ThreadPoolExecutor.java:615)
> >>         at java.lang.Thread.run(Thread.java:745)
> >> ERROR 12-01 10:40:16,874 - Exception in task 0.1 in stage 0.0 (TID 1)
> >> java.lang.RuntimeException: Exception occurred in query execution.Please
> >> check logs.
> >>         at scala.sys.package$.error(package.scala:27)
> >>         at
> >> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<
> >> init>(CarbonScanRDD.scala:226)
> >>         at
> >> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(
> >> CarbonScanRDD.scala:192)
> >>         at
> >> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
> >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
> >>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> >> MapPartitionsRDD.scala:38)
> >>         at
> >> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
> >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
> >>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> >> MapPartitionsRDD.scala:38)
> >>         at
> >> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
> >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
> >>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> >> MapPartitionsRDD.scala:38)
> >>         at
> >> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
> >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
> >>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> >> MapPartitionsRDD.scala:38)
> >>         at
> >> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
> >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
> >>         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> >> scala:66)
> >>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
> >>         at org.apache.spark.executor.Executor$TaskRunner.run(
> >> Executor.scala:227)
> >>         at
> >> java.util.concurrent.ThreadPoolExecutor.runWorker(
> >> ThreadPoolExecutor.java:1145)
> >>         at
> >> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> >> ThreadPoolExecutor.java:615)
> >>         at java.lang.Thread.run(Thread.java:745)
> >> ------------------------------------------------------------
> >> -------------------------------------------------------------------
> >> Same err 'NullPointerException' in following sql:
> >>
> >> select * from store.sale where store_id=100000
> >> select * from store.sale where store_id=100001
> >> select * from store.sale where store_id=100002
> >> select * from store.sale where store_id=100006
> >> select * from store.sale where store_id=100011
> >> select * from store.sale where store_id=299999
> >>
> >> But  fine and can return results in following sql:
> >>
> >> select * from store.sale where store_id=100008
> >> select * from store.sale where store_id=100009
> >> select * from store.sale where store_id=100010
> >> select * from store.sale where store_id=100013
> >> select * from store.sale where store_id=100027
> >> select * from store.sale limit 10
> >> select count(*) from store.sale
> >> select * from store.sale where store_id=100005
> >> select count(*) from store.sale where store_id=100005
> >> select distinct(store_id) from store.sale order by store_id
> >>
> >>
> >> But all work fine when table is cached to memory, not throwing
> >> 'NullPointerException'.
> >>
> >>
> >>
> >>
> >> --
> >> View this message in context: http://apache-carbondata-
> >> mailing-list-archive.1130556.n5.nabble.com/query-err-
> >> NullPointerException-but-fine-after-table-cached-in-memory-tp6032.html
> >> Sent from the Apache CarbonData Mailing List archive mailing list
> archive
> >> at Nabble.com.
> >>
>
>
>
>
>
> --
> View this message in context: http://apache-carbondata-
> mailing-list-archive.1130556.n5.nabble.com/query-err-
> NullPointerException-but-fine-after-table-cached-in-memory-
> tp6032p6035.html
> Sent from the Apache CarbonData Mailing List archive mailing list archive
> at Nabble.com.
>
kumar vishal


________________________________
If you reply to this email, your message will be added to the discussion below:
http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/query-err-NullPointerException-but-fine-after-table-cached-in-memory-tp6032p6108.html
To unsubscribe from query err 'NullPointerException' but fine after table 
cached in memory, click 
here<http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_code&node=6032&code=cGVuZ2xpMDYwNkBvdXRsb29rLmNvbXw2MDMyfDk5OTg3NTMwMQ==>.
NAML<http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>




--
View this message in context: 
http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/query-err-NullPointerException-but-fine-after-table-cached-in-memory-tp6032p6109.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at 
Nabble.com.

Reply via email to