Hi,
Executor log as follows: 

INFO  12-01 15:45:58,574 - Registered signal handlers for [TERM, HUP, INT]
INFO  12-01 15:45:59,380 - Changing view acls to: yarn,spark
INFO  12-01 15:45:59,387 - Changing modify acls to: yarn,spark
INFO  12-01 15:45:59,387 - SecurityManager: authentication disabled; ui acls
disabled; users with view permissions: Set(yarn, spark); users with modify
permissions: Set(yarn, spark)
INFO  12-01 15:46:00,422 - Changing view acls to: yarn,spark
INFO  12-01 15:46:00,422 - Changing modify acls to: yarn,spark
INFO  12-01 15:46:00,422 - SecurityManager: authentication disabled; ui acls
disabled; users with view permissions: Set(yarn, spark); users with modify
permissions: Set(yarn, spark)
INFO  12-01 15:46:00,825 - Slf4jLogger started
INFO  12-01 15:46:00,873 - Starting remoting
INFO  12-01 15:46:01,046 - Remoting started; listening on addresses
:[akka.tcp://sparkExecutorActorSystem@dpnode03:42230]
INFO  12-01 15:46:01,051 - Successfully started service
'sparkExecutorActorSystem' on port 42230.
INFO  12-01 15:46:01,117 - Created local directory at
/data02/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-f1549267-39fd-4f92-a793-db3975c05272
INFO  12-01 15:46:01,117 - Created local directory at
/data03/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-e1f964d4-877c-45e2-88a0-3d9d14ebf47e
INFO  12-01 15:46:01,117 - Created local directory at
/data04/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-63faa7b7-248b-424a-98bb-ab85d9aef0c3
INFO  12-01 15:46:01,118 - Created local directory at
/data05/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-e1ac73f4-54e0-4239-933f-2f16f6997a49
INFO  12-01 15:46:01,118 - Created local directory at
/data06/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-5b8bf328-594a-419b-af4d-d34a8a350dce
INFO  12-01 15:46:01,118 - Created local directory at
/data07/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-6eed5405-1b71-4d70-931a-3c68a8bff14a
INFO  12-01 15:46:01,119 - Created local directory at
/data08/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-5dfcf5a3-d06c-4ae7-a2ec-1a41eb709a26
INFO  12-01 15:46:01,119 - Created local directory at
/data09/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-8ed5f4f3-6c92-4f46-adea-362324bed778
INFO  12-01 15:46:01,119 - Created local directory at
/data10/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-c7367242-4a68-4b45-9a13-cffd4793b46e
INFO  12-01 15:46:01,119 - Created local directory at
/data11/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-2e25390e-835e-4f1e-bca5-96b6ecdc7247
INFO  12-01 15:46:01,120 - Created local directory at
/data12/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-c0911b1e-9c85-4393-bdab-db35a772b4fb
INFO  12-01 15:46:01,141 - MemoryStore started with capacity 5.5 GB
INFO  12-01 15:46:01,632 - Connecting to driver:
spark://CoarseGrainedScheduler@192.168.50.8:39206
INFO  12-01 15:46:01,681 - Successfully registered with driver
INFO  12-01 15:46:01,684 - Starting executor ID 2 on host dpnode03
INFO  12-01 15:46:01,722 - Successfully started service
'org.apache.spark.network.netty.NettyBlockTransferService' on port 47403.
INFO  12-01 15:46:01,723 - Server created on 47403
INFO  12-01 15:46:01,728 - external shuffle service port = 7337
INFO  12-01 15:46:01,728 - Trying to register BlockManager
INFO  12-01 15:46:01,743 - Registered BlockManager
INFO  12-01 15:46:01,747 - Registering executor with local external shuffle
service.
INFO  12-01 15:46:01,778 - Using REPL class URI: http://192.168.50.8:44119
INFO  12-01 15:49:19,059 - Got assigned task 0
INFO  12-01 15:49:19,066 - Running task 0.0 in stage 0.0 (TID 0)
INFO  12-01 15:49:19,270 - Started reading broadcast variable 0
INFO  12-01 15:49:19,442 - Block broadcast_0_piece0 stored as bytes in
memory (estimated size 18.8 KB, free 18.8 KB)
INFO  12-01 15:49:19,468 - Reading broadcast variable 0 took 198 ms
INFO  12-01 15:49:19,616 - Block broadcast_0 stored as values in memory
(estimated size 35.2 KB, free 54.0 KB)
INFO  12-01 15:49:21,375 -
*************************/usr/hdp/2.5.0.0-1245/spark/conf/carbon.properties
INFO  12-01 15:49:21,376 - [Executor task launch
worker-0][partitionID:sale;queryID:1457355617520266_0] Query will be
executed on table: sale
INFO  12-01 15:49:21,379 - [Executor task launch
worker-0][partitionID:sale;queryID:1457355617520266_0] Property file path:
/usr/hdp/2.5.0.0-1245/spark/conf/carbon.properties
INFO  12-01 15:49:21,379 - [Executor task launch
worker-0][partitionID:sale;queryID:1457355617520266_0] ------Using
Carbon.properties --------
INFO  12-01 15:49:21,379 - [Executor task launch
worker-0][partitionID:sale;queryID:1457355617520266_0]
{carbon.number.of.cores.while.loading=6,
carbon.number.of.cores.while.compacting=4, carbon.sort.file.buffer.size=20,
carbon.inmemory.record.size=120000, carbon.sort.size=500000,
carbon.graph.rowset.size=100000, carbon.ddl.base.hdfs.url=/user/spark,
carbon.compaction.level.threshold=8,6, carbon.number.of.cores=4,
carbon.kettle.home=/usr/hdp/2.5.0.0-1245/spark/carbonlib/carbonplugins,
carbon.storelocation=hdfs://julong/carbondata/carbonstore,
carbon.enable.auto.load.merge=true, carbon.enableXXHash=true,
carbon.sort.intermediate.files.limit=100, carbon.major.compaction.size=1024,
carbon.badRecords.location=/opt/Carbon/Spark/badrecords,
carbon.use.local.dir=true, carbon.enable.quick.filter=false}
INFO  12-01 15:49:23,999 - pool-19-thread-1
****************************Total Number Rows In BTREE: 10380500
INFO  12-01 15:49:24,005 - [Executor task launch
worker-0][partitionID:sale;queryID:1457355617520266_0] Column cache size not
configured. Therefore default behavior will be considered and no LRU based
eviction of columns will be done
ERROR 12-01 15:49:24,565 - [Executor task launch
worker-0][partitionID:sale;queryID:1457355617520266_0] 
java.lang.NullPointerException
        at
org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.intialiseInfos(AbstractDetailQueryResultIterator.java:117)
        at
org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.<init>(AbstractDetailQueryResultIterator.java:107)
        at
org.apache.carbondata.scan.result.iterator.DetailQueryResultIterator.<init>(DetailQueryResultIterator.java:43)
        at
org.apache.carbondata.scan.executor.impl.DetailQueryExecutor.execute(DetailQueryExecutor.java:39)
        at
org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:216)
        at
org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
ERROR 12-01 15:49:24,577 - Exception in task 0.0 in stage 0.0 (TID 0)
java.lang.RuntimeException: Exception occurred in query execution.Please
check logs.
        at scala.sys.package$.error(package.scala:27)
        at
org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:226)
        at
org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
INFO  12-01 15:49:34,727 - Got assigned task 3
INFO  12-01 15:49:34,727 - Running task 0.3 in stage 0.0 (TID 3)
INFO  12-01 15:49:34,799 -
*************************/usr/hdp/2.5.0.0-1245/spark/conf/carbon.properties
INFO  12-01 15:49:34,799 - [Executor task launch
worker-0][partitionID:sale;queryID:1457355617520266_0] Query will be
executed on table: sale
ERROR 12-01 15:49:34,813 - [Executor task launch
worker-0][partitionID:sale;queryID:1457355617520266_0] 
java.lang.NullPointerException
        at
org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.intialiseInfos(AbstractDetailQueryResultIterator.java:117)
        at
org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.<init>(AbstractDetailQueryResultIterator.java:107)
        at
org.apache.carbondata.scan.result.iterator.DetailQueryResultIterator.<init>(DetailQueryResultIterator.java:43)
        at
org.apache.carbondata.scan.executor.impl.DetailQueryExecutor.execute(DetailQueryExecutor.java:39)
        at
org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:216)
        at
org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
ERROR 12-01 15:49:34,816 - Exception in task 0.3 in stage 0.0 (TID 3)
java.lang.RuntimeException: Exception occurred in query execution.Please
check logs.
        at scala.sys.package$.error(package.scala:27)
        at
org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:226)
        at
org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)

Thanks.





kumarvishal09 wrote
> Hi,
> Can u please share executor log.
> 
> -Regards
> Kumar Vishal
> 
> On Thu, Jan 12, 2017 at 1:59 PM, Li Peng &lt;

> pengli0606@

> &gt; wrote:
> 
>> Hello,
>>
>> use carbondata 0.2.0, following is the problem:
>>
>> Only one column 'store_id' throws NullPointerException  when query, but
>> it
>> works fine  when some value or table is cached in memory.
>>
>> store_id's  type is int, cardinality is 200 Thousand, is configured about
>> dictionary and inverted index.
>>
>> sql:
>> select
>> order_code,saletype,checkout_date,cashier_code,item_cont,
>> invoice_price,giveamt,saleamt
>> from store.sale where store_id=299998
>>
>> error:
>> ERROR 12-01 10:40:16,861 - [Executor task launch
>> worker-0][partitionID:sale;queryID:1438806645368420_0]
>> java.lang.NullPointerException
>>         at
>> org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultItera
>> tor.intialiseInfos(AbstractDetailQueryResultIterator.java:117)
>>         at
>> org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultItera
>> tor.
> <init>
> (AbstractDetailQueryResultIterator.java:107)
>>         at
>> org.apache.carbondata.scan.result.iterator.DetailQueryResultIterator.<
>> init>(DetailQueryResultIterator.java:43)
>>         at
>> org.apache.carbondata.scan.executor.impl.DetailQueryExecutor.execute(
>> DetailQueryExecutor.java:39)
>>         at
>> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<
>> init>(CarbonScanRDD.scala:216)
>>         at
>> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(
>> CarbonScanRDD.scala:192)
>>         at
>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
>> MapPartitionsRDD.scala:38)
>>         at
>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
>> MapPartitionsRDD.scala:38)
>>         at
>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
>> MapPartitionsRDD.scala:38)
>>         at
>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
>> MapPartitionsRDD.scala:38)
>>         at
>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>>         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
>> scala:66)
>>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
>>         at org.apache.spark.executor.Executor$TaskRunner.run(
>> Executor.scala:227)
>>         at
>> java.util.concurrent.ThreadPoolExecutor.runWorker(
>> ThreadPoolExecutor.java:1145)
>>         at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(
>> ThreadPoolExecutor.java:615)
>>         at java.lang.Thread.run(Thread.java:745)
>> ERROR 12-01 10:40:16,874 - Exception in task 0.1 in stage 0.0 (TID 1)
>> java.lang.RuntimeException: Exception occurred in query execution.Please
>> check logs.
>>         at scala.sys.package$.error(package.scala:27)
>>         at
>> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<
>> init>(CarbonScanRDD.scala:226)
>>         at
>> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(
>> CarbonScanRDD.scala:192)
>>         at
>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
>> MapPartitionsRDD.scala:38)
>>         at
>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
>> MapPartitionsRDD.scala:38)
>>         at
>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
>> MapPartitionsRDD.scala:38)
>>         at
>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
>> MapPartitionsRDD.scala:38)
>>         at
>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>>         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
>> scala:66)
>>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
>>         at org.apache.spark.executor.Executor$TaskRunner.run(
>> Executor.scala:227)
>>         at
>> java.util.concurrent.ThreadPoolExecutor.runWorker(
>> ThreadPoolExecutor.java:1145)
>>         at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(
>> ThreadPoolExecutor.java:615)
>>         at java.lang.Thread.run(Thread.java:745)
>> ------------------------------------------------------------
>> -------------------------------------------------------------------
>> Same err 'NullPointerException' in following sql:
>>
>> select * from store.sale where store_id=100000
>> select * from store.sale where store_id=100001
>> select * from store.sale where store_id=100002
>> select * from store.sale where store_id=100006
>> select * from store.sale where store_id=100011
>> select * from store.sale where store_id=299999
>>
>> But  fine and can return results in following sql:
>>
>> select * from store.sale where store_id=100008
>> select * from store.sale where store_id=100009
>> select * from store.sale where store_id=100010
>> select * from store.sale where store_id=100013
>> select * from store.sale where store_id=100027
>> select * from store.sale limit 10
>> select count(*) from store.sale
>> select * from store.sale where store_id=100005
>> select count(*) from store.sale where store_id=100005
>> select distinct(store_id) from store.sale order by store_id
>>
>>
>> But all work fine when table is cached to memory, not throwing
>> 'NullPointerException'.
>>
>>
>>
>>
>> --
>> View this message in context: http://apache-carbondata-
>> mailing-list-archive.1130556.n5.nabble.com/query-err-
>> NullPointerException-but-fine-after-table-cached-in-memory-tp6032.html
>> Sent from the Apache CarbonData Mailing List archive mailing list archive
>> at Nabble.com.
>>





--
View this message in context: 
http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/query-err-NullPointerException-but-fine-after-table-cached-in-memory-tp6032p6035.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at 
Nabble.com.

Reply via email to