[jira] [Reopened] (CARBONDATA-867) Shows an error on select query with date & timestamp datatype after altering the table.

2017-04-19 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO reopened CARBONDATA-867:
--

select distinct(date1) from uniqdata1 ; 
it is displaying NULL value instead of date.

> Shows an error on select query with date & timestamp datatype after altering 
> the table.
> ---
>
> Key: CARBONDATA-867
> URL: https://issues.apache.org/jira/browse/CARBONDATA-867
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.1.0-incubating
> Environment: Spark2.1
>Reporter: SWATI RAO
> Attachments: 2000_UniqData.csv
>
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 add columns(tmpstmp date) 
> TBLPROPERTIES('DEFAULT.VALUE.tmpstmp'= '2017-01-01');
> 0: jdbc:hive2://192.168.2.126:1> select distinct(tmpstmp) from uniqdata1 ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 44.0 failed 1 times, most recent failure: Lost task 0.0 in 
> stage 44.0 (TID 1038, localhost, executor driver): 
> java.lang.ClassCastException: java.lang.Integer cannot be cast to 
> java.lang.Long
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDirectDictionaryData(RestructureBasedVectorResultCollector.java:151)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingDimensions(RestructureBasedVectorResultCollector.java:114)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:97)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
>   at 
> org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithKeys$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
>  Source)
>   at 
> org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
>   at 
> org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
>   at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
>   at 
> org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
>   at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
>   at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
>   at org.apache.spark.scheduler.Task.run(Task.scala:99)
>   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)
> Driver stacktrace: (state=,code=0)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Closed] (CARBONDATA-864) After adding column using alter query, when we put any column in "Dictionary Exclude" then perform select query on that column then it will throws an exception.

2017-04-19 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-864?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-864.

Resolution: Fixed

> After adding column using alter query, when we put any column in "Dictionary 
> Exclude" then perform select query on that column then it will throws an 
> exception.
> 
>
> Key: CARBONDATA-864
> URL: https://issues.apache.org/jira/browse/CARBONDATA-864
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.1.0-incubating
> Environment: Spark2.1
>Reporter: SWATI RAO
> Attachments: 2000_UniqData.csv, error.png
>
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 add columns(dict int) 
> TBLPROPERTIES('DICTIONARY_INCLUDE'='dict','DEFAULT.VALUE.dict'= '');
> select distinct(dict) from uniqdata1 ;
> when we perform select query on "DICTIONARY_INCLUDE" it is working
> but when we perform select query on "DICTIONARY_EXCLUDE" then it will throws 
> an exception:
> alter table uniqdata1 add columns(nodict string) 
> TBLPROPERTIES('DICTIONARY_EXCLUDE'='nodict', 'DEFAULT.VALUE.NoDict'= 'abcd');
> select distinct(nodict) from uniqdata1 ;
> 0: jdbc:hive2://192.168.2.126:1> select distinct(nodict) from uniqdata1 ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 21.0 failed 1 times, most recent failure: Lost task 0.0 in 
> stage 21.0 (TID 419, localhost, executor driver): 
> java.lang.ArrayIndexOutOfBoundsException: 4186
>   at 
> org.apache.spark.sql.execution.vectorized.OnHeapColumnVector.putByteArray(OnHeapColumnVector.java:401)
>   at 
> org.apache.spark.sql.execution.vectorized.ColumnVector.putByteArray(ColumnVector.java:569)
>   at 
> org.apache.carbondata.spark.vectorreader.ColumnarVectorWrapper.putBytes(ColumnarVectorWrapper.java:77)
>   at 
> org.apache.carbondata.spark.vectorreader.ColumnarVectorWrapper.putBytes(ColumnarVectorWrapper.java:83)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillNoDictionaryData(RestructureBasedVectorResultCollector.java:167)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingDimensions(RestructureBasedVectorResultCollector.java:122)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:97)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
>   at 
> org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithKeys$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
>  Source)
>   at 
> org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
>   at 
> org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
>   at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
>   at 
> org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
>   at 
> 

[jira] [Closed] (CARBONDATA-867) Shows an error on select query with date & timestamp datatype after altering the table.

2017-04-19 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-867.

Resolution: Fixed

> Shows an error on select query with date & timestamp datatype after altering 
> the table.
> ---
>
> Key: CARBONDATA-867
> URL: https://issues.apache.org/jira/browse/CARBONDATA-867
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.1.0-incubating
> Environment: Spark2.1
>Reporter: SWATI RAO
> Attachments: 2000_UniqData.csv
>
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 add columns(tmpstmp date) 
> TBLPROPERTIES('DEFAULT.VALUE.tmpstmp'= '2017-01-01');
> 0: jdbc:hive2://192.168.2.126:1> select distinct(tmpstmp) from uniqdata1 ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 44.0 failed 1 times, most recent failure: Lost task 0.0 in 
> stage 44.0 (TID 1038, localhost, executor driver): 
> java.lang.ClassCastException: java.lang.Integer cannot be cast to 
> java.lang.Long
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDirectDictionaryData(RestructureBasedVectorResultCollector.java:151)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingDimensions(RestructureBasedVectorResultCollector.java:114)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:97)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
>   at 
> org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithKeys$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
>  Source)
>   at 
> org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
>   at 
> org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
>   at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
>   at 
> org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
>   at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
>   at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
>   at org.apache.spark.scheduler.Task.run(Task.scala:99)
>   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)
> Driver stacktrace: (state=,code=0)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Closed] (CARBONDATA-877) String datatype is throwing an error when included in DIctionary_Exclude in a alter query

2017-04-19 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-877?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-877.


> String datatype is throwing an error when included in DIctionary_Exclude in a 
> alter query
> -
>
> Key: CARBONDATA-877
> URL: https://issues.apache.org/jira/browse/CARBONDATA-877
> Project: CarbonData
>  Issue Type: Bug
>Reporter: SWATI RAO
>Assignee: Kunal Kapoor
> Fix For: 1.1.0-incubating
>
> Attachments: 2000_UniqData.csv
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 drop columns(CUST_NAME);
> alter table uniqdata1 add columns(CUST_NAME string) 
> TBLPROPERTIES('DICTIONARY_EXCLUDE'='CUST_NAME', 
> 'DEFAULT.VALUE.CUST_NAME'='testuser') ;
> Column added successfully. But when we execute:
> select distinct(CUST_NAME) from uniqdata1 ; 
> &
> select count(CUST_NAME) from uniqdata1 ;
> it throws an error :
> "Job aborted due to stage failure: Task 0 in stage 9.0 failed 1 times, most 
> recent failure: Lost task 0.0 in stage 9.0 (TID 206, localhost, executor 
> driver): java.lang.ArrayIndexOutOfBoundsException: 4186"



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Commented] (CARBONDATA-868) Select query on decimal datatype is not working fine after adding decimal column using alter

2017-04-19 Thread SWATI RAO (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-868?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15974139#comment-15974139
 ] 

SWATI RAO commented on CARBONDATA-868:
--

@Srigopal Mohanty : Okay I will check it once more and will get back to you.

> Select query on decimal datatype is not working fine after adding decimal 
> column using alter
> 
>
> Key: CARBONDATA-868
> URL: https://issues.apache.org/jira/browse/CARBONDATA-868
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.1.0-incubating
> Environment: Spark2.1
>Reporter: SWATI RAO
> Attachments: 2000_UniqData.csv
>
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 add columns(msrField 
> decimal(5,2))TBLPROPERTIES('DEFAULT.VALUE.msrfield'= '123.45');
> 0: jdbc:hive2://192.168.2.126:1> select msrField from uniqdata1;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 48.0 failed 1 times, most recent failure: Lost task 0.0 in 
> stage 48.0 (TID 1041, localhost, executor driver): 
> java.lang.ArrayIndexOutOfBoundsException: 4186
>   at 
> org.apache.spark.sql.execution.vectorized.OnHeapColumnVector.putInt(OnHeapColumnVector.java:202)
>   at 
> org.apache.spark.sql.execution.vectorized.ColumnVector.putDecimal(ColumnVector.java:608)
>   at 
> org.apache.carbondata.spark.vectorreader.ColumnarVectorWrapper.putDecimal(ColumnarVectorWrapper.java:58)
>   at 
> org.apache.carbondata.spark.vectorreader.ColumnarVectorWrapper.putDecimals(ColumnarVectorWrapper.java:64)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingMeasures(RestructureBasedVectorResultCollector.java:202)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:98)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
>   at 
> org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
>  Source)
>   at 
> org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
>   at 
> org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
>   at 
> org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:231)
>   at 
> org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:225)
>   at 
> org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:826)
>   at 
> org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:826)
>   at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
>   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
>   at org.apache.spark.scheduler.Task.run(Task.scala:99)
>   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> 

[jira] [Updated] (CARBONDATA-918) Select query is not working for Complex datatype

2017-04-12 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-918?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-918:
-
Environment: Spark 2.1  (was: Spark 1.6)

> Select query is not working for Complex datatype
> 
>
> Key: CARBONDATA-918
> URL: https://issues.apache.org/jira/browse/CARBONDATA-918
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.0.0-incubating
> Environment: Spark 2.1
>Reporter: SWATI RAO
> Attachments: Array.csv
>
>
> Select Query is not working in carbondata for complex datatype:
> CARBONDATA:
> 0: jdbc:hive2://hadoop-master:1> create table Array_com (CUST_ID string, 
> YEAR int, MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED 
> string, ARRAY_INT array,ARRAY_STRING array,ARRAY_DATE 
> array,CARD_COUNT int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT 
> double, HQ_DEPOSIT double) STORED BY 'org.apache.carbondata.format' ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (3.567 seconds)
> 0: jdbc:hive2://hadoop-master:1>  LOAD DATA INPATH 
> 'hdfs://192.168.2.145:54310/BabuStore/Data/complex/Array.csv' INTO table 
> Array_com  options ('DELIMITER'=',', 'QUOTECHAR'='"', 
> 'FILEHEADER'='CUST_ID,YEAR,MONTH,AGE,GENDER,EDUCATED,IS_MARRIED,ARRAY_INT,ARRAY_STRING,ARRAY_DATE,CARD_COUNT,DEBIT_COUNT,CREDIT_COUNT,DEPOSIT,HQ_DEPOSIT','COMPLEX_DELIMITER_LEVEL_1'='$')
>  ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (6.541 seconds)
> 0: jdbc:hive2://hadoop-master:1> select array_int[0], array_int[0]+ 10 as 
> a  from array_com ;
> 0: jdbc:hive2://192.168.2.126:1> select * from array_com ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 1 in stage 254.0 failed 1 times, most recent failure: Lost task 1.0 in 
> stage 254.0 (TID 8631, localhost, executor driver): 
> java.lang.NullPointerException
> at 
> org.apache.carbondata.core.datastore.filesystem.AbstractDFSCarbonFile.getLastModifiedTime(AbstractDFSCarbonFile.java:135)
> at 
> org.apache.carbondata.core.datastore.filesystem.AbstractDFSCarbonFile.isFileModified(AbstractDFSCarbonFile.java:210)
> at 
> org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.isDictionaryMetaFileModified(AbstractDictionaryCache.java:119)
> at 
> org.apache.carbondata.core.cache.dictionary.AbstractDictionaryCache.checkAndLoadDictionaryData(AbstractDictionaryCache.java:159)
> at 
> org.apache.carbondata.core.cache.dictionary.ForwardDictionaryCache.getDictionary(ForwardDictionaryCache.java:195)
> at 
> org.apache.carbondata.core.cache.dictionary.ForwardDictionaryCache.get(ForwardDictionaryCache.java:71)
> at 
> org.apache.carbondata.core.cache.dictionary.ForwardDictionaryCache.get(ForwardDictionaryCache.java:40)
> at 
> org.apache.carbondata.hadoop.readsupport.impl.DictionaryDecodeReadSupport.initialize(DictionaryDecodeReadSupport.java:65)
> at 
> org.apache.carbondata.spark.readsupport.SparkRowReadSupportImpl.initialize(SparkRowReadSupportImpl.java:33)
> at 
> org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:76)
> at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
> at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
> at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
> at org.apache.spark.scheduler.Task.run(Task.scala:99)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> Driver stacktrace: (state=,code=0)
> HIVE:
> create table Array_com_h (CUST_ID string, YEAR int, MONTH int, AGE int, 
> GENDER string, EDUCATED string, IS_MARRIED string, ARRAY_INT 
> array,ARRAY_STRING array,ARRAY_DATE array,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double);
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (6.541 seconds)
> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/complex/Array.csv' INTO table 
> Array_com_h;
> +-+--+

[jira] [Updated] (CARBONDATA-904) ArrayIndexOutOfBoundsException

2017-04-11 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-904?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-904:
-
  Attachment: Test_Data1.csv
  Test_Data1_h1.csv
 Environment: Spark1.6
Request participants:   (was: )
   Fix Version/s: 1.1.0-incubating

> ArrayIndexOutOfBoundsException 
> ---
>
> Key: CARBONDATA-904
> URL: https://issues.apache.org/jira/browse/CARBONDATA-904
> Project: CarbonData
>  Issue Type: Bug
> Environment: Spark1.6
>Reporter: SWATI RAO
> Fix For: 1.1.0-incubating
>
> Attachments: Test_Data1_h1.csv, Test_Data1_h1.csv
>
>
> Or operator is not working properly.
> When we execute these query in hive it is working fine but when we execute 
> the same in carbondata it throws an exception:
> java.lang.ArrayIndexOutOfBoundsException
> HIVE:
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary_h1 (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED 
> FIELDS TERMINATED BY ',' ;
> +-+--+
> | result  |
> +-+--+
> +-+--+
> No rows selected (1.177 seconds)
> 0: jdbc:hive2://hadoop-master:1> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE 
> Test_Boundary_h1 ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.437 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) 
> from Test_Boundary_h1 where c6_Timestamp ='2017-07-01 12:07:28' or 
> c6_Timestamp ='2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' 
> group by c6_Timestamp ;
> +++--+
> |  c6_Timestamp  |  _c1   |
> +++--+
> | 2017-07-01 12:07:28.0  | 2017-07-01 12:07:28.0  |
> +++--+
> 1 row selected (1.637 seconds)
> CARBONDATA:
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format' ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (4.48 seconds)
> 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 
> 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table 
> Test_Boundary 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
>  ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (4.445 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) 
> from Test_Boundary where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp 
> =' 2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by 
> c6_Timestamp ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 5.0 failed 4 times, most recent failure: Lost task 0.3 in 
> stage 5.0 (TID 8, hadoop-master): java.lang.RuntimeException: 
> java.util.concurrent.ExecutionException: 
> java.lang.ArrayIndexOutOfBoundsException: 0
>   at 
> org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.updateScanner(AbstractDataBlockIterator.java:136)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:50)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31)
>   at 
> org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41)
>   at 
> org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:79)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> 

[jira] [Updated] (CARBONDATA-904) ArrayIndexOutOfBoundsException

2017-04-11 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-904?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-904:
-
Attachment: (was: Test_Data1.csv)

> ArrayIndexOutOfBoundsException 
> ---
>
> Key: CARBONDATA-904
> URL: https://issues.apache.org/jira/browse/CARBONDATA-904
> Project: CarbonData
>  Issue Type: Bug
> Environment: Spark1.6
>Reporter: SWATI RAO
> Fix For: 1.1.0-incubating
>
> Attachments: Test_Data1_h1.csv, Test_Data1_h1.csv
>
>
> Or operator is not working properly.
> When we execute these query in hive it is working fine but when we execute 
> the same in carbondata it throws an exception:
> java.lang.ArrayIndexOutOfBoundsException
> HIVE:
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary_h1 (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED 
> FIELDS TERMINATED BY ',' ;
> +-+--+
> | result  |
> +-+--+
> +-+--+
> No rows selected (1.177 seconds)
> 0: jdbc:hive2://hadoop-master:1> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE 
> Test_Boundary_h1 ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.437 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) 
> from Test_Boundary_h1 where c6_Timestamp ='2017-07-01 12:07:28' or 
> c6_Timestamp ='2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' 
> group by c6_Timestamp ;
> +++--+
> |  c6_Timestamp  |  _c1   |
> +++--+
> | 2017-07-01 12:07:28.0  | 2017-07-01 12:07:28.0  |
> +++--+
> 1 row selected (1.637 seconds)
> CARBONDATA:
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format' ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (4.48 seconds)
> 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 
> 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table 
> Test_Boundary 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
>  ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (4.445 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) 
> from Test_Boundary where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp 
> =' 2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by 
> c6_Timestamp ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 5.0 failed 4 times, most recent failure: Lost task 0.3 in 
> stage 5.0 (TID 8, hadoop-master): java.lang.RuntimeException: 
> java.util.concurrent.ExecutionException: 
> java.lang.ArrayIndexOutOfBoundsException: 0
>   at 
> org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.updateScanner(AbstractDataBlockIterator.java:136)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:50)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31)
>   at 
> org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41)
>   at 
> org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:79)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> 

[jira] [Updated] (CARBONDATA-904) ArrayIndexOutOfBoundsException

2017-04-11 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-904?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-904:
-
Attachment: (was: Test_Data1.csv)

> ArrayIndexOutOfBoundsException 
> ---
>
> Key: CARBONDATA-904
> URL: https://issues.apache.org/jira/browse/CARBONDATA-904
> Project: CarbonData
>  Issue Type: Bug
> Environment: Spark1.6
>Reporter: SWATI RAO
> Fix For: 1.1.0-incubating
>
> Attachments: Test_Data1_h1.csv, Test_Data1_h1.csv
>
>
> Or operator is not working properly.
> When we execute these query in hive it is working fine but when we execute 
> the same in carbondata it throws an exception:
> java.lang.ArrayIndexOutOfBoundsException
> HIVE:
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary_h1 (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED 
> FIELDS TERMINATED BY ',' ;
> +-+--+
> | result  |
> +-+--+
> +-+--+
> No rows selected (1.177 seconds)
> 0: jdbc:hive2://hadoop-master:1> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE 
> Test_Boundary_h1 ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.437 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) 
> from Test_Boundary_h1 where c6_Timestamp ='2017-07-01 12:07:28' or 
> c6_Timestamp ='2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' 
> group by c6_Timestamp ;
> +++--+
> |  c6_Timestamp  |  _c1   |
> +++--+
> | 2017-07-01 12:07:28.0  | 2017-07-01 12:07:28.0  |
> +++--+
> 1 row selected (1.637 seconds)
> CARBONDATA:
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format' ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (4.48 seconds)
> 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 
> 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table 
> Test_Boundary 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
>  ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (4.445 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) 
> from Test_Boundary where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp 
> =' 2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by 
> c6_Timestamp ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 5.0 failed 4 times, most recent failure: Lost task 0.3 in 
> stage 5.0 (TID 8, hadoop-master): java.lang.RuntimeException: 
> java.util.concurrent.ExecutionException: 
> java.lang.ArrayIndexOutOfBoundsException: 0
>   at 
> org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.updateScanner(AbstractDataBlockIterator.java:136)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:50)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41)
>   at 
> org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31)
>   at 
> org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41)
>   at 
> org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:79)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
>   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>   at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>   at 
> 

[jira] [Created] (CARBONDATA-904) ArrayIndexOutOfBoundsException

2017-04-11 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-904:


 Summary: ArrayIndexOutOfBoundsException 
 Key: CARBONDATA-904
 URL: https://issues.apache.org/jira/browse/CARBONDATA-904
 Project: CarbonData
  Issue Type: Bug
Reporter: SWATI RAO
 Attachments: Test_Data1.csv, Test_Data1_h1.csv

Or operator is not working properly.

When we execute these query in hive it is working fine but when we execute the 
same in carbondata it throws an exception:
java.lang.ArrayIndexOutOfBoundsException

HIVE:
0: jdbc:hive2://hadoop-master:1> create table Test_Boundary_h1 (c1_int 
int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED 
FIELDS TERMINATED BY ',' ;
+-+--+
| result  |
+-+--+
+-+--+
No rows selected (1.177 seconds)
0: jdbc:hive2://hadoop-master:1> load data local inpath 
'/opt/Carbon/CarbonData/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE 
Test_Boundary_h1 ;
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.437 seconds)
0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) from 
Test_Boundary_h1 where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp 
='2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by 
c6_Timestamp ;
+++--+
|  c6_Timestamp  |  _c1   |
+++--+
| 2017-07-01 12:07:28.0  | 2017-07-01 12:07:28.0  |
+++--+
1 row selected (1.637 seconds)

CARBONDATA:
0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int 
int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
'org.apache.carbondata.format' ;
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (4.48 seconds)

0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 
'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table 
Test_Boundary 
OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
 ;
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (4.445 seconds)
0: jdbc:hive2://hadoop-master:1> select c6_Timestamp,max(c6_Timestamp) from 
Test_Boundary where c6_Timestamp ='2017-07-01 12:07:28' or c6_Timestamp =' 
2019-07-05 13:07:30' or c6_Timestamp = '1999-01-06 10:05:29' group by 
c6_Timestamp ;
Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 
0 in stage 5.0 failed 4 times, most recent failure: Lost task 0.3 in stage 5.0 
(TID 8, hadoop-master): java.lang.RuntimeException: 
java.util.concurrent.ExecutionException: 
java.lang.ArrayIndexOutOfBoundsException: 0
at 
org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.updateScanner(AbstractDataBlockIterator.java:136)
at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:50)
at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.next(DataBlockIteratorImpl.java:32)
at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.getBatchResult(DetailQueryResultIterator.java:50)
at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:41)
at 
org.apache.carbondata.core.scan.result.iterator.DetailQueryResultIterator.next(DetailQueryResultIterator.java:31)
at 
org.apache.carbondata.core.scan.result.iterator.ChunkRowIterator.(ChunkRowIterator.java:41)
at 
org.apache.carbondata.hadoop.CarbonRecordReader.initialize(CarbonRecordReader.java:79)
at 
org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:204)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
at 

[jira] [Updated] (CARBONDATA-883) Select query after alter is not working fine with integer datatype.

2017-04-07 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-883?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-883:
-
Description: 
CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, 
DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= 
"256 MB");

LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
table uniqdata OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');

ALTER TABLE uniqdata RENAME TO uniqdata1;


0: jdbc:hive2://192.168.2.126:1> alter table default.uniqdata1 add 
columns(ACTIVE_EMUI_VERSION int) 
TBLPROPERTIES('DEFAULT.VALUE.ACTIVE_EMUI_VERSION'='67890');
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.265 seconds)
0: jdbc:hive2://192.168.2.126:1> select distinct(ACTIVE_EMUI_VERSION) from 
uniqdata1 ;
Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 
0 in stage 22.0 failed 1 times, most recent failure: Lost task 0.0 in stage 
22.0 (TID 1211, localhost, executor driver): java.lang.ClassCastException: 
java.lang.Long cannot be cast to java.lang.Integer
at 
org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingMeasures(RestructureBasedVectorResultCollector.java:193)
at 
org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:98)
at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
at 
org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
at 
org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
at 
org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
at 
org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
 Source)
at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithKeys$(Unknown
 Source)
at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
 Source)
at 
org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
at 
org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
at org.apache.spark.scheduler.Task.run(Task.scala:99)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)

Driver stacktrace: (state=,code=0)
0: jdbc:hive2://192.168.2.126:1> 


> Select query after alter is not working fine with integer datatype.
> ---
>
> Key: CARBONDATA-883
> URL: https://issues.apache.org/jira/browse/CARBONDATA-883
> Project: CarbonData
>  Issue Type: Bug
> Environment: CREATE TABLE uniqdata (CUST_ID int,CUST_NAME 
> String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, 
> BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), 
> DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 
> double,INTEGER_COLUMN1 int) STORED BY 'org.apache.carbondata.format' 
> TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 

[jira] [Created] (CARBONDATA-883) Select query after alter is not working fine with integer datatype.

2017-04-07 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-883:


 Summary: Select query after alter is not working fine with integer 
datatype.
 Key: CARBONDATA-883
 URL: https://issues.apache.org/jira/browse/CARBONDATA-883
 Project: CarbonData
  Issue Type: Bug
 Environment: CREATE TABLE uniqdata (CUST_ID int,CUST_NAME 
String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 
bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= 
"256 MB");

LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
table uniqdata OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');

ALTER TABLE uniqdata RENAME TO uniqdata1;


0: jdbc:hive2://192.168.2.126:1> alter table default.uniqdata1 add 
columns(ACTIVE_EMUI_VERSION int) 
TBLPROPERTIES('DEFAULT.VALUE.ACTIVE_EMUI_VERSION'='67890');
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.265 seconds)
0: jdbc:hive2://192.168.2.126:1> select distinct(ACTIVE_EMUI_VERSION) from 
uniqdata1 ;
Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 
0 in stage 22.0 failed 1 times, most recent failure: Lost task 0.0 in stage 
22.0 (TID 1211, localhost, executor driver): java.lang.ClassCastException: 
java.lang.Long cannot be cast to java.lang.Integer
at 
org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingMeasures(RestructureBasedVectorResultCollector.java:193)
at 
org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:98)
at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
at 
org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
at 
org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
at 
org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
at 
org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
 Source)
at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithKeys$(Unknown
 Source)
at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
 Source)
at 
org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
at 
org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
at org.apache.spark.scheduler.Task.run(Task.scala:99)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)

Driver stacktrace: (state=,code=0)
0: jdbc:hive2://192.168.2.126:1> 

Reporter: SWATI RAO






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Issue Comment Deleted] (CARBONDATA-867) Shows an error on select query with date & timestamp datatype after altering the table.

2017-04-07 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-867:
-
Comment: was deleted

(was: Now it displays "NULL" values


0: jdbc:hive2://192.168.2.126:1> alter table uniqdata1 drop 
columns(ACTIVE_EMUI_VERSION);
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.28 seconds)
0: jdbc:hive2://192.168.2.126:1> alter table uniqdata1 add 
columns(ACTIVE_EMUI_VERSION timestamp) TBLPROPERTIES 
('DEFAULT.VALUE.ACTIVE_EMUI_VERSION'= '2017-01-01');
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.273 seconds)
0: jdbc:hive2://192.168.2.126:1> select distinct(ACTIVE_EMUI_VERSION) from 
uniqdata1 ;
+--+--+
| ACTIVE_EMUI_VERSION  |
+--+--+
| NULL |
+--+--+
1 row selected (0.816 seconds)
0: jdbc:hive2://192.168.2.126:1> 
)

> Shows an error on select query with date & timestamp datatype after altering 
> the table.
> ---
>
> Key: CARBONDATA-867
> URL: https://issues.apache.org/jira/browse/CARBONDATA-867
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.1.0-incubating
> Environment: Spark2.1
>Reporter: SWATI RAO
> Attachments: 2000_UniqData.csv
>
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 add columns(tmpstmp date) 
> TBLPROPERTIES('DEFAULT.VALUE.tmpstmp'= '2017-01-01');
> 0: jdbc:hive2://192.168.2.126:1> select distinct(tmpstmp) from uniqdata1 ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 44.0 failed 1 times, most recent failure: Lost task 0.0 in 
> stage 44.0 (TID 1038, localhost, executor driver): 
> java.lang.ClassCastException: java.lang.Integer cannot be cast to 
> java.lang.Long
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDirectDictionaryData(RestructureBasedVectorResultCollector.java:151)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingDimensions(RestructureBasedVectorResultCollector.java:114)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:97)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
>   at 
> org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithKeys$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
>  Source)
>   at 
> org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
>   at 
> org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
>   at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
>   at 
> org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
>   at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
>   at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
>   at org.apache.spark.scheduler.Task.run(Task.scala:99)
>   at 

[jira] [Commented] (CARBONDATA-867) Shows an error on select query with date & timestamp datatype after altering the table.

2017-04-07 Thread SWATI RAO (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-867?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15960436#comment-15960436
 ] 

SWATI RAO commented on CARBONDATA-867:
--

0: jdbc:hive2://192.168.2.126:1> alter table uniqdata1 drop 
columns(ACTIVE_EMUI_VERSION);
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.276 seconds)
0: jdbc:hive2://192.168.2.126:1> alter table uniqdata1 add 
columns(ACTIVE_EMUI_VERSION timestamp) TBLPROPERTIES 
('DEFAULT.VALUE.ACTIVE_EMUI_VERSION'= '2017-01-01 00-01-00');
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.279 seconds)
0: jdbc:hive2://192.168.2.126:1> select distinct(ACTIVE_EMUI_VERSION) from 
uniqdata1 ;
+--+--+
| ACTIVE_EMUI_VERSION  |
+--+--+
| NULL |
+--+--+
1 row selected (0.687 seconds)
0: jdbc:hive2://192.168.2.126:1> 

> Shows an error on select query with date & timestamp datatype after altering 
> the table.
> ---
>
> Key: CARBONDATA-867
> URL: https://issues.apache.org/jira/browse/CARBONDATA-867
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.1.0-incubating
> Environment: Spark2.1
>Reporter: SWATI RAO
> Attachments: 2000_UniqData.csv
>
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 add columns(tmpstmp date) 
> TBLPROPERTIES('DEFAULT.VALUE.tmpstmp'= '2017-01-01');
> 0: jdbc:hive2://192.168.2.126:1> select distinct(tmpstmp) from uniqdata1 ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 44.0 failed 1 times, most recent failure: Lost task 0.0 in 
> stage 44.0 (TID 1038, localhost, executor driver): 
> java.lang.ClassCastException: java.lang.Integer cannot be cast to 
> java.lang.Long
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDirectDictionaryData(RestructureBasedVectorResultCollector.java:151)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingDimensions(RestructureBasedVectorResultCollector.java:114)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:97)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
>   at 
> org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithKeys$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
>  Source)
>   at 
> org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
>   at 
> org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
>   at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
>   at 
> org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
>   at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
>   at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
>   at org.apache.spark.scheduler.Task.run(Task.scala:99)
>   at 

[jira] [Commented] (CARBONDATA-867) Shows an error on select query with date & timestamp datatype after altering the table.

2017-04-07 Thread SWATI RAO (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-867?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15960434#comment-15960434
 ] 

SWATI RAO commented on CARBONDATA-867:
--

Now it displays "NULL" values


0: jdbc:hive2://192.168.2.126:1> alter table uniqdata1 drop 
columns(ACTIVE_EMUI_VERSION);
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.28 seconds)
0: jdbc:hive2://192.168.2.126:1> alter table uniqdata1 add 
columns(ACTIVE_EMUI_VERSION timestamp) TBLPROPERTIES 
('DEFAULT.VALUE.ACTIVE_EMUI_VERSION'= '2017-01-01');
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.273 seconds)
0: jdbc:hive2://192.168.2.126:1> select distinct(ACTIVE_EMUI_VERSION) from 
uniqdata1 ;
+--+--+
| ACTIVE_EMUI_VERSION  |
+--+--+
| NULL |
+--+--+
1 row selected (0.816 seconds)
0: jdbc:hive2://192.168.2.126:1> 


> Shows an error on select query with date & timestamp datatype after altering 
> the table.
> ---
>
> Key: CARBONDATA-867
> URL: https://issues.apache.org/jira/browse/CARBONDATA-867
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.1.0-incubating
> Environment: Spark2.1
>Reporter: SWATI RAO
> Attachments: 2000_UniqData.csv
>
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 add columns(tmpstmp date) 
> TBLPROPERTIES('DEFAULT.VALUE.tmpstmp'= '2017-01-01');
> 0: jdbc:hive2://192.168.2.126:1> select distinct(tmpstmp) from uniqdata1 ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 44.0 failed 1 times, most recent failure: Lost task 0.0 in 
> stage 44.0 (TID 1038, localhost, executor driver): 
> java.lang.ClassCastException: java.lang.Integer cannot be cast to 
> java.lang.Long
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDirectDictionaryData(RestructureBasedVectorResultCollector.java:151)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingDimensions(RestructureBasedVectorResultCollector.java:114)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:97)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
>   at 
> org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithKeys$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
>  Source)
>   at 
> org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
>   at 
> org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
>   at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
>   at 
> org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
>   at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
>   at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
>   at org.apache.spark.scheduler.Task.run(Task.scala:99)
>   at 

[jira] [Issue Comment Deleted] (CARBONDATA-864) After adding column using alter query, when we put any column in "Dictionary Exclude" then perform select query on that column then it will throws an ex

2017-04-07 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-864?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-864:
-
Comment: was deleted

(was: When we put string datatype in dictionary_exclude it will display "NULL" 
values when perform select query.

CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, 
DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= 
"256 MB");



LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
table uniqdata OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');LOAD
 DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into table 
uniqdata OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');


ALTER TABLE uniqdata RENAME TO uniqdata1;


alter table uniqdata1 drop columns(ACTIVE_EMUI_VERSION);


alter table uniqdata1 add columns(ACTIVE_EMUI_VERSION string) 
TBLPROPERTIES('DICTIONARY_EXCLUDE'='ACTIVE_EMUI_VERSION', 
'DEFAULT.VALUE.(ACTIVE_EMUI_VERSION'='abcd');


select distinct(ACTIVE_EMUI_VERSION) from uniqdata1 ;

Results:

0: jdbc:hive2://192.168.2.126:1> alter table uniqdata1 drop 
columns(ACTIVE_EMUI_VERSION);
Error: java.lang.RuntimeException: Column active_emui_version does not exists 
in the table default.uniqdata1 (state=,code=0)
0: jdbc:hive2://192.168.2.126:1>
0: jdbc:hive2://192.168.2.126:1> alter table uniqdata1 add 
columns(ACTIVE_EMUI_VERSION string) 
TBLPROPERTIES('DICTIONARY_EXCLUDE'='ACTIVE_EMUI_VERSION', 
'DEFAULT.VALUE.(ACTIVE_EMUI_VERSION'='abcd');
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.308 seconds)
0: jdbc:hive2://192.168.2.126:1> select distinct(ACTIVE_EMUI_VERSION) from 
uniqdata1 ;
+--+--+
| ACTIVE_EMUI_VERSION  |
+--+--+
| NULL |
+--+--+
1 row selected (0.88 seconds)
0: jdbc:hive2://192.168.2.126:1>
)

> After adding column using alter query, when we put any column in "Dictionary 
> Exclude" then perform select query on that column then it will throws an 
> exception.
> 
>
> Key: CARBONDATA-864
> URL: https://issues.apache.org/jira/browse/CARBONDATA-864
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.1.0-incubating
> Environment: Spark2.1
>Reporter: SWATI RAO
> Attachments: 2000_UniqData.csv, error.png
>
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 add columns(dict int) 
> TBLPROPERTIES('DICTIONARY_INCLUDE'='dict','DEFAULT.VALUE.dict'= '');
> select distinct(dict) from uniqdata1 ;
> when we perform select query on "DICTIONARY_INCLUDE" it is working
> but when we perform select query on "DICTIONARY_EXCLUDE" then it will throws 
> an exception:
> alter table uniqdata1 add columns(nodict string) 
> TBLPROPERTIES('DICTIONARY_EXCLUDE'='nodict', 'DEFAULT.VALUE.NoDict'= 'abcd');
> select distinct(nodict) from uniqdata1 ;
> 0: jdbc:hive2://192.168.2.126:1> select distinct(nodict) from uniqdata1 ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 21.0 failed 1 times, most recent failure: Lost task 0.0 in 
> stage 21.0 (TID 419, localhost, executor driver): 
> java.lang.ArrayIndexOutOfBoundsException: 4186
>   at 
> org.apache.spark.sql.execution.vectorized.OnHeapColumnVector.putByteArray(OnHeapColumnVector.java:401)
>   at 
> 

[jira] [Commented] (CARBONDATA-864) After adding column using alter query, when we put any column in "Dictionary Exclude" then perform select query on that column then it will throws an exception.

2017-04-07 Thread SWATI RAO (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-864?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15960416#comment-15960416
 ] 

SWATI RAO commented on CARBONDATA-864:
--

When we put string datatype in dictionary_exclude it will display "NULL" values 
when perform select query.

CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, 
DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= 
"256 MB");



LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
table uniqdata OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');LOAD
 DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into table 
uniqdata OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');


ALTER TABLE uniqdata RENAME TO uniqdata1;


alter table uniqdata1 drop columns(ACTIVE_EMUI_VERSION);


alter table uniqdata1 add columns(ACTIVE_EMUI_VERSION string) 
TBLPROPERTIES('DICTIONARY_EXCLUDE'='ACTIVE_EMUI_VERSION', 
'DEFAULT.VALUE.(ACTIVE_EMUI_VERSION'='abcd');


select distinct(ACTIVE_EMUI_VERSION) from uniqdata1 ;

Results:

0: jdbc:hive2://192.168.2.126:1> alter table uniqdata1 drop 
columns(ACTIVE_EMUI_VERSION);
Error: java.lang.RuntimeException: Column active_emui_version does not exists 
in the table default.uniqdata1 (state=,code=0)
0: jdbc:hive2://192.168.2.126:1>
0: jdbc:hive2://192.168.2.126:1> alter table uniqdata1 add 
columns(ACTIVE_EMUI_VERSION string) 
TBLPROPERTIES('DICTIONARY_EXCLUDE'='ACTIVE_EMUI_VERSION', 
'DEFAULT.VALUE.(ACTIVE_EMUI_VERSION'='abcd');
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.308 seconds)
0: jdbc:hive2://192.168.2.126:1> select distinct(ACTIVE_EMUI_VERSION) from 
uniqdata1 ;
+--+--+
| ACTIVE_EMUI_VERSION  |
+--+--+
| NULL |
+--+--+
1 row selected (0.88 seconds)
0: jdbc:hive2://192.168.2.126:1>


> After adding column using alter query, when we put any column in "Dictionary 
> Exclude" then perform select query on that column then it will throws an 
> exception.
> 
>
> Key: CARBONDATA-864
> URL: https://issues.apache.org/jira/browse/CARBONDATA-864
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.1.0-incubating
> Environment: Spark2.1
>Reporter: SWATI RAO
> Attachments: 2000_UniqData.csv, error.png
>
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 add columns(dict int) 
> TBLPROPERTIES('DICTIONARY_INCLUDE'='dict','DEFAULT.VALUE.dict'= '');
> select distinct(dict) from uniqdata1 ;
> when we perform select query on "DICTIONARY_INCLUDE" it is working
> but when we perform select query on "DICTIONARY_EXCLUDE" then it will throws 
> an exception:
> alter table uniqdata1 add columns(nodict string) 
> TBLPROPERTIES('DICTIONARY_EXCLUDE'='nodict', 'DEFAULT.VALUE.NoDict'= 'abcd');
> select distinct(nodict) from uniqdata1 ;
> 0: jdbc:hive2://192.168.2.126:1> select distinct(nodict) from uniqdata1 ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 21.0 failed 1 times, most recent failure: Lost task 0.0 in 
> stage 21.0 (TID 419, localhost, executor driver): 
> java.lang.ArrayIndexOutOfBoundsException: 4186
>   at 
> org.apache.spark.sql.execution.vectorized.OnHeapColumnVector.putByteArray(OnHeapColumnVector.java:401)
>   at 
> 

[jira] [Updated] (CARBONDATA-877) String datatype is throwing an error when included in DIctionary_Exclude in a alter query

2017-04-06 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-877?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-877:
-
Description: 
CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, 
DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= 
"256 MB");

LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
table uniqdata OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');

ALTER TABLE uniqdata RENAME TO uniqdata1;

alter table uniqdata1 drop columns(CUST_NAME);

alter table uniqdata1 add columns(CUST_NAME string) 
TBLPROPERTIES('DICTIONARY_EXCLUDE'='CUST_NAME', 
'DEFAULT.VALUE.CUST_NAME'='testuser') ;

Column added successfully. But when we execute:
select distinct(CUST_NAME) from uniqdata1 ; 
&
select count(CUST_NAME) from uniqdata1 ;

it throws an error :
"Job aborted due to stage failure: Task 0 in stage 9.0 failed 1 times, most 
recent failure: Lost task 0.0 in stage 9.0 (TID 206, localhost, executor 
driver): java.lang.ArrayIndexOutOfBoundsException: 4186"



  was:
CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, 
DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= 
"256 MB");

LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
table uniqdata OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');

ALTER TABLE uniqdata RENAME TO uniqdata1;

alter table uniqdata1 drop columns(CUST_NAME);

alter table uniqdata1 add columns(CUST_NAME string) 
TBLPROPERTIES('DICTIONARY_EXCLUDE'='CUST_NAME', 
'DEFAULT.VALUE.CUST_NAME'='testuser') ;

Column added successfully. But when we execute:
select distinct(CUST_NAME) from uniqdata1 ; 
&
select count(CUST_NAME) from uniqdata1 ;

it throws an error :
"Job aborted due to stage failure: Task 0 in stage 9.0 failed 1 times, most 
recent failure: Lost task 0.0 in stage 9.0 (TID 206, localhost, executor 
driver): java.lang.ArrayIndexOutOfBoundsException: 4186"




alter table uniqdata1 add columns(CUST_NAME string) 
TBLPROPERTIES('DICTIONARY_EXCLUDE'='CUST_NAME', 
'DEFAULT.VALUE.CUST_NAME'='testuser');



> String datatype is throwing an error when included in DIctionary_Exclude in a 
> alter query
> -
>
> Key: CARBONDATA-877
> URL: https://issues.apache.org/jira/browse/CARBONDATA-877
> Project: CarbonData
>  Issue Type: Bug
>Reporter: SWATI RAO
> Attachments: 2000_UniqData.csv
>
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 drop columns(CUST_NAME);
> alter table uniqdata1 add columns(CUST_NAME string) 
> TBLPROPERTIES('DICTIONARY_EXCLUDE'='CUST_NAME', 
> 'DEFAULT.VALUE.CUST_NAME'='testuser') ;
> Column added successfully. But when we execute:
> select distinct(CUST_NAME) from uniqdata1 ; 
> &
> select count(CUST_NAME) from uniqdata1 ;
> it throws an error :
> "Job aborted due to stage failure: Task 0 in stage 9.0 failed 1 times, most 
> recent failure: Lost task 0.0 in stage 9.0 (TID 206, localhost, executor 
> driver): java.lang.ArrayIndexOutOfBoundsException: 4186"



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-877) String datatype is throwing an error when included in DIctionary_Exclude in a alter query

2017-04-06 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-877:


 Summary: String datatype is throwing an error when included in 
DIctionary_Exclude in a alter query
 Key: CARBONDATA-877
 URL: https://issues.apache.org/jira/browse/CARBONDATA-877
 Project: CarbonData
  Issue Type: Bug
Reporter: SWATI RAO
 Attachments: 2000_UniqData.csv

CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, 
DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= 
"256 MB");

LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
table uniqdata OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');

ALTER TABLE uniqdata RENAME TO uniqdata1;

alter table uniqdata1 drop columns(CUST_NAME);

alter table uniqdata1 add columns(CUST_NAME string) 
TBLPROPERTIES('DICTIONARY_EXCLUDE'='CUST_NAME', 
'DEFAULT.VALUE.CUST_NAME'='testuser') ;

Column added successfully. But when we execute:
select distinct(CUST_NAME) from uniqdata1 ; 
&
select count(CUST_NAME) from uniqdata1 ;

it throws an error :
"Job aborted due to stage failure: Task 0 in stage 9.0 failed 1 times, most 
recent failure: Lost task 0.0 in stage 9.0 (TID 206, localhost, executor 
driver): java.lang.ArrayIndexOutOfBoundsException: 4186"




alter table uniqdata1 add columns(CUST_NAME string) 
TBLPROPERTIES('DICTIONARY_EXCLUDE'='CUST_NAME', 
'DEFAULT.VALUE.CUST_NAME'='testuser');




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-868) Select query on decimal datatype is not working fine after adding decimal column using alter

2017-04-05 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-868:


 Summary: Select query on decimal datatype is not working fine 
after adding decimal column using alter
 Key: CARBONDATA-868
 URL: https://issues.apache.org/jira/browse/CARBONDATA-868
 Project: CarbonData
  Issue Type: Bug
Affects Versions: 1.1.0-incubating
 Environment: Spark2.1
Reporter: SWATI RAO
 Attachments: 2000_UniqData.csv

CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, 
DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= 
"256 MB");

LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
table uniqdata OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');

ALTER TABLE uniqdata RENAME TO uniqdata1;

alter table uniqdata1 add columns(msrField 
decimal(5,2))TBLPROPERTIES('DEFAULT.VALUE.msrfield'= '123.45');

0: jdbc:hive2://192.168.2.126:1> select msrField from uniqdata1;
Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 
0 in stage 48.0 failed 1 times, most recent failure: Lost task 0.0 in stage 
48.0 (TID 1041, localhost, executor driver): 
java.lang.ArrayIndexOutOfBoundsException: 4186
at 
org.apache.spark.sql.execution.vectorized.OnHeapColumnVector.putInt(OnHeapColumnVector.java:202)
at 
org.apache.spark.sql.execution.vectorized.ColumnVector.putDecimal(ColumnVector.java:608)
at 
org.apache.carbondata.spark.vectorreader.ColumnarVectorWrapper.putDecimal(ColumnarVectorWrapper.java:58)
at 
org.apache.carbondata.spark.vectorreader.ColumnarVectorWrapper.putDecimals(ColumnarVectorWrapper.java:64)
at 
org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingMeasures(RestructureBasedVectorResultCollector.java:202)
at 
org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:98)
at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
at 
org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
at 
org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
at 
org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
at 
org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
 Source)
at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
 Source)
at 
org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
at 
org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:231)
at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:225)
at 
org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:826)
at 
org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:826)
at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
at org.apache.spark.scheduler.Task.run(Task.scala:99)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)

Driver stacktrace: (state=,code=0)




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Updated] (CARBONDATA-864) After adding column using alter query, when we put any column in "Dictionary Exclude" then perform select query on that column then it will throws an exception.

2017-04-05 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-864?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-864:
-
Request participants:   (was: )
 Summary: After adding column using alter query, when we put 
any column in "Dictionary Exclude" then perform select query on that column 
then it will throws an exception.  (was: If we put any column in "Dictionary 
Exclude" then perform select query on that column then it will throws an 
exception.)

> After adding column using alter query, when we put any column in "Dictionary 
> Exclude" then perform select query on that column then it will throws an 
> exception.
> 
>
> Key: CARBONDATA-864
> URL: https://issues.apache.org/jira/browse/CARBONDATA-864
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.1.0-incubating
> Environment: Spark2.1
>Reporter: SWATI RAO
> Attachments: 2000_UniqData.csv, error.png
>
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 add columns(dict int) 
> TBLPROPERTIES('DICTIONARY_INCLUDE'='dict','DEFAULT.VALUE.dict'= '');
> select distinct(dict) from uniqdata1 ;
> when we perform select query on "DICTIONARY_INCLUDE" it is working
> but when we perform select query on "DICTIONARY_EXCLUDE" then it will throws 
> an exception:
> alter table uniqdata1 add columns(nodict string) 
> TBLPROPERTIES('DICTIONARY_EXCLUDE'='nodict', 'DEFAULT.VALUE.NoDict'= 'abcd');
> select distinct(nodict) from uniqdata1 ;
> 0: jdbc:hive2://192.168.2.126:1> select distinct(nodict) from uniqdata1 ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 21.0 failed 1 times, most recent failure: Lost task 0.0 in 
> stage 21.0 (TID 419, localhost, executor driver): 
> java.lang.ArrayIndexOutOfBoundsException: 4186
>   at 
> org.apache.spark.sql.execution.vectorized.OnHeapColumnVector.putByteArray(OnHeapColumnVector.java:401)
>   at 
> org.apache.spark.sql.execution.vectorized.ColumnVector.putByteArray(ColumnVector.java:569)
>   at 
> org.apache.carbondata.spark.vectorreader.ColumnarVectorWrapper.putBytes(ColumnarVectorWrapper.java:77)
>   at 
> org.apache.carbondata.spark.vectorreader.ColumnarVectorWrapper.putBytes(ColumnarVectorWrapper.java:83)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillNoDictionaryData(RestructureBasedVectorResultCollector.java:167)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingDimensions(RestructureBasedVectorResultCollector.java:122)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:97)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
>   at 
> org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithKeys$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
>  Source)
>   at 
> org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
>   at 
> 

[jira] [Updated] (CARBONDATA-867) Shows an error on select query with date & timestamp datatype after altering the table.

2017-04-05 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-867:
-
Request participants:   (was: )
 Summary: Shows an error on select query with date & timestamp 
datatype after altering the table.  (was: Shows an error on select query with 
date & timestamp datatype)

> Shows an error on select query with date & timestamp datatype after altering 
> the table.
> ---
>
> Key: CARBONDATA-867
> URL: https://issues.apache.org/jira/browse/CARBONDATA-867
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.1.0-incubating
> Environment: Spark2.1
>Reporter: SWATI RAO
> Attachments: 2000_UniqData.csv
>
>
> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
> string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
> bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
> decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
> int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
> ("TABLE_BLOCKSIZE"= "256 MB");
> LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
> table uniqdata OPTIONS('DELIMITER'=',' , 
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> ALTER TABLE uniqdata RENAME TO uniqdata1;
> alter table uniqdata1 add columns(tmpstmp date) 
> TBLPROPERTIES('DEFAULT.VALUE.tmpstmp'= '2017-01-01');
> 0: jdbc:hive2://192.168.2.126:1> select distinct(tmpstmp) from uniqdata1 ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 44.0 failed 1 times, most recent failure: Lost task 0.0 in 
> stage 44.0 (TID 1038, localhost, executor driver): 
> java.lang.ClassCastException: java.lang.Integer cannot be cast to 
> java.lang.Long
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDirectDictionaryData(RestructureBasedVectorResultCollector.java:151)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingDimensions(RestructureBasedVectorResultCollector.java:114)
>   at 
> org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:97)
>   at 
> org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
>   at 
> org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
>   at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
>   at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithKeys$(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
>  Source)
>   at 
> org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
>   at 
> org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
>   at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
>   at 
> org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
>   at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
>   at 
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
>   at org.apache.spark.scheduler.Task.run(Task.scala:99)
>   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)
> Driver stacktrace: (state=,code=0)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-867) Shows an error on select query with date & timestamp datatype

2017-04-05 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-867:


 Summary: Shows an error on select query with date & timestamp 
datatype
 Key: CARBONDATA-867
 URL: https://issues.apache.org/jira/browse/CARBONDATA-867
 Project: CarbonData
  Issue Type: Bug
Affects Versions: 1.1.0-incubating
 Environment: Spark2.1
Reporter: SWATI RAO
 Attachments: 2000_UniqData.csv

CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, 
DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= 
"256 MB");

LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
table uniqdata OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');

ALTER TABLE uniqdata RENAME TO uniqdata1;

alter table uniqdata1 add columns(tmpstmp date) 
TBLPROPERTIES('DEFAULT.VALUE.tmpstmp'= '2017-01-01');

0: jdbc:hive2://192.168.2.126:1> select distinct(tmpstmp) from uniqdata1 ;
Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 
0 in stage 44.0 failed 1 times, most recent failure: Lost task 0.0 in stage 
44.0 (TID 1038, localhost, executor driver): java.lang.ClassCastException: 
java.lang.Integer cannot be cast to java.lang.Long
at 
org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDirectDictionaryData(RestructureBasedVectorResultCollector.java:151)
at 
org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.fillDataForNonExistingDimensions(RestructureBasedVectorResultCollector.java:114)
at 
org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectVectorBatch(RestructureBasedVectorResultCollector.java:97)
at 
org.apache.carbondata.core.scan.processor.impl.DataBlockIteratorImpl.processNextBatch(DataBlockIteratorImpl.java:65)
at 
org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:46)
at 
org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:246)
at 
org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:140)
at 
org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:222)
at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
 Source)
at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.agg_doAggregateWithKeys$(Unknown
 Source)
at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
 Source)
at 
org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
at 
org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:377)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
at org.apache.spark.scheduler.Task.run(Task.scala:99)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)

Driver stacktrace: (state=,code=0)




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-847) Select query not working properly after alter.

2017-04-03 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-847:


 Summary: Select query not working properly after alter.
 Key: CARBONDATA-847
 URL: https://issues.apache.org/jira/browse/CARBONDATA-847
 Project: CarbonData
  Issue Type: Bug
Affects Versions: 1.1.0-incubating
 Environment: Spark2.1
Reporter: SWATI RAO
 Attachments: 2000_UniqData.csv

Execute these set of queries: 

CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, 
DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= 
"256 MB");

LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
table uniqdata OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');

ALTER TABLE uniqdata RENAME TO uniqdata1;

alter table uniqdata1 add columns(dict int) 
TBLPROPERTIES('DICTIONARY_INCLUDE'='dict','DEFAULT.VALUE.dict'= '');

select distinct(dict) from uniqdata2 ;

it will display the result but when we perform :

select * from uniqdata1 ;

it will display an error message :

Job aborted due to stage failure: Task 3 in stage 59.0 failed 1 times, most 
recent failure: Lost task 3.0 in stage 59.0 (TID 714, localhost, executor 
driver): java.lang.NullPointerException



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Closed] (CARBONDATA-370) Abnormal behaviour of datatype Int & bigInt

2017-03-02 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-370.

Resolution: Fixed

> Abnormal behaviour of datatype Int & bigInt
> ---
>
> Key: CARBONDATA-370
> URL: https://issues.apache.org/jira/browse/CARBONDATA-370
> Project: CarbonData
>  Issue Type: Bug
>Reporter: SWATI RAO
>Assignee: anubhav tarar
>
> The specified data-type range for Int & bigInt is:
> INT –––> -2147483648 to 2147483647
> BIGINT ––> -922337203685477580807 to 922337203685477580806
>  but I found abnormal behaviour in these data-type . 
> 1.When I insert beyond range in Int it allows range even beyond bigInt. 
> 2.When I insert into bigInt it doesn’t accept all values following in its 
> range.
>  



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-712) 'BAD_RECORDS_ACTION'='REDIRECT' is not working properly.

2017-02-17 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-712:


 Summary: 'BAD_RECORDS_ACTION'='REDIRECT' is not working properly.
 Key: CARBONDATA-712
 URL: https://issues.apache.org/jira/browse/CARBONDATA-712
 Project: CarbonData
  Issue Type: Bug
  Components: data-load
Affects Versions: 1.0.0-incubating
 Environment: Spark 1.6
Reporter: SWATI RAO
Priority: Minor
 Attachments: 2000_UniqData.csv, uniq_shared_dictionary.zip

When we tried to load data using 'BAD_RECORDS_ACTION'='REDIRECT' the bad record 
were not written in the file. I have set the property as 
carbon.badRecords.location=/opt/Carbon/Spark/badrecords.
PFA for bad_record files which is empty & query csv.

CREATE TABLE uniq_shared_dictionary (CUST_ID int,CUST_NAME 
String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 
bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format' 
TBLPROPERTIES('DICTIONARY_INCLUDE'='CUST_ID,Double_COLUMN2,DECIMAL_COLUMN2','columnproperties.CUST_ID.shared_column'='shared.CUST_ID','columnproperties.decimal_column2.shared_column'='shared.decimal_column2')

LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
table uniq_shared_dictionary OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_LOGGER_ENABLE'='TRUE', 
'BAD_RECORDS_ACTION'='REDIRECT','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1','SINGLE_Pass'='true')

select Cust_Id from uniq_shared_dictionary ;

The bad_records are also not loaded in the table.





--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-711) Inconsistent data load when single_pass='true'

2017-02-17 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-711:


 Summary: Inconsistent data load when single_pass='true'
 Key: CARBONDATA-711
 URL: https://issues.apache.org/jira/browse/CARBONDATA-711
 Project: CarbonData
  Issue Type: Bug
  Components: data-load
Affects Versions: 1.1.0-incubating
 Environment: Spark 1.6
Reporter: SWATI RAO
Priority: Minor
 Attachments: 2000_UniqData.csv

 When we perform dataload with Single_pass='true' , it repeats some of the 
values in the table whereas the csv contains empty value for that column. PFA 
csv which is used for dataloading. And below is the create , load . and select 
query.


CREATE TABLE uniq_shared_dictionary (CUST_ID int,CUST_NAME 
String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 
bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format' 
TBLPROPERTIES('DICTIONARY_INCLUDE'='CUST_ID,Double_COLUMN2,DECIMAL_COLUMN2','columnproperties.CUST_ID.shared_column'='shared.CUST_ID','columnproperties.decimal_column2.shared_column'='shared.decimal_column2')
 ;



 LOAD DATA INPATH 
'hdfs://192.168.2.145:54310/BabuStore/Data/uniqdata/2000_UniqData.csv' into 
table uniq_shared_dictionary OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_LOGGER_ENABLE'='TRUE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1','SINGLE_Pass'='true')
 ;

Output: 
0: jdbc:hive2://hadoop-master:1> select CUST_ID from uniq_shared_dictionary 
;
+--+--+
| Cust_Id  |
+--+--+
| 8999 |
| 8999 |
| 8999 |
| 8999 |
| 8999 |
| 8999 |
| 8999 |
| 8999 |
| 8999 |
| 8999 |
| 8999 |
| 8999 |
| 9000 |
| 9001 |
| 9002 |
| 9003 |
| 9004 |
| 9005 |
| 9006 |
| 9007 |
| 9008 |
| 9009 |
| 9010 |
| 9011 |
| 9012 |
| 9013 |
| 9014 |
| 9015 |
| 9016 |
| 9017 |
| 9018 |
| 9019 |
| 9020 |
| 9021 |
| 9022 |
| 9023 |
| 9024 |
| 9025 |
| 9026 |
| 9027 |
| 9028 |
| 9029 |
| 9030 |
| 9031 |
| 9032 |
| 9033 |
| 9034 |
| 9035 |
| 9036 |
| 9037 |
| 9038 |
| 9039 |
| 9040 |
| 9041 |
| 9042 |
| 9043 |
| 9044 |
| 9045 |
| 9046 |
| 9047 |
| 9048 |
| 9049 |
| 9050 |
| 9051 |
| 9052 |
| 9053 |
| 9054 |
| 9055 |
| 9056 |
| 9057 |
| 9058 |
| 9059 |
| 9060 |
| 9061 |
| 9062 |
| 9063 |
| 9064 |
| 9065 |
| 9066 |
| 9067 |
| 9068 |
| 9069 |
| 9070 |
| 9071 |
| 9072 |
| 9073 |
| 9074 |
| 9075 |
| 9076 |
| 9077 |
| 9078 |
| 9079 |
| 9080 |
| 9081 |
| 9082 |
| 9083 |
| 9084 |
| 9085 |
| 9086 |
| 9087 |
+--+--+
| Cust_Id  |
+--+--+
| 9088 |
| 9089 |
| 9090 |
| 9091 |
| 9092 |
| 9093 |
| 9094 |
| 9095 |
| 9096 |
| 9097 |
| 9098 |
| 9099 |
| 9100 |
| 9101 |
| 9102 |
| 9103 |
| 9104 |
| 9105 |
| 9106 |
| 9107 |
| 9108 |
| 9109 |
| 9110 |
| 9111 |
| 9112 |
| 9113 |
| 9114 |
| 9115 |
| 9116 |
| 9117 |
| 9118 |
| 9119 |
| 9120 |
| 9121 |
| 9122 |
| 9123 |
| 9124 |
| 9125 |
| 9126 |
| 9127 |
| 9128 |
| 9129 |
| 9130 |
| 9131 |
| 9132 |
| 9133 |
| 9134 |
| 9135 |
| 9136 |
| 9137 |
| 9138 |
| 9139 |
| 9140 |
| 9141 |
| 9142 |
| 9143 |
| 9144 |
| 9145 |
| 9146 |
| 9147 |
| 9148 |
| 9149 |
| 9150 |
| 9151 |
| 9152 |
| 9153 |
| 9154 |
| 9155 |
| 9156 |
| 9157 |
| 9158 |
| 9159 |
| 9160 |
| 9161 |
| 9162 |
| 9163 |
| 9164 |
| 9165 |
| 9166 |
| 9167 |
| 9168 |
| 9169 |
| 9170 |
| 9171 |
| 9172 |
| 9173 |
| 9174 |
| 9175 |
| 9176 |
| 9177 |
| 9178 |
| 9179 |
| 9180 |
| 9181 |
| 9182 |
| 9183 |
| 9184 |
| 9185 |
| 9186 |
| 9187 |
+--+--+
| Cust_Id  |
+--+--+
| 9188 |
| 9189 |
| 9190 |
| 9191 |
| 9192 |
| 9193 |
| 9194 |
| 9195 |
| 9196 |
| 9197 |
| 9198 |
| 9199 |
| 9200 |
| 9201 |
| 9202 |
| 9203 |
| 9204 |
| 9205 |
| 9206 |
| 9207 |
| 9208 |
| 9209 |
| 9210 |
| 9211 |
| 9212 |
| 9213 |
| 9214 |
| 9215 |
| 9216 |
| 9217 |
| 9218 |
| 9219 |
| 9220 |
| 9221 |
| 9222 |

[jira] [Created] (CARBONDATA-708) Between operator does not work properly in carbondata.

2017-02-15 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-708:


 Summary: Between operator does not work properly in carbondata.
 Key: CARBONDATA-708
 URL: https://issues.apache.org/jira/browse/CARBONDATA-708
 Project: CarbonData
  Issue Type: Bug
  Components: sql
Affects Versions: 1.1.0-incubating
 Environment: Spark 2.1
Reporter: SWATI RAO
Priority: Minor
 Attachments: 100_hive_test.csv

Incorrect result displays.

Steps to reproduce:

1:Create table using following Command

" create table Carbon_automation (imei string,deviceInformationId int,MAC 
string,deviceColor string,device_backColor string,modelId string,marketName 
string,AMSize string,ROMSize string,CUPAudit string,CPIClocked string,series 
string,productionDate timestamp,bomCode string,internalModels string, 
deliveryTime string, channelsId string, channelsName string , deliveryAreaId 
string, deliveryCountry string, deliveryProvince string, deliveryCity 
string,deliveryDistrict string, deliveryStreet string, oxSingleNumber string, 
ActiveCheckTime string, ActiveAreaId string, ActiveCountry string, 
ActiveProvince string, Activecity string, ActiveDistrict string, ActiveStreet 
string, ActiveOperatorId string, Active_releaseId string, Active_EMUIVersion 
string, Active_operaSysVersion string, Active_BacVerNumber string, 
Active_BacFlashVer string, Active_webUIVersion string, Active_webUITypeCarrVer 
string,Active_webTypeDataVerNumber string, Active_operatorsVersion string, 
Active_phonePADPartitionedVersions string, Latest_YEAR int, Latest_MONTH int, 
Latest_DAY int, Latest_HOUR string, Latest_areaId string, Latest_country 
string, Latest_province string, Latest_city string, Latest_district string, 
Latest_street string, Latest_releaseId string, Latest_EMUIVersion string, 
Latest_operaSysVersion string, Latest_BacVerNumber string, Latest_BacFlashVer 
string, Latest_webUIVersion string, Latest_webUITypeCarrVer string, 
Latest_webTypeDataVerNumber string, Latest_operatorsVersion string, 
Latest_phonePADPartitionedVersions string, Latest_operatorId string, 
gamePointDescription string,gamePointId double,contractNumber double,imei_count 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
('DICTIONARY_INCLUDE'='deviceInformationId,Latest_YEAR,Latest_MONTH,Latest_DAY')"

2:Load Data with following command

" LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/HiveData' INTO TABLE 
Carbon_automation 
OPTIONS('DELIMITER'=',','QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='imei,deviceInformationId,MAC,deviceColor,device_backColor,modelId,marketName,AMSize,ROMSize,CUPAudit,CPIClocked,series,productionDate,bomCode,internalModels,deliveryTime,channelsId,channelsName,deliveryAreaId,deliveryCountry,deliveryProvince,deliveryCity,deliveryDistrict,deliveryStreet,oxSingleNumber,contractNumber,ActiveCheckTime,ActiveAreaId,ActiveCountry,ActiveProvince,Activecity,ActiveDistrict,ActiveStreet,ActiveOperatorId,Active_releaseId,Active_EMUIVersion,Active_operaSysVersion,Active_BacVerNumber,Active_BacFlashVer,Active_webUIVersion,Active_webUITypeCarrVer,Active_webTypeDataVerNumber,Active_operatorsVersion,Active_phonePADPartitionedVersions,Latest_YEAR,Latest_MONTH,Latest_DAY,Latest_HOUR,Latest_areaId,Latest_country,Latest_province,Latest_city,Latest_district,Latest_street,Latest_releaseId,Latest_EMUIVersion,Latest_operaSysVersion,Latest_BacVerNumber,Latest_BacFlashVer,Latest_webUIVersion,Latest_webUITypeCarrVer,Latest_webTypeDataVerNumber,Latest_operatorsVersion,Latest_phonePADPartitionedVersions,Latest_operatorId,gamePointId,gamePointDescription,imei_count')"

3:Run the Query
select Latest_DAY,Latest_HOUR,count(distinct AMSize) as 
AMSize_number,sum(gamePointId+contractNumber) as total from Carbon_automation 
where Latest_HOUR between 12 and 15 group by Latest_DAY,Latest_HOUR order by 
total desc

4:No Result display:
+-+--+++--+
| Latest_DAY  | Latest_HOUR  | AMSize_number  | total  |
+-+--+++--+
+-+--+++--+
No rows selected (2.133 seconds).

5:CSV Attached "100_hive_test.csv"

Expected Result:Correct Result should be display.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (CARBONDATA-707) Less ( < ) than operator does not work properly in carbondata.

2017-02-15 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-707:


 Summary: Less ( < ) than operator does not work properly in 
carbondata. 
 Key: CARBONDATA-707
 URL: https://issues.apache.org/jira/browse/CARBONDATA-707
 Project: CarbonData
  Issue Type: Bug
  Components: sql
Affects Versions: 1.1.0-incubating
 Environment: Spark 2.1
Reporter: SWATI RAO
Priority: Minor
 Attachments: 100_hive_test.csv

Incorrect result displays 

Steps to Reproduces:
1:Create table using following Command
" create table Carbon_automation (imei string,deviceInformationId int,MAC 
string,deviceColor string,device_backColor string,modelId string,marketName 
string,AMSize string,ROMSize string,CUPAudit string,CPIClocked string,series 
string,productionDate timestamp,bomCode string,internalModels string, 
deliveryTime string, channelsId string, channelsName string , deliveryAreaId 
string, deliveryCountry string, deliveryProvince string, deliveryCity 
string,deliveryDistrict string, deliveryStreet string, oxSingleNumber string, 
ActiveCheckTime string, ActiveAreaId string, ActiveCountry string, 
ActiveProvince string, Activecity string, ActiveDistrict string, ActiveStreet 
string, ActiveOperatorId string, Active_releaseId string, Active_EMUIVersion 
string, Active_operaSysVersion string, Active_BacVerNumber string, 
Active_BacFlashVer string, Active_webUIVersion string, Active_webUITypeCarrVer 
string,Active_webTypeDataVerNumber string, Active_operatorsVersion string, 
Active_phonePADPartitionedVersions string, Latest_YEAR int, Latest_MONTH int, 
Latest_DAY int, Latest_HOUR string, Latest_areaId string, Latest_country 
string, Latest_province string, Latest_city string, Latest_district string, 
Latest_street string, Latest_releaseId string, Latest_EMUIVersion string, 
Latest_operaSysVersion string, Latest_BacVerNumber string, Latest_BacFlashVer 
string, Latest_webUIVersion string, Latest_webUITypeCarrVer string, 
Latest_webTypeDataVerNumber string, Latest_operatorsVersion string, 
Latest_phonePADPartitionedVersions string, Latest_operatorId string, 
gamePointDescription string,gamePointId double,contractNumber double,imei_count 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
('DICTIONARY_INCLUDE'='deviceInformationId,Latest_YEAR,Latest_MONTH,Latest_DAY')"

2:Load Data with following command
" LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/HiveData' INTO TABLE 
Carbon_automation 
OPTIONS('DELIMITER'=',','QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='imei,deviceInformationId,MAC,deviceColor,device_backColor,modelId,marketName,AMSize,ROMSize,CUPAudit,CPIClocked,series,productionDate,bomCode,internalModels,deliveryTime,channelsId,channelsName,deliveryAreaId,deliveryCountry,deliveryProvince,deliveryCity,deliveryDistrict,deliveryStreet,oxSingleNumber,contractNumber,ActiveCheckTime,ActiveAreaId,ActiveCountry,ActiveProvince,Activecity,ActiveDistrict,ActiveStreet,ActiveOperatorId,Active_releaseId,Active_EMUIVersion,Active_operaSysVersion,Active_BacVerNumber,Active_BacFlashVer,Active_webUIVersion,Active_webUITypeCarrVer,Active_webTypeDataVerNumber,Active_operatorsVersion,Active_phonePADPartitionedVersions,Latest_YEAR,Latest_MONTH,Latest_DAY,Latest_HOUR,Latest_areaId,Latest_country,Latest_province,Latest_city,Latest_district,Latest_street,Latest_releaseId,Latest_EMUIVersion,Latest_operaSysVersion,Latest_BacVerNumber,Latest_BacFlashVer,Latest_webUIVersion,Latest_webUITypeCarrVer,Latest_webTypeDataVerNumber,Latest_operatorsVersion,Latest_phonePADPartitionedVersions,Latest_operatorId,gamePointId,gamePointDescription,imei_count')"

3:Run the Query 
" Select imei,gamePointId, channelsId,series from Carbon_automation where  
channelsId < 4 ORDER BY gamePointId limit 5 "

4:Incorrect Result displays as follows:
++--+-+--+--+
|imei| gamePointId  | channelsId  |  series  |
++--+-+--+--+
| 1AA100050  | 29.0 | 1   | 2Series  |
| 1AA100014  | 151.0| 3   | 5Series  |
| 1AA100011  | 202.0| 1   | 0Series  |
| 1AA100018  | 441.0| 4   | 8Series  |
| 1AA100060  | 538.0| 4   | 8Series  |
++--+-+--+--+
5 rows selected (0.237 seconds)

5:CSV Attached: "100_hive_test.csv"

Expected Result: It should not display channel id 4 as per query.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Updated] (CARBONDATA-706) Mulitiple OR operators does not work properly in carbondata

2017-02-15 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-706?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-706:
-
Description: 
Incorrect result displays with multiple OR operators. 

Steps to Reproduces:
1:Create table using following Command
 " create table Carbon_automation (imei string,deviceInformationId int,MAC 
string,deviceColor string,device_backColor string,modelId string,marketName 
string,AMSize string,ROMSize string,CUPAudit string,CPIClocked string,series 
string,productionDate timestamp,bomCode string,internalModels string, 
deliveryTime string, channelsId string, channelsName string , deliveryAreaId 
string, deliveryCountry string, deliveryProvince string, deliveryCity 
string,deliveryDistrict string, deliveryStreet string, oxSingleNumber string, 
ActiveCheckTime string, ActiveAreaId string, ActiveCountry string, 
ActiveProvince string, Activecity string, ActiveDistrict string, ActiveStreet 
string, ActiveOperatorId string, Active_releaseId string, Active_EMUIVersion 
string, Active_operaSysVersion string, Active_BacVerNumber string, 
Active_BacFlashVer string, Active_webUIVersion string, Active_webUITypeCarrVer 
string,Active_webTypeDataVerNumber string, Active_operatorsVersion string, 
Active_phonePADPartitionedVersions string, Latest_YEAR int, Latest_MONTH int, 
Latest_DAY int, Latest_HOUR string, Latest_areaId string, Latest_country 
string, Latest_province string, Latest_city string, Latest_district string, 
Latest_street string, Latest_releaseId string, Latest_EMUIVersion string, 
Latest_operaSysVersion string, Latest_BacVerNumber string, Latest_BacFlashVer 
string, Latest_webUIVersion string, Latest_webUITypeCarrVer string, 
Latest_webTypeDataVerNumber string, Latest_operatorsVersion string, 
Latest_phonePADPartitionedVersions string, Latest_operatorId string, 
gamePointDescription string,gamePointId double,contractNumber double,imei_count 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
('DICTIONARY_INCLUDE'='deviceInformationId,Latest_YEAR,Latest_MONTH,Latest_DAY')"

2:Load Data with following command
 " LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/HiveData' INTO TABLE 
Carbon_automation 
OPTIONS('DELIMITER'=',','QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='imei,deviceInformationId,MAC,deviceColor,device_backColor,modelId,marketName,AMSize,ROMSize,CUPAudit,CPIClocked,series,productionDate,bomCode,internalModels,deliveryTime,channelsId,channelsName,deliveryAreaId,deliveryCountry,deliveryProvince,deliveryCity,deliveryDistrict,deliveryStreet,oxSingleNumber,contractNumber,ActiveCheckTime,ActiveAreaId,ActiveCountry,ActiveProvince,Activecity,ActiveDistrict,ActiveStreet,ActiveOperatorId,Active_releaseId,Active_EMUIVersion,Active_operaSysVersion,Active_BacVerNumber,Active_BacFlashVer,Active_webUIVersion,Active_webUITypeCarrVer,Active_webTypeDataVerNumber,Active_operatorsVersion,Active_phonePADPartitionedVersions,Latest_YEAR,Latest_MONTH,Latest_DAY,Latest_HOUR,Latest_areaId,Latest_country,Latest_province,Latest_city,Latest_district,Latest_street,Latest_releaseId,Latest_EMUIVersion,Latest_operaSysVersion,Latest_BacVerNumber,Latest_BacFlashVer,Latest_webUIVersion,Latest_webUITypeCarrVer,Latest_webTypeDataVerNumber,Latest_operatorsVersion,Latest_phonePADPartitionedVersions,Latest_operatorId,gamePointId,gamePointDescription,imei_count')"

3:Now run the Select Query:
" select imei,gamePointId, channelsId,series  from Carbon_automation where 
channelsId >=10 OR channelsId <=1 or series='7Series' "

4:Result Displays 

" 0: jdbc:hive2://localhost:1> select imei,gamePointId, channelsId,series  
from Carbon_automation where channelsId >=10 OR channelsId <=1 or 
series='7Series';
+-+--+-+--+--+
|imei | gamePointId  | channelsId  |  series  |
+-+--+-+--+--+
| 1AA1| 2738.562 | 4   | 7Series  |
| 1AA10   | 1714.635 | 4   | 7Series  |
| 1AA100  | 1271.0   | 6   | 5Series  |
| 1AA1000 | 692.0| 3   | 5Series  |
| 1AA1| 2175.0   | 1   | 7Series  |
| 1AA10   | 136.0| 6   | 9Series  |
| 1AA100  | 1600.0   | 6   | 7Series  |
| 1AA11   | 505.0| 7   | 0Series  |
| 1AA12   | 1341.0   | 3   | 0Series  |
| 1AA13   | 2239.0   | 3   | 5Series  |
| 1AA14   | 2970.0   | 2   | 4Series  |
| 1AA15   | 2593.0   | 1   | 1Series  |
| 1AA16   | 2572.0   | 2   | 6Series  |
| 1AA17   | 1991.0   | 3   | 9Series  |
| 1AA18   | 1442.0   | 1   | 8Series  |
| 1AA19   | 1841.0   | 3   | 0Series  |
| 1AA10001| 298.0| 5   | 2Series  |
| 1AA100010   | 79.0 | 6   | 3Series  |
| 1AA100011   | 202.0| 1   | 

[jira] [Created] (CARBONDATA-706) Mulitiple OR operators does not work properly in carbondata

2017-02-15 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-706:


 Summary: Mulitiple OR operators does not work properly in 
carbondata
 Key: CARBONDATA-706
 URL: https://issues.apache.org/jira/browse/CARBONDATA-706
 Project: CarbonData
  Issue Type: Bug
  Components: sql
Affects Versions: 1.1.0-incubating
 Environment: Spark 2.1
Reporter: SWATI RAO
Priority: Minor
 Attachments: 100_hive_test.csv

Multiple OR operators result does not match with hive.

Steps to Reproduces:
1:Create table using following Command
 " create table Carbon_automation (imei string,deviceInformationId int,MAC 
string,deviceColor string,device_backColor string,modelId string,marketName 
string,AMSize string,ROMSize string,CUPAudit string,CPIClocked string,series 
string,productionDate timestamp,bomCode string,internalModels string, 
deliveryTime string, channelsId string, channelsName string , deliveryAreaId 
string, deliveryCountry string, deliveryProvince string, deliveryCity 
string,deliveryDistrict string, deliveryStreet string, oxSingleNumber string, 
ActiveCheckTime string, ActiveAreaId string, ActiveCountry string, 
ActiveProvince string, Activecity string, ActiveDistrict string, ActiveStreet 
string, ActiveOperatorId string, Active_releaseId string, Active_EMUIVersion 
string, Active_operaSysVersion string, Active_BacVerNumber string, 
Active_BacFlashVer string, Active_webUIVersion string, Active_webUITypeCarrVer 
string,Active_webTypeDataVerNumber string, Active_operatorsVersion string, 
Active_phonePADPartitionedVersions string, Latest_YEAR int, Latest_MONTH int, 
Latest_DAY int, Latest_HOUR string, Latest_areaId string, Latest_country 
string, Latest_province string, Latest_city string, Latest_district string, 
Latest_street string, Latest_releaseId string, Latest_EMUIVersion string, 
Latest_operaSysVersion string, Latest_BacVerNumber string, Latest_BacFlashVer 
string, Latest_webUIVersion string, Latest_webUITypeCarrVer string, 
Latest_webTypeDataVerNumber string, Latest_operatorsVersion string, 
Latest_phonePADPartitionedVersions string, Latest_operatorId string, 
gamePointDescription string,gamePointId double,contractNumber double,imei_count 
int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
('DICTIONARY_INCLUDE'='deviceInformationId,Latest_YEAR,Latest_MONTH,Latest_DAY')"

2:Load Data with following command
 " LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/HiveData' INTO TABLE 
Carbon_automation 
OPTIONS('DELIMITER'=',','QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='imei,deviceInformationId,MAC,deviceColor,device_backColor,modelId,marketName,AMSize,ROMSize,CUPAudit,CPIClocked,series,productionDate,bomCode,internalModels,deliveryTime,channelsId,channelsName,deliveryAreaId,deliveryCountry,deliveryProvince,deliveryCity,deliveryDistrict,deliveryStreet,oxSingleNumber,contractNumber,ActiveCheckTime,ActiveAreaId,ActiveCountry,ActiveProvince,Activecity,ActiveDistrict,ActiveStreet,ActiveOperatorId,Active_releaseId,Active_EMUIVersion,Active_operaSysVersion,Active_BacVerNumber,Active_BacFlashVer,Active_webUIVersion,Active_webUITypeCarrVer,Active_webTypeDataVerNumber,Active_operatorsVersion,Active_phonePADPartitionedVersions,Latest_YEAR,Latest_MONTH,Latest_DAY,Latest_HOUR,Latest_areaId,Latest_country,Latest_province,Latest_city,Latest_district,Latest_street,Latest_releaseId,Latest_EMUIVersion,Latest_operaSysVersion,Latest_BacVerNumber,Latest_BacFlashVer,Latest_webUIVersion,Latest_webUITypeCarrVer,Latest_webTypeDataVerNumber,Latest_operatorsVersion,Latest_phonePADPartitionedVersions,Latest_operatorId,gamePointId,gamePointDescription,imei_count')"

3:Now run the Select Query:
" select imei,gamePointId, channelsId,series  from Carbon_automation where 
channelsId >=10 OR channelsId <=1 or series='7Series' "

4:Result Displays 

" 0: jdbc:hive2://localhost:1> select imei,gamePointId, channelsId,series  
from Carbon_automation where channelsId >=10 OR channelsId <=1 or 
series='7Series';
+-+--+-+--+--+
|imei | gamePointId  | channelsId  |  series  |
+-+--+-+--+--+
| 1AA1| 2738.562 | 4   | 7Series  |
| 1AA10   | 1714.635 | 4   | 7Series  |
| 1AA100  | 1271.0   | 6   | 5Series  |
| 1AA1000 | 692.0| 3   | 5Series  |
| 1AA1| 2175.0   | 1   | 7Series  |
| 1AA10   | 136.0| 6   | 9Series  |
| 1AA100  | 1600.0   | 6   | 7Series  |
| 1AA11   | 505.0| 7   | 0Series  |
| 1AA12   | 1341.0   | 3   | 0Series  |
| 1AA13   | 2239.0   | 3   | 5Series  |
| 1AA14   | 2970.0   | 2   | 4Series  |
| 1AA15   | 2593.0   | 1   | 1Series  |
| 1AA16   | 2572.0   | 2   | 6Series  |
| 1AA17   | 

[jira] [Created] (CARBONDATA-704) data mismatch between hive and carbondata after loading for bigint values

2017-02-13 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-704:


 Summary: data mismatch between hive and carbondata after loading 
for bigint values
 Key: CARBONDATA-704
 URL: https://issues.apache.org/jira/browse/CARBONDATA-704
 Project: CarbonData
  Issue Type: Bug
  Components: data-load
Affects Versions: 1.0.0-incubating
Reporter: SWATI RAO
 Attachments: Test_Data1 (4).csv

carbondata

0: jdbc:hive2://localhost:1> create table Test_Boundary (c1_int 
int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
'org.apache.carbondata.format' ;

0: jdbc:hive2://localhost:1>  LOAD DATA INPATH 
'hdfs://localhost:54310/Test_Data1.csv' INTO table Test_Boundary OPTIONS

('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='');

0: jdbc:hive2://localhost:1> select c2_Bigint from Test_Boundary;
+--+--+
|  c2_Bigint   |
+--+--+
| NULL |
| NULL |
| NULL |
| 9223372036854775807  |
| 9223372036854775807  |
| 9223372036854775807  |
| 9223372036854775807  |
| 9223372036854775807  |
| 9223372036854775807  |
| 9223372036854775807  |
| 9223372036854775807  |
| 9223372036854775807  |
| 9223372036854775807  |
| 9223372036854775807  |
| 9223372036854775807  |
| 9223372036854775807  |
+--+--+

but in hive

create table Test_Boundary_hive (c1_int int,c2_Bigint Bigint,c3_Decimal 
Decimal(38,30),c4_double double,c5_string string,c6_Timestamp 
Timestamp,c7_Datatype_Desc string)  ROW FORMAT DELIMITED FIELDS TERMINATED BY 
",";

LOAD DATA LOCAL INPATH 'Test_Data1.csv' into table Test_Boundary_hive;

select c2_Bigint from Test_Boundary_hive;

+---+--+
|   c2_Bigint   |
+---+--+
| 1234  |
| 2345  |
| 3456  |
| 4567  |
| 9223372036854775807   |
| -9223372036854775808  |
| -9223372036854775807  |
| -9223372036854775806  |
| -9223372036854775805  |
| 0 |
| 9223372036854775807   |
| 9223372036854775807   |
| 9223372036854775807   |
| NULL  |
| NULL  |
| NULL  |
+---+--+




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Closed] (CARBONDATA-419) Show tables in query for non-existing database do not show any exception

2017-02-09 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-419?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-419.

Resolution: Fixed

invalid issue

> Show tables in query for non-existing database do not show any exception 
> -
>
> Key: CARBONDATA-419
> URL: https://issues.apache.org/jira/browse/CARBONDATA-419
> Project: CarbonData
>  Issue Type: Bug
>  Components: data-query
>Reporter: SWATI RAO
>Assignee: Manohar Vanam
>
> When executing command 'show tables in' for an non-existing database, it do 
> not show any error,
> Example: show tables in test_tableMg;
> (where test_tableMg database do not exist)
> whereas when executing the same query on hive, it says : 
> FAILED: SemanticException [Error 10072]: Database does not exist: test_tableMg



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Closed] (CARBONDATA-678) Corr function is not working for double datatype.

2017-02-09 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-678?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-678.

Resolution: Fixed

this issue is invalid

> Corr function is not working for double datatype.
> -
>
> Key: CARBONDATA-678
> URL: https://issues.apache.org/jira/browse/CARBONDATA-678
> Project: CarbonData
>  Issue Type: Bug
>  Components: sql
>Affects Versions: 1.0.0-incubating
> Environment: Spark 1.6
>Reporter: SWATI RAO
>Priority: Trivial
> Fix For: 1.1.0-incubating
>
> Attachments: 2000_UniqData.csv, 3000_1_UniqData.csv, 
> 4000_UniqData.csv, 5000_UniqData.csv, 6000_UniqData.csv, 7000_UniqData.csv
>
>
> We have created a table uniqdata_h and then we upload data in this table then 
> we perform select query having data-type double then it displays null value 
> but when we perform select query with integer or bigint data-type it displays 
> a value.
> "Double" data-type is not working.
> CREATE QUERY :
> 0: jdbc:hive2://localhost:1> CREATE TABLE uniqdata_h (CUST_ID 
> int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ 
> timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 
> decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, 
> Double_COLUMN2 double,INTEGER_COLUMN1 int) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY ',';
> +-+--+
> | result  |
> +-+--+
> +-+--+
> No rows selected (4.249 seconds)
> LOAD 1 :
> 0: jdbc:hive2://localhost:1> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/uniqdata/2000_UniqData.csv' into table 
> uniqdata_h;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (1.176 seconds)
> LOAD 2 :
> 0: jdbc:hive2://localhost:1> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/uniqdata/4000_UniqData.csv' into table 
> uniqdata_h;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.485 seconds)
> LOAD 3 :
> 0: jdbc:hive2://localhost:1> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/uniqdata/5000_UniqData.csv' into table 
> uniqdata_h;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.365 seconds)
> LOAD 4 :
> 0: jdbc:hive2://localhost:1> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/uniqdata/6000_UniqData.csv' into table 
> uniqdata_h;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.594 seconds)
> LOAD 5 :
> 0: jdbc:hive2://localhost:1> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/uniqdata/7000_UniqData.csv' into table 
> uniqdata_h;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.753 seconds)
> LOAD 6 :
> 0: jdbc:hive2://localhost:1> load data local inpath 
> '/opt/Carbon/CarbonData/TestData/Data/uniqdata/3000_1_UniqData.csv' into 
> table uniqdata_h;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.417 seconds)
> SELECT QUERY : double data-type displays NULL value
> 0: jdbc:hive2://localhost:1> select corr(Double_COLUMN1,Double_COLUMN1)  
> as a from uniqdata_h ;
> +---+--+
> |   a   |
> +---+--+
> | NULL  |
> +---+--+
> whereas select query with integer & bigdata type displays some value.
> 0: jdbc:hive2://localhost:1> select corr(integer_COLUMN1,bigint_COLUMN1)  
> as a from uniqdata_h ;
> +-+--+
> |  a  |
> +-+--+
> | 1.0002  |
> +-+--+
> 1 row selected (0.54 seconds)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Closed] (CARBONDATA-693) Issue in select query for function Ceiling & Floor

2017-02-09 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-693.

Resolution: Fixed

this issue is invalid

> Issue in select query for function Ceiling & Floor
> --
>
> Key: CARBONDATA-693
> URL: https://issues.apache.org/jira/browse/CARBONDATA-693
> Project: CarbonData
>  Issue Type: Bug
>  Components: sql
>Affects Versions: 1.0.0-incubating
> Environment: Spark1.6
>Reporter: SWATI RAO
>Priority: Trivial
> Attachments: Test_Data1.csv
>
>
> When executed in Hive  it is working Fine, But when we execute same query in 
> carbondata it gives an error.
> PushUP_FILTER_Test_Boundary_TC059
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary2 (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ;
> +-+--+
> | result  |
> +-+--+
> +-+--+
> No rows selected (0.101 seconds)
> 0: jdbc:hive2://hadoop-master:1> show tables ;
> +-+--+--+
> |tableName| isTemporary  |
> +-+--+--+
> | test_boundary   | false|
> | test_boundary2  | false|
> +-+--+--+
> 2 rows selected (0.02 seconds)
> 0: jdbc:hive2://hadoop-master:1> load data local inpath 
> '/home/server/Desktop/Test_Data1.csv' into table test_boundary2 ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.201 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c3_Decimal from Test_Boundary 
> where floor(c3_Decimal)=0.00 or floor(c3_Decimal) IS NULL ;
> +-+--+
> | c3_Decimal  |
> +-+--+
> +-+--+
> No rows selected (0.315 seconds)
> 0: jdbc:hive2://hadoop-master:1> select count(*) from Test_Boundary ;
> +--+--+
> | _c0  |
> +--+--+
> | 0|
> +--+--+
> CarbonData 
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format' ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.535 seconds)
> 0: jdbc:hive2://hadoop-master:1> select count(*) from Test_Boundary ;
> +--+--+
> | _c0  |
> +--+--+
> | 0|
> +--+--+
> 1 row selected (1.883 seconds)
>  
> hdfs://192.168.2.145:54310hdfs://192.168.2.145:54310/HDFS_URL/BabuStore/Data/Test_Data1.csv
>  (state=,code=0)
> 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 
> 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table 
> Test_Boundary 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
>  ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (2.436 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c3_Decimal from Test_Boundary 
> where floor(c3_Decimal)=0.00 or floor(c3_Decimal) IS NULL ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 7.0 failed 4 times, most recent failure: Lost task 0.3 in 
> stage 7.0 (TID 16, hadoop-master): org.apache.spark.sql.AnalysisException: 
> Decimal scale (0) cannot be greater than precision (-28).;
>   at org.apache.spark.sql.types.PrecisionInfo.(DecimalType.scala:32)
>   at org.apache.spark.sql.types.DecimalType.(DecimalType.scala:68)
>   at 
> org.apache.spark.sql.types.DecimalType$.bounded(DecimalType.scala:155)
>   at org.apache.spark.sql.types.Decimal.floor(Decimal.scala:326)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificPredicate.eval(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.codegen.GeneratePredicate$$anonfun$create$2.apply(GeneratePredicate.scala:68)
>   at 
> org.apache.spark.sql.catalyst.expressions.codegen.GeneratePredicate$$anonfun$create$2.apply(GeneratePredicate.scala:68)
>   at 
> org.apache.spark.sql.execution.Filter$$anonfun$2$$anonfun$apply$2.apply(basicOperators.scala:74)
>   at 
> org.apache.spark.sql.execution.Filter$$anonfun$2$$anonfun$apply$2.apply(basicOperators.scala:72)
>   at scala.collection.Iterator$$anon$14.hasNext(Iterator.scala:390)
>   at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
>   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
>   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
>   at 
> scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
>   at 
> scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
>   at 
> 

[jira] [Updated] (CARBONDATA-693) Issue in select query for function Ceiling & Floor

2017-02-07 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-693:
-
Request participants:   (was: )
Priority: Minor  (was: Major)

> Issue in select query for function Ceiling & Floor
> --
>
> Key: CARBONDATA-693
> URL: https://issues.apache.org/jira/browse/CARBONDATA-693
> Project: CarbonData
>  Issue Type: Bug
>  Components: sql
>Affects Versions: 1.0.0-incubating
> Environment: Spark1.6
>Reporter: SWATI RAO
>Priority: Minor
> Attachments: Test_Data1.csv
>
>
> When executed in Hive  it is working Fine, But when we execute same query in 
> carbondata it gives an error.
> PushUP_FILTER_Test_Boundary_TC059
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary2 (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ;
> +-+--+
> | result  |
> +-+--+
> +-+--+
> No rows selected (0.101 seconds)
> 0: jdbc:hive2://hadoop-master:1> show tables ;
> +-+--+--+
> |tableName| isTemporary  |
> +-+--+--+
> | test_boundary   | false|
> | test_boundary2  | false|
> +-+--+--+
> 2 rows selected (0.02 seconds)
> 0: jdbc:hive2://hadoop-master:1> load data local inpath 
> '/home/server/Desktop/Test_Data1.csv' into table test_boundary2 ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.201 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c3_Decimal from Test_Boundary 
> where floor(c3_Decimal)=0.00 or floor(c3_Decimal) IS NULL ;
> +-+--+
> | c3_Decimal  |
> +-+--+
> +-+--+
> No rows selected (0.315 seconds)
> 0: jdbc:hive2://hadoop-master:1> select count(*) from Test_Boundary ;
> +--+--+
> | _c0  |
> +--+--+
> | 0|
> +--+--+
> CarbonData 
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format' ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.535 seconds)
> 0: jdbc:hive2://hadoop-master:1> select count(*) from Test_Boundary ;
> +--+--+
> | _c0  |
> +--+--+
> | 0|
> +--+--+
> 1 row selected (1.883 seconds)
>  
> hdfs://192.168.2.145:54310hdfs://192.168.2.145:54310/HDFS_URL/BabuStore/Data/Test_Data1.csv
>  (state=,code=0)
> 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 
> 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table 
> Test_Boundary 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
>  ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (2.436 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c3_Decimal from Test_Boundary 
> where floor(c3_Decimal)=0.00 or floor(c3_Decimal) IS NULL ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 7.0 failed 4 times, most recent failure: Lost task 0.3 in 
> stage 7.0 (TID 16, hadoop-master): org.apache.spark.sql.AnalysisException: 
> Decimal scale (0) cannot be greater than precision (-28).;
>   at org.apache.spark.sql.types.PrecisionInfo.(DecimalType.scala:32)
>   at org.apache.spark.sql.types.DecimalType.(DecimalType.scala:68)
>   at 
> org.apache.spark.sql.types.DecimalType$.bounded(DecimalType.scala:155)
>   at org.apache.spark.sql.types.Decimal.floor(Decimal.scala:326)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificPredicate.eval(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.codegen.GeneratePredicate$$anonfun$create$2.apply(GeneratePredicate.scala:68)
>   at 
> org.apache.spark.sql.catalyst.expressions.codegen.GeneratePredicate$$anonfun$create$2.apply(GeneratePredicate.scala:68)
>   at 
> org.apache.spark.sql.execution.Filter$$anonfun$2$$anonfun$apply$2.apply(basicOperators.scala:74)
>   at 
> org.apache.spark.sql.execution.Filter$$anonfun$2$$anonfun$apply$2.apply(basicOperators.scala:72)
>   at scala.collection.Iterator$$anon$14.hasNext(Iterator.scala:390)
>   at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
>   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
>   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
>   at 
> scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
>   at 
> scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
>   at 
> 

[jira] [Updated] (CARBONDATA-693) Issue in select query for function Ceiling & Floor

2017-02-07 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-693:
-
Request participants:   (was: )
Priority: Trivial  (was: Minor)

> Issue in select query for function Ceiling & Floor
> --
>
> Key: CARBONDATA-693
> URL: https://issues.apache.org/jira/browse/CARBONDATA-693
> Project: CarbonData
>  Issue Type: Bug
>  Components: sql
>Affects Versions: 1.0.0-incubating
> Environment: Spark1.6
>Reporter: SWATI RAO
>Priority: Trivial
> Attachments: Test_Data1.csv
>
>
> When executed in Hive  it is working Fine, But when we execute same query in 
> carbondata it gives an error.
> PushUP_FILTER_Test_Boundary_TC059
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary2 (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ;
> +-+--+
> | result  |
> +-+--+
> +-+--+
> No rows selected (0.101 seconds)
> 0: jdbc:hive2://hadoop-master:1> show tables ;
> +-+--+--+
> |tableName| isTemporary  |
> +-+--+--+
> | test_boundary   | false|
> | test_boundary2  | false|
> +-+--+--+
> 2 rows selected (0.02 seconds)
> 0: jdbc:hive2://hadoop-master:1> load data local inpath 
> '/home/server/Desktop/Test_Data1.csv' into table test_boundary2 ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.201 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c3_Decimal from Test_Boundary 
> where floor(c3_Decimal)=0.00 or floor(c3_Decimal) IS NULL ;
> +-+--+
> | c3_Decimal  |
> +-+--+
> +-+--+
> No rows selected (0.315 seconds)
> 0: jdbc:hive2://hadoop-master:1> select count(*) from Test_Boundary ;
> +--+--+
> | _c0  |
> +--+--+
> | 0|
> +--+--+
> CarbonData 
> 0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int 
> int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format' ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.535 seconds)
> 0: jdbc:hive2://hadoop-master:1> select count(*) from Test_Boundary ;
> +--+--+
> | _c0  |
> +--+--+
> | 0|
> +--+--+
> 1 row selected (1.883 seconds)
>  
> hdfs://192.168.2.145:54310hdfs://192.168.2.145:54310/HDFS_URL/BabuStore/Data/Test_Data1.csv
>  (state=,code=0)
> 0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 
> 'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table 
> Test_Boundary 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
>  ;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (2.436 seconds)
> 0: jdbc:hive2://hadoop-master:1> select c3_Decimal from Test_Boundary 
> where floor(c3_Decimal)=0.00 or floor(c3_Decimal) IS NULL ;
> Error: org.apache.spark.SparkException: Job aborted due to stage failure: 
> Task 0 in stage 7.0 failed 4 times, most recent failure: Lost task 0.3 in 
> stage 7.0 (TID 16, hadoop-master): org.apache.spark.sql.AnalysisException: 
> Decimal scale (0) cannot be greater than precision (-28).;
>   at org.apache.spark.sql.types.PrecisionInfo.(DecimalType.scala:32)
>   at org.apache.spark.sql.types.DecimalType.(DecimalType.scala:68)
>   at 
> org.apache.spark.sql.types.DecimalType$.bounded(DecimalType.scala:155)
>   at org.apache.spark.sql.types.Decimal.floor(Decimal.scala:326)
>   at 
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificPredicate.eval(Unknown
>  Source)
>   at 
> org.apache.spark.sql.catalyst.expressions.codegen.GeneratePredicate$$anonfun$create$2.apply(GeneratePredicate.scala:68)
>   at 
> org.apache.spark.sql.catalyst.expressions.codegen.GeneratePredicate$$anonfun$create$2.apply(GeneratePredicate.scala:68)
>   at 
> org.apache.spark.sql.execution.Filter$$anonfun$2$$anonfun$apply$2.apply(basicOperators.scala:74)
>   at 
> org.apache.spark.sql.execution.Filter$$anonfun$2$$anonfun$apply$2.apply(basicOperators.scala:72)
>   at scala.collection.Iterator$$anon$14.hasNext(Iterator.scala:390)
>   at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
>   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
>   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
>   at 
> scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
>   at 
> scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
>   at 
> 

[jira] [Updated] (CARBONDATA-665) Comparision Failure occurs when we execute the same query in hive and Carbondata

2017-02-07 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-665?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-665:
-
Request participants:   (was: )
Priority: Minor  (was: Major)

> Comparision Failure occurs when we execute the same query in hive and 
> Carbondata
> 
>
> Key: CARBONDATA-665
> URL: https://issues.apache.org/jira/browse/CARBONDATA-665
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.0.0-incubating
> Environment: Spark 1.6
>Reporter: SWATI RAO
>Priority: Minor
> Attachments: Test_Data1.csv
>
>
> Orderby is not working , so records are not coming in sequence as well there 
> is data difference and some values being stored as null
> Data itself is stored incorrectly and is different from Hive
> Spark version :1.6.2
> Create 1 query : create table Test_Boundary (c1_int int,c2_Bigint 
> Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format'
> Load 1 Query : LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO 
> table Test_Boundary 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
> Create 2 query : create table Test_Boundary1 (c1_int int,c2_Bigint 
> Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format'
> Load 2 query:  LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO 
> table Test_Boundary1 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
> Select Query : 
> select 
> c1_int,c2_Bigint,c3_Decimal,c4_double,c5_string,c6_Timestamp,c7_Datatype_Desc 
> from Test_Boundary where c2_bigint=c2_bigint



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Updated] (CARBONDATA-665) Comparision Failure occurs when we execute the same query in hive and Carbondata

2017-02-07 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-665?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-665:
-
Request participants:   (was: )

> Comparision Failure occurs when we execute the same query in hive and 
> Carbondata
> 
>
> Key: CARBONDATA-665
> URL: https://issues.apache.org/jira/browse/CARBONDATA-665
> Project: CarbonData
>  Issue Type: Bug
>Affects Versions: 1.0.0-incubating
> Environment: Spark 1.6
>Reporter: SWATI RAO
>Priority: Minor
> Attachments: Test_Data1.csv
>
>
> Orderby is not working , so records are not coming in sequence as well there 
> is data difference and some values being stored as null
> Data itself is stored incorrectly and is different from Hive
> Spark version :1.6.2
> Create 1 query : create table Test_Boundary (c1_int int,c2_Bigint 
> Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format'
> Load 1 Query : LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO 
> table Test_Boundary 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
> Create 2 query : create table Test_Boundary1 (c1_int int,c2_Bigint 
> Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format'
> Load 2 query:  LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO 
> table Test_Boundary1 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
> Select Query : 
> select 
> c1_int,c2_Bigint,c3_Decimal,c4_double,c5_string,c6_Timestamp,c7_Datatype_Desc 
> from Test_Boundary where c2_bigint=c2_bigint



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Updated] (CARBONDATA-668) Dataloads fails for this particular query

2017-02-07 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-668?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-668:
-
Description: 
Dataloads  fail for this query 


Create Query: create table smart_500_DE (MSISDN string,IMSI string,IMEI 
string,INTERFACEID int,GROUPID int,GGSN_ID double,SGSN_ID 
double,SESSION_INDICATOR int,BEGIN_TIME double,BEGIN_TIME_MSEL int,END_TIME 
double,END_TIME_MSEL int,PROT_CATEGORY int,PROT_TYPE int,L7_CARRIER_PROT 
int,SUB_PROT_TYPE int,SID double,PROBEID double,ENCRYPT_VERSION 
int,ROAMING_TYPE int,ROAM_DIRECTION int,MS_IP string,SERVER_IP string,MS_PORT 
int,SERVER_DECIMAL Decimal,APN string,SGSN_SIG_IP 
string,GGSN_SIG_IP_BigInt_NEGATIVE bigint,SGSN_USER_IP string,GGSN_USER_IP 
string,MCC string,MNC string,RAT int,LAC string,RAC string,SAC string,CI 
string,BROWSER_TIMESTAMP timestamp,TCP_CONN_STATES 
int,TCP_STATES_BIGINTPOSITIVE int,TCP_WIN_SIZE int,TCP_MSS int,TCP_CONN_TIMES 
int,TCP_CONN_2_FAILED_TIMES int,TCP_CONN_3_FAILED_TIMES int,HOST 
string,STREAMING_URL string,GET_STREAMING_FAILED_CODE int,GET_STREAMING_FLAG 
int,GET_NUM int,GET_SUCCEED_NUM int,GET_RETRANS_NUM int,GET_TIMEOUT_NUM 
int,INTBUFFER_FST_FLAG int,INTBUFFER_FULL_FLAG int,STALL_NUM 
int,VIDEO_FRAME_RATE int,VIDEO_CODEC_ID string,VIDEO_WIDTH int,VIDEO_HEIGHT 
int,AUDIO_CODEC_ID string,MEDIA_FILE_TYPE int,PLAY_STATE int,STREAMING_FLAG 
int,TCP_STATUS_INDICATOR int,DISCONNECTION_FLAG int,FAILURE_CODE int,FLAG 
int,TAC string,ECI string,TCP_SYN_TIME_MSEL int,TCP_FST_SYN_DIRECTION 
int,RAN_NE_USER_IP string,HOMEMCC string,HOMEMNC string,CHARGE_FLAG 
int,PREPAID_FLAG int,USER_AGENT string,MS_WIN_STAT_TOTAL_NUM 
int,MS_WIN_STAT_SMALL_NUM int,MS_ACK_TO_1STGET_DELAY 
int,SERVER_ACK_TO_1STDATA_DELAY int,STREAMING_TYPE int,SOURCE_VIDEO_QUALITY 
int,TETHERING_FLAG int,CARRIER_ID double,LAYER1ID int,LAYER2ID int,LAYER3ID 
int,LAYER4ID int,LAYER5ID int,LAYER6ID int,CHARGING_RULE_BASE_NAME string,SP 
string,EXTENDED_URL string,SV string,FIRST_SAI_CGI_ECGI 
string,EXTENDED_URL_OTHER string,SIGNALING_USE_FLAG int,DNS_RETRANS_NUM 
int,DNS_FAIL_CODE int,FIRST_RAT int,MS_INDICATOR string,LAST_SAI_CGI_ECGI 
string,LAST_RAT int,FIRST_LONGITUDE double,FIRST_LATITUDE double,FIRST_ALTITUDE 
int,FIRST_RASTERLONGITUDE double,FIRST_RASTERLATITUDE 
double,FIRST_RASTERALTITUDE int,FIRST_FREQUENCYSPOT int,FIRST_CLUTTER 
int,FIRST_USERBEHAVIOR int,FIRST_SPEED int,FIRST_CREDIBILITY int,LAST_LONGITUDE 
double,LAST_LATITUDE double,LAST_ALTITUDE int,LAST_RASTERLONGITUDE 
double,LAST_RASTERLATITUDE double,LAST_RASTERALTITUDE int,LAST_FREQUENCYSPOT 
int,LAST_CLUTTER int,LAST_USERBEHAVIOR int,LAST_SPEED int,LAST_CREDIBILITY 
int,IMEI_CIPHERTEXT string,APP_ID int,DOMAIN_NAME string,STREAMING_CACHE_IP 
string,STOP_LONGER_THAN_MIN_THRESHOLD int,STOP_LONGER_THAN_MAX_THRESHOLD 
int,PLAY_END_STAT int,STOP_START_TIME1 double,STOP_END_TIME1 
double,STOP_START_TIME2 double,STOP_END_TIME2 double,STOP_START_TIME3 
double,STOP_END_TIME3 double,STOP_START_TIME4 double,STOP_END_TIME4 
double,STOP_START_TIME5 double,STOP_END_TIME5 double,STOP_START_TIME6 
double,STOP_END_TIME6 double,STOP_START_TIME7 double,STOP_END_TIME7 
double,STOP_START_TIME8 double,STOP_END_TIME8 double,STOP_START_TIME9 
double,STOP_END_TIME9 double,STOP_START_TIME10 double,STOP_END_TIME10 
double,FAIL_CLASS double,RECORD_TYPE double,NODATA_COUNT 
double,VIDEO_NODATA_DURATION double,VIDEO_SMOOTH_DURATION 
double,VIDEO_SD_DURATION double,VIDEO_HD_DURATION double,VIDEO_UHD_DURATION 
double,VIDEO_FHD_DURATION double,FLUCTUATION double,START_DOWNLOAD_THROUGHPUT 
double,L7_UL_GOODPUT_FULL_MSS double,SESSIONKEY string,FIRST_UCELLID 
double,LAST_UCELLID double,UCELLID1 double,LONGITUDE1 double,LATITUDE1 
double,UCELLID2 double,LONGITUDE2 double,LATITUDE2 double,UCELLID3 
double,LONGITUDE3 double,LATITUDE3 double,UCELLID4 double,LONGITUDE4 
double,LATITUDE4 double,UCELLID5 double,LONGITUDE5 double,LATITUDE5 
double,UCELLID6 double,LONGITUDE6 double,LATITUDE6 double,UCELLID7 
double,LONGITUDE7 double,LATITUDE7 double,UCELLID8 double,LONGITUDE8 
double,LATITUDE8 double,UCELLID9 double,LONGITUDE9 double,LATITUDE9 
double,UCELLID10 double,LONGITUDE10 double,LATITUDE10 
double,INTBUFFER_FULL_DELAY double,STALL_DURATION double,STREAMING_DW_PACKETS 
double,STREAMING_DOWNLOAD_DELAY double,PLAY_DURATION double,STREAMING_QUALITY 
int,VIDEO_DATA_RATE double,AUDIO_DATA_RATE double,STREAMING_FILESIZE 
double,STREAMING_DURATIOIN double,TCP_SYN_TIME double,TCP_RTT_STEP1 
double,CHARGE_ID double,UL_REVERSE_TO_DL_DELAY double,DL_REVERSE_TO_UL_DELAY 
double,DATATRANS_DW_GOODPUT double,DATATRANS_DW_TOTAL_DURATION 
double,SUM_FRAGMENT_INTERVAL double,TCP_FIN_TIMES double,TCP_RESET_TIMES 
double,URL_CLASSIFICATION double,STREAMING_LQ_DURATIOIN double,MAX_DNS_DELAY 
double,MAX_DNS2SYN double,MAX_LATANCY_OF_LINK_SETUP double,MAX_SYNACK2FIRSTACK 
double,MAX_SYNACK2LASTACK double,MAX_ACK2GET_DELAY 
double,MAX_FRAG_INTERVAL_PREDELAY 

[jira] [Created] (CARBONDATA-693) Issue in select query for function Ceiling & Floor

2017-02-03 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-693:


 Summary: Issue in select query for function Ceiling & Floor
 Key: CARBONDATA-693
 URL: https://issues.apache.org/jira/browse/CARBONDATA-693
 Project: CarbonData
  Issue Type: Bug
  Components: sql
Affects Versions: 1.0.0-incubating
 Environment: Spark1.6
Reporter: SWATI RAO
 Attachments: Test_Data1.csv

When executed in Hive  it is working Fine, But when we execute same query in 
carbondata it gives an error.
PushUP_FILTER_Test_Boundary_TC059

0: jdbc:hive2://hadoop-master:1> create table Test_Boundary2 (c1_int 
int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ;
+-+--+
| result  |
+-+--+
+-+--+
No rows selected (0.101 seconds)



0: jdbc:hive2://hadoop-master:1> show tables ;
+-+--+--+
|tableName| isTemporary  |
+-+--+--+
| test_boundary   | false|
| test_boundary2  | false|
+-+--+--+
2 rows selected (0.02 seconds)


0: jdbc:hive2://hadoop-master:1> load data local inpath 
'/home/server/Desktop/Test_Data1.csv' into table test_boundary2 ;
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.201 seconds)


0: jdbc:hive2://hadoop-master:1> select c3_Decimal from Test_Boundary where 
floor(c3_Decimal)=0.00 or floor(c3_Decimal) IS NULL ;
+-+--+
| c3_Decimal  |
+-+--+
+-+--+
No rows selected (0.315 seconds)


0: jdbc:hive2://hadoop-master:1> select count(*) from Test_Boundary ;
+--+--+
| _c0  |
+--+--+
| 0|
+--+--+


CarbonData 

0: jdbc:hive2://hadoop-master:1> create table Test_Boundary (c1_int 
int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
'org.apache.carbondata.format' ;
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.535 seconds)
0: jdbc:hive2://hadoop-master:1> select count(*) from Test_Boundary ;
+--+--+
| _c0  |
+--+--+
| 0|
+--+--+
1 row selected (1.883 seconds)
 
hdfs://192.168.2.145:54310hdfs://192.168.2.145:54310/HDFS_URL/BabuStore/Data/Test_Data1.csv
 (state=,code=0)
0: jdbc:hive2://hadoop-master:1> LOAD DATA INPATH 
'hdfs://192.168.2.145:54310/BabuStore/Data/Test_Data1.csv' INTO table 
Test_Boundary 
OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
 ;
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (2.436 seconds)


0: jdbc:hive2://hadoop-master:1> select c3_Decimal from Test_Boundary where 
floor(c3_Decimal)=0.00 or floor(c3_Decimal) IS NULL ;
Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 
0 in stage 7.0 failed 4 times, most recent failure: Lost task 0.3 in stage 7.0 
(TID 16, hadoop-master): org.apache.spark.sql.AnalysisException: Decimal scale 
(0) cannot be greater than precision (-28).;
at org.apache.spark.sql.types.PrecisionInfo.(DecimalType.scala:32)
at org.apache.spark.sql.types.DecimalType.(DecimalType.scala:68)
at 
org.apache.spark.sql.types.DecimalType$.bounded(DecimalType.scala:155)
at org.apache.spark.sql.types.Decimal.floor(Decimal.scala:326)
at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificPredicate.eval(Unknown
 Source)
at 
org.apache.spark.sql.catalyst.expressions.codegen.GeneratePredicate$$anonfun$create$2.apply(GeneratePredicate.scala:68)
at 
org.apache.spark.sql.catalyst.expressions.codegen.GeneratePredicate$$anonfun$create$2.apply(GeneratePredicate.scala:68)
at 
org.apache.spark.sql.execution.Filter$$anonfun$2$$anonfun$apply$2.apply(basicOperators.scala:74)
at 
org.apache.spark.sql.execution.Filter$$anonfun$2$$anonfun$apply$2.apply(basicOperators.scala:72)
at scala.collection.Iterator$$anon$14.hasNext(Iterator.scala:390)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
at scala.collection.Iterator$class.foreach(Iterator.scala:727)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
at 
scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
at 
scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
at 
scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
at scala.collection.AbstractIterator.to(Iterator.scala:1157)
at 
scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
at 

[jira] [Updated] (CARBONDATA-668) Dataloads fail when no. of column in load query is greater than the no. of column in create table

2017-01-20 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-668?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-668:
-
   Priority: Minor  (was: Major)
Description: 
Dataloads  fail when no. of column in load query is greater than the no. of 
column in create table


Create Query: create table smart_500_DE (MSISDN string,IMSI string,IMEI 
string,INTERFACEID int,GROUPID int,GGSN_ID double,SGSN_ID 
double,SESSION_INDICATOR int,BEGIN_TIME double,BEGIN_TIME_MSEL int,END_TIME 
double,END_TIME_MSEL int,PROT_CATEGORY int,PROT_TYPE int,L7_CARRIER_PROT 
int,SUB_PROT_TYPE int,SID double,PROBEID double,ENCRYPT_VERSION 
int,ROAMING_TYPE int,ROAM_DIRECTION int,MS_IP string,SERVER_IP string,MS_PORT 
int,SERVER_DECIMAL Decimal,APN string,SGSN_SIG_IP 
string,GGSN_SIG_IP_BigInt_NEGATIVE bigint,SGSN_USER_IP string,GGSN_USER_IP 
string,MCC string,MNC string,RAT int,LAC string,RAC string,SAC string,CI 
string,BROWSER_TIMESTAMP timestamp,TCP_CONN_STATES 
int,TCP_STATES_BIGINTPOSITIVE int,TCP_WIN_SIZE int,TCP_MSS int,TCP_CONN_TIMES 
int,TCP_CONN_2_FAILED_TIMES int,TCP_CONN_3_FAILED_TIMES int,HOST 
string,STREAMING_URL string,GET_STREAMING_FAILED_CODE int,GET_STREAMING_FLAG 
int,GET_NUM int,GET_SUCCEED_NUM int,GET_RETRANS_NUM int,GET_TIMEOUT_NUM 
int,INTBUFFER_FST_FLAG int,INTBUFFER_FULL_FLAG int,STALL_NUM 
int,VIDEO_FRAME_RATE int,VIDEO_CODEC_ID string,VIDEO_WIDTH int,VIDEO_HEIGHT 
int,AUDIO_CODEC_ID string,MEDIA_FILE_TYPE int,PLAY_STATE int,STREAMING_FLAG 
int,TCP_STATUS_INDICATOR int,DISCONNECTION_FLAG int,FAILURE_CODE int,FLAG 
int,TAC string,ECI string,TCP_SYN_TIME_MSEL int,TCP_FST_SYN_DIRECTION 
int,RAN_NE_USER_IP string,HOMEMCC string,HOMEMNC string,CHARGE_FLAG 
int,PREPAID_FLAG int,USER_AGENT string,MS_WIN_STAT_TOTAL_NUM 
int,MS_WIN_STAT_SMALL_NUM int,MS_ACK_TO_1STGET_DELAY 
int,SERVER_ACK_TO_1STDATA_DELAY int,STREAMING_TYPE int,SOURCE_VIDEO_QUALITY 
int,TETHERING_FLAG int,CARRIER_ID double,LAYER1ID int,LAYER2ID int,LAYER3ID 
int,LAYER4ID int,LAYER5ID int,LAYER6ID int,CHARGING_RULE_BASE_NAME string,SP 
string,EXTENDED_URL string,SV string,FIRST_SAI_CGI_ECGI 
string,EXTENDED_URL_OTHER string,SIGNALING_USE_FLAG int,DNS_RETRANS_NUM 
int,DNS_FAIL_CODE int,FIRST_RAT int,MS_INDICATOR string,LAST_SAI_CGI_ECGI 
string,LAST_RAT int,FIRST_LONGITUDE double,FIRST_LATITUDE double,FIRST_ALTITUDE 
int,FIRST_RASTERLONGITUDE double,FIRST_RASTERLATITUDE 
double,FIRST_RASTERALTITUDE int,FIRST_FREQUENCYSPOT int,FIRST_CLUTTER 
int,FIRST_USERBEHAVIOR int,FIRST_SPEED int,FIRST_CREDIBILITY int,LAST_LONGITUDE 
double,LAST_LATITUDE double,LAST_ALTITUDE int,LAST_RASTERLONGITUDE 
double,LAST_RASTERLATITUDE double,LAST_RASTERALTITUDE int,LAST_FREQUENCYSPOT 
int,LAST_CLUTTER int,LAST_USERBEHAVIOR int,LAST_SPEED int,LAST_CREDIBILITY 
int,IMEI_CIPHERTEXT string,APP_ID int,DOMAIN_NAME string,STREAMING_CACHE_IP 
string,STOP_LONGER_THAN_MIN_THRESHOLD int,STOP_LONGER_THAN_MAX_THRESHOLD 
int,PLAY_END_STAT int,STOP_START_TIME1 double,STOP_END_TIME1 
double,STOP_START_TIME2 double,STOP_END_TIME2 double,STOP_START_TIME3 
double,STOP_END_TIME3 double,STOP_START_TIME4 double,STOP_END_TIME4 
double,STOP_START_TIME5 double,STOP_END_TIME5 double,STOP_START_TIME6 
double,STOP_END_TIME6 double,STOP_START_TIME7 double,STOP_END_TIME7 
double,STOP_START_TIME8 double,STOP_END_TIME8 double,STOP_START_TIME9 
double,STOP_END_TIME9 double,STOP_START_TIME10 double,STOP_END_TIME10 
double,FAIL_CLASS double,RECORD_TYPE double,NODATA_COUNT 
double,VIDEO_NODATA_DURATION double,VIDEO_SMOOTH_DURATION 
double,VIDEO_SD_DURATION double,VIDEO_HD_DURATION double,VIDEO_UHD_DURATION 
double,VIDEO_FHD_DURATION double,FLUCTUATION double,START_DOWNLOAD_THROUGHPUT 
double,L7_UL_GOODPUT_FULL_MSS double,SESSIONKEY string,FIRST_UCELLID 
double,LAST_UCELLID double,UCELLID1 double,LONGITUDE1 double,LATITUDE1 
double,UCELLID2 double,LONGITUDE2 double,LATITUDE2 double,UCELLID3 
double,LONGITUDE3 double,LATITUDE3 double,UCELLID4 double,LONGITUDE4 
double,LATITUDE4 double,UCELLID5 double,LONGITUDE5 double,LATITUDE5 
double,UCELLID6 double,LONGITUDE6 double,LATITUDE6 double,UCELLID7 
double,LONGITUDE7 double,LATITUDE7 double,UCELLID8 double,LONGITUDE8 
double,LATITUDE8 double,UCELLID9 double,LONGITUDE9 double,LATITUDE9 
double,UCELLID10 double,LONGITUDE10 double,LATITUDE10 
double,INTBUFFER_FULL_DELAY double,STALL_DURATION double,STREAMING_DW_PACKETS 
double,STREAMING_DOWNLOAD_DELAY double,PLAY_DURATION double,STREAMING_QUALITY 
int,VIDEO_DATA_RATE double,AUDIO_DATA_RATE double,STREAMING_FILESIZE 
double,STREAMING_DURATIOIN double,TCP_SYN_TIME double,TCP_RTT_STEP1 
double,CHARGE_ID double,UL_REVERSE_TO_DL_DELAY double,DL_REVERSE_TO_UL_DELAY 
double,DATATRANS_DW_GOODPUT double,DATATRANS_DW_TOTAL_DURATION 
double,SUM_FRAGMENT_INTERVAL double,TCP_FIN_TIMES double,TCP_RESET_TIMES 
double,URL_CLASSIFICATION double,STREAMING_LQ_DURATIOIN double,MAX_DNS_DELAY 
double,MAX_DNS2SYN double,MAX_LATANCY_OF_LINK_SETUP 

[jira] [Updated] (CARBONDATA-668) Data loading

2017-01-20 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-668?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-668:
-
Summary: Data loading   (was: Data loading failed on the latest build)

> Data loading 
> -
>
> Key: CARBONDATA-668
> URL: https://issues.apache.org/jira/browse/CARBONDATA-668
> Project: CarbonData
>  Issue Type: Bug
>  Components: data-load
> Environment: Spark-1.6
>Reporter: SWATI RAO
> Attachments: seq_500Records.csv
>
>
> Create Query: create table smart_500_DE (MSISDN string,IMSI string,IMEI 
> string,INTERFACEID int,GROUPID int,GGSN_ID double,SGSN_ID 
> double,SESSION_INDICATOR int,BEGIN_TIME double,BEGIN_TIME_MSEL int,END_TIME 
> double,END_TIME_MSEL int,PROT_CATEGORY int,PROT_TYPE int,L7_CARRIER_PROT 
> int,SUB_PROT_TYPE int,SID double,PROBEID double,ENCRYPT_VERSION 
> int,ROAMING_TYPE int,ROAM_DIRECTION int,MS_IP string,SERVER_IP string,MS_PORT 
> int,SERVER_DECIMAL Decimal,APN string,SGSN_SIG_IP 
> string,GGSN_SIG_IP_BigInt_NEGATIVE bigint,SGSN_USER_IP string,GGSN_USER_IP 
> string,MCC string,MNC string,RAT int,LAC string,RAC string,SAC string,CI 
> string,BROWSER_TIMESTAMP timestamp,TCP_CONN_STATES 
> int,TCP_STATES_BIGINTPOSITIVE int,TCP_WIN_SIZE int,TCP_MSS int,TCP_CONN_TIMES 
> int,TCP_CONN_2_FAILED_TIMES int,TCP_CONN_3_FAILED_TIMES int,HOST 
> string,STREAMING_URL string,GET_STREAMING_FAILED_CODE int,GET_STREAMING_FLAG 
> int,GET_NUM int,GET_SUCCEED_NUM int,GET_RETRANS_NUM int,GET_TIMEOUT_NUM 
> int,INTBUFFER_FST_FLAG int,INTBUFFER_FULL_FLAG int,STALL_NUM 
> int,VIDEO_FRAME_RATE int,VIDEO_CODEC_ID string,VIDEO_WIDTH int,VIDEO_HEIGHT 
> int,AUDIO_CODEC_ID string,MEDIA_FILE_TYPE int,PLAY_STATE int,STREAMING_FLAG 
> int,TCP_STATUS_INDICATOR int,DISCONNECTION_FLAG int,FAILURE_CODE int,FLAG 
> int,TAC string,ECI string,TCP_SYN_TIME_MSEL int,TCP_FST_SYN_DIRECTION 
> int,RAN_NE_USER_IP string,HOMEMCC string,HOMEMNC string,CHARGE_FLAG 
> int,PREPAID_FLAG int,USER_AGENT string,MS_WIN_STAT_TOTAL_NUM 
> int,MS_WIN_STAT_SMALL_NUM int,MS_ACK_TO_1STGET_DELAY 
> int,SERVER_ACK_TO_1STDATA_DELAY int,STREAMING_TYPE int,SOURCE_VIDEO_QUALITY 
> int,TETHERING_FLAG int,CARRIER_ID double,LAYER1ID int,LAYER2ID int,LAYER3ID 
> int,LAYER4ID int,LAYER5ID int,LAYER6ID int,CHARGING_RULE_BASE_NAME string,SP 
> string,EXTENDED_URL string,SV string,FIRST_SAI_CGI_ECGI 
> string,EXTENDED_URL_OTHER string,SIGNALING_USE_FLAG int,DNS_RETRANS_NUM 
> int,DNS_FAIL_CODE int,FIRST_RAT int,MS_INDICATOR string,LAST_SAI_CGI_ECGI 
> string,LAST_RAT int,FIRST_LONGITUDE double,FIRST_LATITUDE 
> double,FIRST_ALTITUDE int,FIRST_RASTERLONGITUDE double,FIRST_RASTERLATITUDE 
> double,FIRST_RASTERALTITUDE int,FIRST_FREQUENCYSPOT int,FIRST_CLUTTER 
> int,FIRST_USERBEHAVIOR int,FIRST_SPEED int,FIRST_CREDIBILITY 
> int,LAST_LONGITUDE double,LAST_LATITUDE double,LAST_ALTITUDE 
> int,LAST_RASTERLONGITUDE double,LAST_RASTERLATITUDE 
> double,LAST_RASTERALTITUDE int,LAST_FREQUENCYSPOT int,LAST_CLUTTER 
> int,LAST_USERBEHAVIOR int,LAST_SPEED int,LAST_CREDIBILITY int,IMEI_CIPHERTEXT 
> string,APP_ID int,DOMAIN_NAME string,STREAMING_CACHE_IP 
> string,STOP_LONGER_THAN_MIN_THRESHOLD int,STOP_LONGER_THAN_MAX_THRESHOLD 
> int,PLAY_END_STAT int,STOP_START_TIME1 double,STOP_END_TIME1 
> double,STOP_START_TIME2 double,STOP_END_TIME2 double,STOP_START_TIME3 
> double,STOP_END_TIME3 double,STOP_START_TIME4 double,STOP_END_TIME4 
> double,STOP_START_TIME5 double,STOP_END_TIME5 double,STOP_START_TIME6 
> double,STOP_END_TIME6 double,STOP_START_TIME7 double,STOP_END_TIME7 
> double,STOP_START_TIME8 double,STOP_END_TIME8 double,STOP_START_TIME9 
> double,STOP_END_TIME9 double,STOP_START_TIME10 double,STOP_END_TIME10 
> double,FAIL_CLASS double,RECORD_TYPE double,NODATA_COUNT 
> double,VIDEO_NODATA_DURATION double,VIDEO_SMOOTH_DURATION 
> double,VIDEO_SD_DURATION double,VIDEO_HD_DURATION double,VIDEO_UHD_DURATION 
> double,VIDEO_FHD_DURATION double,FLUCTUATION double,START_DOWNLOAD_THROUGHPUT 
> double,L7_UL_GOODPUT_FULL_MSS double,SESSIONKEY string,FIRST_UCELLID 
> double,LAST_UCELLID double,UCELLID1 double,LONGITUDE1 double,LATITUDE1 
> double,UCELLID2 double,LONGITUDE2 double,LATITUDE2 double,UCELLID3 
> double,LONGITUDE3 double,LATITUDE3 double,UCELLID4 double,LONGITUDE4 
> double,LATITUDE4 double,UCELLID5 double,LONGITUDE5 double,LATITUDE5 
> double,UCELLID6 double,LONGITUDE6 double,LATITUDE6 double,UCELLID7 
> double,LONGITUDE7 double,LATITUDE7 double,UCELLID8 double,LONGITUDE8 
> double,LATITUDE8 double,UCELLID9 double,LONGITUDE9 double,LATITUDE9 
> double,UCELLID10 double,LONGITUDE10 double,LATITUDE10 
> double,INTBUFFER_FULL_DELAY double,STALL_DURATION double,STREAMING_DW_PACKETS 
> double,STREAMING_DOWNLOAD_DELAY double,PLAY_DURATION double,STREAMING_QUALITY 
> int,VIDEO_DATA_RATE double,AUDIO_DATA_RATE double,STREAMING_FILESIZE 
> double,STREAMING_DURATIOIN 

[jira] [Updated] (CARBONDATA-668) Data loading failed on the latest build

2017-01-20 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-668?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-668:
-
 Attachment: seq_500Records.csv
Environment: Spark-1.6
Description: 
Create Query: create table smart_500_DE (MSISDN string,IMSI string,IMEI 
string,INTERFACEID int,GROUPID int,GGSN_ID double,SGSN_ID 
double,SESSION_INDICATOR int,BEGIN_TIME double,BEGIN_TIME_MSEL int,END_TIME 
double,END_TIME_MSEL int,PROT_CATEGORY int,PROT_TYPE int,L7_CARRIER_PROT 
int,SUB_PROT_TYPE int,SID double,PROBEID double,ENCRYPT_VERSION 
int,ROAMING_TYPE int,ROAM_DIRECTION int,MS_IP string,SERVER_IP string,MS_PORT 
int,SERVER_DECIMAL Decimal,APN string,SGSN_SIG_IP 
string,GGSN_SIG_IP_BigInt_NEGATIVE bigint,SGSN_USER_IP string,GGSN_USER_IP 
string,MCC string,MNC string,RAT int,LAC string,RAC string,SAC string,CI 
string,BROWSER_TIMESTAMP timestamp,TCP_CONN_STATES 
int,TCP_STATES_BIGINTPOSITIVE int,TCP_WIN_SIZE int,TCP_MSS int,TCP_CONN_TIMES 
int,TCP_CONN_2_FAILED_TIMES int,TCP_CONN_3_FAILED_TIMES int,HOST 
string,STREAMING_URL string,GET_STREAMING_FAILED_CODE int,GET_STREAMING_FLAG 
int,GET_NUM int,GET_SUCCEED_NUM int,GET_RETRANS_NUM int,GET_TIMEOUT_NUM 
int,INTBUFFER_FST_FLAG int,INTBUFFER_FULL_FLAG int,STALL_NUM 
int,VIDEO_FRAME_RATE int,VIDEO_CODEC_ID string,VIDEO_WIDTH int,VIDEO_HEIGHT 
int,AUDIO_CODEC_ID string,MEDIA_FILE_TYPE int,PLAY_STATE int,STREAMING_FLAG 
int,TCP_STATUS_INDICATOR int,DISCONNECTION_FLAG int,FAILURE_CODE int,FLAG 
int,TAC string,ECI string,TCP_SYN_TIME_MSEL int,TCP_FST_SYN_DIRECTION 
int,RAN_NE_USER_IP string,HOMEMCC string,HOMEMNC string,CHARGE_FLAG 
int,PREPAID_FLAG int,USER_AGENT string,MS_WIN_STAT_TOTAL_NUM 
int,MS_WIN_STAT_SMALL_NUM int,MS_ACK_TO_1STGET_DELAY 
int,SERVER_ACK_TO_1STDATA_DELAY int,STREAMING_TYPE int,SOURCE_VIDEO_QUALITY 
int,TETHERING_FLAG int,CARRIER_ID double,LAYER1ID int,LAYER2ID int,LAYER3ID 
int,LAYER4ID int,LAYER5ID int,LAYER6ID int,CHARGING_RULE_BASE_NAME string,SP 
string,EXTENDED_URL string,SV string,FIRST_SAI_CGI_ECGI 
string,EXTENDED_URL_OTHER string,SIGNALING_USE_FLAG int,DNS_RETRANS_NUM 
int,DNS_FAIL_CODE int,FIRST_RAT int,MS_INDICATOR string,LAST_SAI_CGI_ECGI 
string,LAST_RAT int,FIRST_LONGITUDE double,FIRST_LATITUDE double,FIRST_ALTITUDE 
int,FIRST_RASTERLONGITUDE double,FIRST_RASTERLATITUDE 
double,FIRST_RASTERALTITUDE int,FIRST_FREQUENCYSPOT int,FIRST_CLUTTER 
int,FIRST_USERBEHAVIOR int,FIRST_SPEED int,FIRST_CREDIBILITY int,LAST_LONGITUDE 
double,LAST_LATITUDE double,LAST_ALTITUDE int,LAST_RASTERLONGITUDE 
double,LAST_RASTERLATITUDE double,LAST_RASTERALTITUDE int,LAST_FREQUENCYSPOT 
int,LAST_CLUTTER int,LAST_USERBEHAVIOR int,LAST_SPEED int,LAST_CREDIBILITY 
int,IMEI_CIPHERTEXT string,APP_ID int,DOMAIN_NAME string,STREAMING_CACHE_IP 
string,STOP_LONGER_THAN_MIN_THRESHOLD int,STOP_LONGER_THAN_MAX_THRESHOLD 
int,PLAY_END_STAT int,STOP_START_TIME1 double,STOP_END_TIME1 
double,STOP_START_TIME2 double,STOP_END_TIME2 double,STOP_START_TIME3 
double,STOP_END_TIME3 double,STOP_START_TIME4 double,STOP_END_TIME4 
double,STOP_START_TIME5 double,STOP_END_TIME5 double,STOP_START_TIME6 
double,STOP_END_TIME6 double,STOP_START_TIME7 double,STOP_END_TIME7 
double,STOP_START_TIME8 double,STOP_END_TIME8 double,STOP_START_TIME9 
double,STOP_END_TIME9 double,STOP_START_TIME10 double,STOP_END_TIME10 
double,FAIL_CLASS double,RECORD_TYPE double,NODATA_COUNT 
double,VIDEO_NODATA_DURATION double,VIDEO_SMOOTH_DURATION 
double,VIDEO_SD_DURATION double,VIDEO_HD_DURATION double,VIDEO_UHD_DURATION 
double,VIDEO_FHD_DURATION double,FLUCTUATION double,START_DOWNLOAD_THROUGHPUT 
double,L7_UL_GOODPUT_FULL_MSS double,SESSIONKEY string,FIRST_UCELLID 
double,LAST_UCELLID double,UCELLID1 double,LONGITUDE1 double,LATITUDE1 
double,UCELLID2 double,LONGITUDE2 double,LATITUDE2 double,UCELLID3 
double,LONGITUDE3 double,LATITUDE3 double,UCELLID4 double,LONGITUDE4 
double,LATITUDE4 double,UCELLID5 double,LONGITUDE5 double,LATITUDE5 
double,UCELLID6 double,LONGITUDE6 double,LATITUDE6 double,UCELLID7 
double,LONGITUDE7 double,LATITUDE7 double,UCELLID8 double,LONGITUDE8 
double,LATITUDE8 double,UCELLID9 double,LONGITUDE9 double,LATITUDE9 
double,UCELLID10 double,LONGITUDE10 double,LATITUDE10 
double,INTBUFFER_FULL_DELAY double,STALL_DURATION double,STREAMING_DW_PACKETS 
double,STREAMING_DOWNLOAD_DELAY double,PLAY_DURATION double,STREAMING_QUALITY 
int,VIDEO_DATA_RATE double,AUDIO_DATA_RATE double,STREAMING_FILESIZE 
double,STREAMING_DURATIOIN double,TCP_SYN_TIME double,TCP_RTT_STEP1 
double,CHARGE_ID double,UL_REVERSE_TO_DL_DELAY double,DL_REVERSE_TO_UL_DELAY 
double,DATATRANS_DW_GOODPUT double,DATATRANS_DW_TOTAL_DURATION 
double,SUM_FRAGMENT_INTERVAL double,TCP_FIN_TIMES double,TCP_RESET_TIMES 
double,URL_CLASSIFICATION double,STREAMING_LQ_DURATIOIN double,MAX_DNS_DELAY 
double,MAX_DNS2SYN double,MAX_LATANCY_OF_LINK_SETUP double,MAX_SYNACK2FIRSTACK 
double,MAX_SYNACK2LASTACK double,MAX_ACK2GET_DELAY 

[jira] [Created] (CARBONDATA-668) Data loading failed on the latest build

2017-01-20 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-668:


 Summary: Data loading failed on the latest build
 Key: CARBONDATA-668
 URL: https://issues.apache.org/jira/browse/CARBONDATA-668
 Project: CarbonData
  Issue Type: Bug
Reporter: SWATI RAO


Create Query: create table smart_500_DE (MSISDN string,IMSI string,IMEI 
string,INTERFACEID int,GROUPID int,GGSN_ID double,SGSN_ID 
double,SESSION_INDICATOR int,BEGIN_TIME double,BEGIN_TIME_MSEL int,END_TIME 
double,END_TIME_MSEL int,PROT_CATEGORY int,PROT_TYPE int,L7_CARRIER_PROT 
int,SUB_PROT_TYPE int,SID double,PROBEID double,ENCRYPT_VERSION 
int,ROAMING_TYPE int,ROAM_DIRECTION int,MS_IP string,SERVER_IP string,MS_PORT 
int,SERVER_DECIMAL Decimal,APN string,SGSN_SIG_IP 
string,GGSN_SIG_IP_BigInt_NEGATIVE bigint,SGSN_USER_IP string,GGSN_USER_IP 
string,MCC string,MNC string,RAT int,LAC string,RAC string,SAC string,CI 
string,BROWSER_TIMESTAMP timestamp,TCP_CONN_STATES 
int,TCP_STATES_BIGINTPOSITIVE int,TCP_WIN_SIZE int,TCP_MSS int,TCP_CONN_TIMES 
int,TCP_CONN_2_FAILED_TIMES int,TCP_CONN_3_FAILED_TIMES int,HOST 
string,STREAMING_URL string,GET_STREAMING_FAILED_CODE int,GET_STREAMING_FLAG 
int,GET_NUM int,GET_SUCCEED_NUM int,GET_RETRANS_NUM int,GET_TIMEOUT_NUM 
int,INTBUFFER_FST_FLAG int,INTBUFFER_FULL_FLAG int,STALL_NUM 
int,VIDEO_FRAME_RATE int,VIDEO_CODEC_ID string,VIDEO_WIDTH int,VIDEO_HEIGHT 
int,AUDIO_CODEC_ID string,MEDIA_FILE_TYPE int,PLAY_STATE int,STREAMING_FLAG 
int,TCP_STATUS_INDICATOR int,DISCONNECTION_FLAG int,FAILURE_CODE int,FLAG 
int,TAC string,ECI string,TCP_SYN_TIME_MSEL int,TCP_FST_SYN_DIRECTION 
int,RAN_NE_USER_IP string,HOMEMCC string,HOMEMNC string,CHARGE_FLAG 
int,PREPAID_FLAG int,USER_AGENT string,MS_WIN_STAT_TOTAL_NUM 
int,MS_WIN_STAT_SMALL_NUM int,MS_ACK_TO_1STGET_DELAY 
int,SERVER_ACK_TO_1STDATA_DELAY int,STREAMING_TYPE int,SOURCE_VIDEO_QUALITY 
int,TETHERING_FLAG int,CARRIER_ID double,LAYER1ID int,LAYER2ID int,LAYER3ID 
int,LAYER4ID int,LAYER5ID int,LAYER6ID int,CHARGING_RULE_BASE_NAME string,SP 
string,EXTENDED_URL string,SV string,FIRST_SAI_CGI_ECGI 
string,EXTENDED_URL_OTHER string,SIGNALING_USE_FLAG int,DNS_RETRANS_NUM 
int,DNS_FAIL_CODE int,FIRST_RAT int,MS_INDICATOR string,LAST_SAI_CGI_ECGI 
string,LAST_RAT int,FIRST_LONGITUDE double,FIRST_LATITUDE double,FIRST_ALTITUDE 
int,FIRST_RASTERLONGITUDE double,FIRST_RASTERLATITUDE 
double,FIRST_RASTERALTITUDE int,FIRST_FREQUENCYSPOT int,FIRST_CLUTTER 
int,FIRST_USERBEHAVIOR int,FIRST_SPEED int,FIRST_CREDIBILITY int,LAST_LONGITUDE 
double,LAST_LATITUDE double,LAST_ALTITUDE int,LAST_RASTERLONGITUDE 
double,LAST_RASTERLATITUDE double,LAST_RASTERALTITUDE int,LAST_FREQUENCYSPOT 
int,LAST_CLUTTER int,LAST_USERBEHAVIOR int,LAST_SPEED int,LAST_CREDIBILITY 
int,IMEI_CIPHERTEXT string,APP_ID int,DOMAIN_NAME string,STREAMING_CACHE_IP 
string,STOP_LONGER_THAN_MIN_THRESHOLD int,STOP_LONGER_THAN_MAX_THRESHOLD 
int,PLAY_END_STAT int,STOP_START_TIME1 double,STOP_END_TIME1 
double,STOP_START_TIME2 double,STOP_END_TIME2 double,STOP_START_TIME3 
double,STOP_END_TIME3 double,STOP_START_TIME4 double,STOP_END_TIME4 
double,STOP_START_TIME5 double,STOP_END_TIME5 double,STOP_START_TIME6 
double,STOP_END_TIME6 double,STOP_START_TIME7 double,STOP_END_TIME7 
double,STOP_START_TIME8 double,STOP_END_TIME8 double,STOP_START_TIME9 
double,STOP_END_TIME9 double,STOP_START_TIME10 double,STOP_END_TIME10 
double,FAIL_CLASS double,RECORD_TYPE double,NODATA_COUNT 
double,VIDEO_NODATA_DURATION double,VIDEO_SMOOTH_DURATION 
double,VIDEO_SD_DURATION double,VIDEO_HD_DURATION double,VIDEO_UHD_DURATION 
double,VIDEO_FHD_DURATION double,FLUCTUATION double,START_DOWNLOAD_THROUGHPUT 
double,L7_UL_GOODPUT_FULL_MSS double,SESSIONKEY string,FIRST_UCELLID 
double,LAST_UCELLID double,UCELLID1 double,LONGITUDE1 double,LATITUDE1 
double,UCELLID2 double,LONGITUDE2 double,LATITUDE2 double,UCELLID3 
double,LONGITUDE3 double,LATITUDE3 double,UCELLID4 double,LONGITUDE4 
double,LATITUDE4 double,UCELLID5 double,LONGITUDE5 double,LATITUDE5 
double,UCELLID6 double,LONGITUDE6 double,LATITUDE6 double,UCELLID7 
double,LONGITUDE7 double,LATITUDE7 double,UCELLID8 double,LONGITUDE8 
double,LATITUDE8 double,UCELLID9 double,LONGITUDE9 double,LATITUDE9 
double,UCELLID10 double,LONGITUDE10 double,LATITUDE10 
double,INTBUFFER_FULL_DELAY double,STALL_DURATION double,STREAMING_DW_PACKETS 
double,STREAMING_DOWNLOAD_DELAY double,PLAY_DURATION double,STREAMING_QUALITY 
int,VIDEO_DATA_RATE double,AUDIO_DATA_RATE double,STREAMING_FILESIZE 
double,STREAMING_DURATIOIN double,TCP_SYN_TIME double,TCP_RTT_STEP1 
double,CHARGE_ID double,UL_REVERSE_TO_DL_DELAY double,DL_REVERSE_TO_UL_DELAY 
double,DATATRANS_DW_GOODPUT double,DATATRANS_DW_TOTAL_DURATION 
double,SUM_FRAGMENT_INTERVAL double,TCP_FIN_TIMES double,TCP_RESET_TIMES 
double,URL_CLASSIFICATION double,STREAMING_LQ_DURATIOIN double,MAX_DNS_DELAY 
double,MAX_DNS2SYN double,MAX_LATANCY_OF_LINK_SETUP double,MAX_SYNACK2FIRSTACK 

[jira] [Updated] (CARBONDATA-665) Comparision Failure occurs when we execute the same query in hive and Carbondata

2017-01-19 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-665?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-665:
-
 Attachment: Test_Data1.csv
Description: 
Orderby is not working , so records are not coming in sequence as well there is 
data difference and some values being stored as null

Data itself is stored incorrectly and is different from Hive
Spark version :1.6.2


Create 1 query : create table Test_Boundary (c1_int int,c2_Bigint 
Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string string,c6_Timestamp 
Timestamp,c7_Datatype_Desc string) STORED BY 'org.apache.carbondata.format'

Load 1 Query : LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO 
table Test_Boundary 
OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')

Create 2 query : create table Test_Boundary1 (c1_int int,c2_Bigint 
Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string string,c6_Timestamp 
Timestamp,c7_Datatype_Desc string) STORED BY 'org.apache.carbondata.format'

Load 2 query:  LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO 
table Test_Boundary1 
OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')

Select Query : 
select 
c1_int,c2_Bigint,c3_Decimal,c4_double,c5_string,c6_Timestamp,c7_Datatype_Desc 
from Test_Boundary where c2_bigint=c2_bigint

  was:
 Create 1 query : create table Test_Boundary (c1_int int,c2_Bigint 
Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string string,c6_Timestamp 
Timestamp,c7_Datatype_Desc string) STORED BY 'org.apache.carbondata.format'

Load 1 Query : LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO 
table Test_Boundary 
OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')

Create 2 query : create table Test_Boundary1 (c1_int int,c2_Bigint 
Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string string,c6_Timestamp 
Timestamp,c7_Datatype_Desc string) STORED BY 'org.apache.carbondata.format'

Load 2 query:  LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO 
table Test_Boundary1 
OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')

Select Query : 
select 
c1_int,c2_Bigint,c3_Decimal,c4_double,c5_string,c6_Timestamp,c7_Datatype_Desc 
from Test_Boundary where c2_bigint=c2_bigint


> Comparision Failure occurs when we execute the same query in hive and 
> Carbondata
> 
>
> Key: CARBONDATA-665
> URL: https://issues.apache.org/jira/browse/CARBONDATA-665
> Project: CarbonData
>  Issue Type: Bug
>Reporter: SWATI RAO
> Attachments: Test_Data1.csv
>
>
> Orderby is not working , so records are not coming in sequence as well there 
> is data difference and some values being stored as null
> Data itself is stored incorrectly and is different from Hive
> Spark version :1.6.2
> Create 1 query : create table Test_Boundary (c1_int int,c2_Bigint 
> Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format'
> Load 1 Query : LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO 
> table Test_Boundary 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
> Create 2 query : create table Test_Boundary1 (c1_int int,c2_Bigint 
> Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string 
> string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 
> 'org.apache.carbondata.format'
> Load 2 query:  LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO 
> table Test_Boundary1 
> OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='')
> Select Query : 
> select 
> c1_int,c2_Bigint,c3_Decimal,c4_double,c5_string,c6_Timestamp,c7_Datatype_Desc 
> from Test_Boundary where c2_bigint=c2_bigint



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (CARBONDATA-395) Unit Test cases for package org.apache.carbondata.scan.expression.ExpressionResult

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-395?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-395.


> Unit Test cases for package 
> org.apache.carbondata.scan.expression.ExpressionResult
> --
>
> Key: CARBONDATA-395
> URL: https://issues.apache.org/jira/browse/CARBONDATA-395
> Project: CarbonData
>  Issue Type: Test
>Reporter: SWATI RAO
>Priority: Trivial
> Fix For: 1.0.0-incubating
>
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (CARBONDATA-410) Implement test cases for core.datastore.file system

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-410?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-410.


> Implement test cases for core.datastore.file system
> ---
>
> Key: CARBONDATA-410
> URL: https://issues.apache.org/jira/browse/CARBONDATA-410
> Project: CarbonData
>  Issue Type: Test
>Reporter: SWATI RAO
> Fix For: 1.0.0-incubating
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (CARBONDATA-416) Add unit test case for result.impl package

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-416?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-416.


> Add unit test case for result.impl package
> --
>
> Key: CARBONDATA-416
> URL: https://issues.apache.org/jira/browse/CARBONDATA-416
> Project: CarbonData
>  Issue Type: Test
>Reporter: SWATI RAO
>Priority: Trivial
> Fix For: 1.0.0-incubating
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Reopened] (CARBONDATA-427) In documentation ,applicable is not define properly in Parameters of Compaction

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-427?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO reopened CARBONDATA-427:
--

> In documentation ,applicable is not define properly in Parameters of 
> Compaction 
> 
>
> Key: CARBONDATA-427
> URL: https://issues.apache.org/jira/browse/CARBONDATA-427
> Project: CarbonData
>  Issue Type: Bug
>  Components: docs
>Reporter: SWATI RAO
>
> Applicable is not define properly in Parameters of Compaction like
> In carbon.numberof.preserve.segments  applicable is Minor/Major, it should be 
> minor
> # carbon.allowed.compaction.days applicable is Minor/Major, it should be 
> minor
> #carbon.number.of.cores.while.compacting applicable is Minor/Major, it should 
> be minor



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (CARBONDATA-427) In documentation ,applicable is not define properly in Parameters of Compaction

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-427?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-427.

Resolution: Fixed

> In documentation ,applicable is not define properly in Parameters of 
> Compaction 
> 
>
> Key: CARBONDATA-427
> URL: https://issues.apache.org/jira/browse/CARBONDATA-427
> Project: CarbonData
>  Issue Type: Bug
>  Components: docs
>Reporter: SWATI RAO
>
> Applicable is not define properly in Parameters of Compaction like
> In carbon.numberof.preserve.segments  applicable is Minor/Major, it should be 
> minor
> # carbon.allowed.compaction.days applicable is Minor/Major, it should be 
> minor
> #carbon.number.of.cores.while.compacting applicable is Minor/Major, it should 
> be minor



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Reopened] (CARBONDATA-419) Show tables in query for non-existing database do not show any exception

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-419?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO reopened CARBONDATA-419:
--

> Show tables in query for non-existing database do not show any exception 
> -
>
> Key: CARBONDATA-419
> URL: https://issues.apache.org/jira/browse/CARBONDATA-419
> Project: CarbonData
>  Issue Type: Bug
>  Components: data-query
>Reporter: SWATI RAO
>Assignee: Manohar Vanam
>
> When executing command 'show tables in' for an non-existing database, it do 
> not show any error,
> Example: show tables in test_tableMg;
> (where test_tableMg database do not exist)
> whereas when executing the same query on hive, it says : 
> FAILED: SemanticException [Error 10072]: Database does not exist: test_tableMg



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (CARBONDATA-442) Query result mismatching with Hive

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-442?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-442.


> Query result mismatching with Hive
> --
>
> Key: CARBONDATA-442
> URL: https://issues.apache.org/jira/browse/CARBONDATA-442
> Project: CarbonData
>  Issue Type: Bug
>Reporter: SWATI RAO
> Fix For: 1.0.0-incubating
>
>
> => I created table using following command : 
> create table Carbon_automation_test5 (imei string,deviceInformationId int,MAC 
> string,deviceColor string,device_backColor string,modelId string,marketName 
> string,AMSize string,ROMSize string,CUPAudit string,CPIClocked string,series 
> string,productionDate string,bomCode string,internalModels string, 
> deliveryTime string, channelsId string,channelsName string , deliveryAreaId 
> string, deliveryCountry string, deliveryProvince string, deliveryCity 
> string,deliveryDistrict string, deliveryStreet string,oxSingleNumber string, 
> ActiveCheckTime string, ActiveAreaId string, ActiveCountry string, 
> ActiveProvince string, Activecity string, ActiveDistrict string, ActiveStreet 
> string, ActiveOperatorId string, Active_releaseId string, Active_EMUIVersion 
> string,Active_operaSysVersion string, Active_BacVerNumber string, 
> Active_BacFlashVer string,Active_webUIVersion string, Active_webUITypeCarrVer 
> string,Active_webTypeDataVerNumber string, Active_operatorsVersion string, 
> Active_phonePADPartitionedVersions string,Latest_YEAR int, Latest_MONTH int, 
> Latest_DAY int, Latest_HOUR string, Latest_areaId string, Latest_country 
> string, Latest_province string, Latest_city string,Latest_district string, 
> Latest_street string, Latest_releaseId string,Latest_EMUIVersion string, 
> Latest_operaSysVersion string, Latest_BacVerNumber string,Latest_BacFlashVer 
> string, Latest_webUIVersion string, Latest_webUITypeCarrVer 
> string,Latest_webTypeDataVerNumber string, Latest_operatorsVersion 
> string,Latest_phonePADPartitionedVersions string, Latest_operatorId 
> string,gamePointDescription string, gamePointId int,contractNumber int) 
> stored by 'org.apache.carbondata.format' 
> => Load csv to table : 
> LOAD DATA INPATH 'hdfs://localhost:54310/user/hduser/100_olap.csv' INTO table 
> Carbon_automation_test5 OPTIONS('DELIMITER'= ',' ,'QUOTECHAR'= '"', 
> 'FILEHEADER'= 
> 'imei,deviceInformationId,MAC,deviceColor,device_backColor,modelId,marketName,AMSize,ROMSize,CUPAudit,CPIClocked,series,productionDate,bomCode,internalModels,deliveryTime,channelsId,channelsName,deliveryAreaId,deliveryCountry,deliveryProvince,deliveryCity,deliveryDistrict,deliveryStreet,oxSingleNumber,contractNumber,ActiveCheckTime,ActiveAreaId,ActiveCountry,ActiveProvince,Activecity,ActiveDistrict,ActiveStreet,ActiveOperatorId,Active_releaseId,Active_EMUIVersion,Active_operaSysVersion,Active_BacVerNumber,Active_BacFlashVer,Active_webUIVersion,Active_webUITypeCarrVer,Active_webTypeDataVerNumber,Active_operatorsVersion,Active_phonePADPartitionedVersions,Latest_YEAR,Latest_MONTH,Latest_DAY,Latest_HOUR,Latest_areaId,Latest_country,Latest_province,Latest_city,Latest_district,Latest_street,Latest_releaseId,Latest_EMUIVersion,Latest_operaSysVersion,Latest_BacVerNumber,Latest_BacFlashVer,Latest_webUIVersion,Latest_webUITypeCarrVer,Latest_webTypeDataVerNumber,Latest_operatorsVersion,Latest_phonePADPartitionedVersions,Latest_operatorId,gamePointId,gamePointDescription')
> =>now executed SELECT querry : 
> SELECT Carbon_automation_test5.AMSize AS AMSize, 
> Carbon_automation_test5.ActiveCountry AS ActiveCountry, 
> Carbon_automation_test5.Activecity AS Activecity , 
> SUM(Carbon_automation_test5.gamePointId) AS Sum_gamePointId FROM ( SELECT 
> AMSize,ActiveCountry,gamePointId, Activecity FROM (select * from 
> Carbon_automation_test5) SUB_QRY ) Carbon_automation_test5 INNER JOIN ( 
> SELECT ActiveCountry, Activecity, AMSize FROM (select * from 
> Carbon_automation_test5) SUB_QRY ) Carbon_automation_vmall_test1 ON 
> Carbon_automation_test5.AMSize = Carbon_automation_vmall_test1.AMSize WHERE 
> NOT(Carbon_automation_test5.AMSize <= '3RAM size') GROUP BY 
> Carbon_automation_test5.AMSize, Carbon_automation_test5.ActiveCountry, 
> Carbon_automation_test5.Activecity ORDER BY Carbon_automation_test5.AMSize 
> ASC, Carbon_automation_test5.ActiveCountry ASC, 
> Carbon_automation_test5.Activecity ASC;
> +++-+--+--+
> |   AMSize   | ActiveCountry  | Activecity  | Sum_gamePointId  |
> +++-+--+--+
> | 4RAM size  | Chinese| changsha| 200860   |
> | 4RAM size  | Chinese| guangzhou   | 38016|
> | 4RAM size  | Chinese| shenzhen| 49610|
> | 4RAM size  | Chinese| wuhan   | 117568   |
> | 4RAM size  | 

[jira] [Closed] (CARBONDATA-444) Improved integration test-case for AllDataTypesTestCase1

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-444?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-444.


> Improved integration test-case for AllDataTypesTestCase1
> 
>
> Key: CARBONDATA-444
> URL: https://issues.apache.org/jira/browse/CARBONDATA-444
> Project: CarbonData
>  Issue Type: Task
>Reporter: SWATI RAO
> Fix For: 1.0.0-incubating
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (CARBONDATA-438) Add unit test for scan.scanner.impl package

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-438?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-438.


> Add unit test for scan.scanner.impl package
> ---
>
> Key: CARBONDATA-438
> URL: https://issues.apache.org/jira/browse/CARBONDATA-438
> Project: CarbonData
>  Issue Type: Test
>Reporter: SWATI RAO
>Assignee: Kunal Kapoor
>Priority: Trivial
> Fix For: 1.0.0-incubating
>
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (CARBONDATA-460) Add Unit Tests For core.writer.sortindex package

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-460?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-460.


> Add Unit Tests For core.writer.sortindex package 
> -
>
> Key: CARBONDATA-460
> URL: https://issues.apache.org/jira/browse/CARBONDATA-460
> Project: CarbonData
>  Issue Type: Test
>Reporter: SWATI RAO
>Priority: Trivial
> Fix For: 1.0.0-incubating
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (CARBONDATA-472) Improve code coverage for core.cache package.

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-472.


> Improve code coverage for core.cache package.
> -
>
> Key: CARBONDATA-472
> URL: https://issues.apache.org/jira/browse/CARBONDATA-472
> Project: CarbonData
>  Issue Type: Test
>Reporter: SWATI RAO
>Assignee: Kunal Kapoor
>Priority: Trivial
> Fix For: 1.0.0-incubating
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (CARBONDATA-445) Improved integration test-case for AllDataTypesTestCase3

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-445.


>  Improved integration test-case for AllDataTypesTestCase3
> -
>
> Key: CARBONDATA-445
> URL: https://issues.apache.org/jira/browse/CARBONDATA-445
> Project: CarbonData
>  Issue Type: Task
>Reporter: SWATI RAO
> Fix For: 1.0.0-incubating
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (CARBONDATA-450) Increase Test Coverage for Core.reader module

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-450?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-450.


> Increase Test Coverage for Core.reader module
> -
>
> Key: CARBONDATA-450
> URL: https://issues.apache.org/jira/browse/CARBONDATA-450
> Project: CarbonData
>  Issue Type: Test
>Reporter: SWATI RAO
> Fix For: 1.0.0-incubating
>
>  Time Spent: 1h 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (CARBONDATA-483) Add Unit Tests For core.carbon.metadata package

2017-01-18 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-483.


> Add Unit Tests For core.carbon.metadata package
> ---
>
> Key: CARBONDATA-483
> URL: https://issues.apache.org/jira/browse/CARBONDATA-483
> Project: CarbonData
>  Issue Type: Test
>Reporter: SWATI RAO
>Priority: Trivial
> Fix For: 1.0.0-incubating
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (CARBONDATA-394) Carbon Loading data from files having invalid extensions or no extension

2016-12-15 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-394?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO closed CARBONDATA-394.

Resolution: Not A Problem

> Carbon Loading data from files having invalid extensions or no extension
> 
>
> Key: CARBONDATA-394
> URL: https://issues.apache.org/jira/browse/CARBONDATA-394
> Project: CarbonData
>  Issue Type: Bug
>Reporter: SWATI RAO
>Assignee: xbkaishui
>Priority: Trivial
>
> When I try to run the following queries :
> LOAD DATA inpath 'hdfs://localhost:54310/user/hive/warehouse/file1.csv.csv' 
> INTO table empdata options('DELIMITER'=',', 'FILEHEADER'='id, 
> name','QUOTECHAR'='"');
> LOAD DATA inpath 
> 'hdfs://localhost:54310/user/hive/warehouse/file2.csv.csv.csv.csv' INTO table 
> empdata options('DELIMITER'=',', 'FILEHEADER'='id, name','QUOTECHAR'='"');
>  LOAD DATA inpath 'hdfs://localhost:54310/user/hive/warehouse/file3.txttt' 
> INTO table empdata options('DELIMITER'=',', 'FILEHEADER'='id, 
> name','QUOTECHAR'='"');
> LOAD DATA inpath 'hdfs://localhost:54310/user/hive/warehouse/file4' INTO 
> table empdata options('DELIMITER'=',', 'FILEHEADER'='id, 
> name','QUOTECHAR'='"');
> LOAD DATA inpath 
> 'hdfs://localhost:54310/user/hive/warehouse/file5.txt.bat.csv' INTO table 
> empdata options('DELIMITER'=',', 'FILEHEADER'='id, name','QUOTECHAR'='"');
> We should get Input File Errors, but the data is loaded successfully into the 
> Carbon table.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CARBONDATA-394) Carbon Loading data from files having invalid extensions or no extension

2016-12-15 Thread SWATI RAO (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-394?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15750987#comment-15750987
 ] 

SWATI RAO commented on CARBONDATA-394:
--

So we are closing this issue.

> Carbon Loading data from files having invalid extensions or no extension
> 
>
> Key: CARBONDATA-394
> URL: https://issues.apache.org/jira/browse/CARBONDATA-394
> Project: CarbonData
>  Issue Type: Bug
>Reporter: SWATI RAO
>Assignee: xbkaishui
>Priority: Trivial
>
> When I try to run the following queries :
> LOAD DATA inpath 'hdfs://localhost:54310/user/hive/warehouse/file1.csv.csv' 
> INTO table empdata options('DELIMITER'=',', 'FILEHEADER'='id, 
> name','QUOTECHAR'='"');
> LOAD DATA inpath 
> 'hdfs://localhost:54310/user/hive/warehouse/file2.csv.csv.csv.csv' INTO table 
> empdata options('DELIMITER'=',', 'FILEHEADER'='id, name','QUOTECHAR'='"');
>  LOAD DATA inpath 'hdfs://localhost:54310/user/hive/warehouse/file3.txttt' 
> INTO table empdata options('DELIMITER'=',', 'FILEHEADER'='id, 
> name','QUOTECHAR'='"');
> LOAD DATA inpath 'hdfs://localhost:54310/user/hive/warehouse/file4' INTO 
> table empdata options('DELIMITER'=',', 'FILEHEADER'='id, 
> name','QUOTECHAR'='"');
> LOAD DATA inpath 
> 'hdfs://localhost:54310/user/hive/warehouse/file5.txt.bat.csv' INTO table 
> empdata options('DELIMITER'=',', 'FILEHEADER'='id, name','QUOTECHAR'='"');
> We should get Input File Errors, but the data is loaded successfully into the 
> Carbon table.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-483) Add Unit Tests For core.carbon.metadata package

2016-12-02 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-483:


 Summary: Add Unit Tests For core.carbon.metadata package
 Key: CARBONDATA-483
 URL: https://issues.apache.org/jira/browse/CARBONDATA-483
 Project: CarbonData
  Issue Type: Test
Reporter: SWATI RAO
Priority: Trivial






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CARBONDATA-361) SMALL INT data showing unsupported datatype

2016-11-29 Thread SWATI RAO (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-361?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707814#comment-15707814
 ] 

SWATI RAO commented on CARBONDATA-361:
--

[~ravi.pesala] : As mentioned on your wiki page which is lastly updated on 10th 
Nov'16: SMALL INT data type should work but currently it isnot working. Either 
update the wiki page or we need to resolve this bug.

https://cwiki.apache.org/confluence/display/CARBONDATA/Carbon+Data+Types

> SMALL INT data showing unsupported datatype
> ---
>
> Key: CARBONDATA-361
> URL: https://issues.apache.org/jira/browse/CARBONDATA-361
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Prabhat Kashyap
>
> When I tried executing the following query with SMALLINT datatype
> create table testTable2(id SMALLINT, name String) stored by 'carbondata' ;
> It is giving me following exception.
> Error: org.apache.carbondata.spark.exception.MalformedCarbonCommandException: 
> Unsupported data type : FieldSchema(name:id, type:smallint, 
> comment:null).getType (state=,code=0)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-472) Improve code coverage for core.cache package.

2016-11-29 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-472:


 Summary: Improve code coverage for core.cache package.
 Key: CARBONDATA-472
 URL: https://issues.apache.org/jira/browse/CARBONDATA-472
 Project: CarbonData
  Issue Type: Test
Reporter: SWATI RAO
Priority: Trivial






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-457) Add Unit Tests For core.writer package

2016-11-28 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-457:


 Summary: Add Unit Tests For core.writer package 
 Key: CARBONDATA-457
 URL: https://issues.apache.org/jira/browse/CARBONDATA-457
 Project: CarbonData
  Issue Type: Test
Reporter: SWATI RAO
Priority: Trivial






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-450) Increase Test Coverage for Core.reader module

2016-11-25 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-450:


 Summary: Increase Test Coverage for Core.reader module
 Key: CARBONDATA-450
 URL: https://issues.apache.org/jira/browse/CARBONDATA-450
 Project: CarbonData
  Issue Type: Test
Reporter: SWATI RAO






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CARBONDATA-444) Improved integration test-case for AllDataTypesTestCase1

2016-11-23 Thread SWATI RAO (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-444?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SWATI RAO updated CARBONDATA-444:
-
Issue Type: Task  (was: Bug)

> Improved integration test-case for AllDataTypesTestCase1
> 
>
> Key: CARBONDATA-444
> URL: https://issues.apache.org/jira/browse/CARBONDATA-444
> Project: CarbonData
>  Issue Type: Task
>Reporter: SWATI RAO
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-444) Improved integration test-case for AllDataTypesTestCase1

2016-11-23 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-444:


 Summary: Improved integration test-case for AllDataTypesTestCase1
 Key: CARBONDATA-444
 URL: https://issues.apache.org/jira/browse/CARBONDATA-444
 Project: CarbonData
  Issue Type: Bug
Reporter: SWATI RAO






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-438) Add unit test for scan.scanner.impl package

2016-11-22 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-438:


 Summary: Add unit test for scan.scanner.impl package
 Key: CARBONDATA-438
 URL: https://issues.apache.org/jira/browse/CARBONDATA-438
 Project: CarbonData
  Issue Type: Test
Reporter: SWATI RAO
Priority: Trivial






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-434) Update test cases for AllDataTypesTestCase2

2016-11-21 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-434:


 Summary: Update test cases for AllDataTypesTestCase2
 Key: CARBONDATA-434
 URL: https://issues.apache.org/jira/browse/CARBONDATA-434
 Project: CarbonData
  Issue Type: Improvement
Reporter: SWATI RAO






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-435) improve integration test case for AllDataTypesTestCase4

2016-11-21 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-435:


 Summary: improve integration test case for AllDataTypesTestCase4
 Key: CARBONDATA-435
 URL: https://issues.apache.org/jira/browse/CARBONDATA-435
 Project: CarbonData
  Issue Type: Improvement
Reporter: SWATI RAO






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-427) In documentation ,applicable is not define properly in Parameters of Compaction

2016-11-20 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-427:


 Summary: In documentation ,applicable is not define properly in 
Parameters of Compaction 
 Key: CARBONDATA-427
 URL: https://issues.apache.org/jira/browse/CARBONDATA-427
 Project: CarbonData
  Issue Type: Bug
  Components: docs
Reporter: SWATI RAO


Applicable is not define properly in Parameters of Compaction like

In carbon.numberof.preserve.segmentsapplicable is Minor/Major, it should be 
minor

# carbon.allowed.compaction.days applicable is Minor/Major, it should be 
minor

#carbon.number.of.cores.while.compacting applicable is Minor/Major, it should 
be minor



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-410) Implement test cases for core.datastore.file system

2016-11-14 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-410:


 Summary: Implement test cases for core.datastore.file system
 Key: CARBONDATA-410
 URL: https://issues.apache.org/jira/browse/CARBONDATA-410
 Project: CarbonData
  Issue Type: Task
Reporter: SWATI RAO






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-408) Unable to create view from a table

2016-11-11 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-408:


 Summary: Unable to create view from a table
 Key: CARBONDATA-408
 URL: https://issues.apache.org/jira/browse/CARBONDATA-408
 Project: CarbonData
  Issue Type: Bug
Reporter: SWATI RAO
Priority: Trivial


When we tried to execute the following query to create view in carbon :

create view emp_view AS Select name,sal from demo2;

NOTE :demo2 table contains following columns: 
id Int,
name String, 
sal decimal

we got the following exception:
Error: org.apache.spark.sql.execution.QueryExecutionException: FAILED: 
SemanticException [Error 10004]: Line 1:31 Invalid table alias or column 
reference 'name': (possible column names are: col) (state=,code=0)

where as we are able to create view in hive using the same query.





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-407) Missing documentation for data load using data frames

2016-11-10 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-407:


 Summary: Missing documentation for data load using data frames
 Key: CARBONDATA-407
 URL: https://issues.apache.org/jira/browse/CARBONDATA-407
 Project: CarbonData
  Issue Type: Bug
  Components: data-load, docs
Reporter: SWATI RAO


Problem: Could not load data from a json file using data frames.

Suggestion:
Data can be loaded using data frames for the following cases :
1) Using case class (we have created a sample example for this)
2) From Json Data
3) Load data from existing parquet table to carbon data
4) Load data from one carbon table to another.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-395) Unit Test cases for package org.apache.carbondata.scan.expression.ExpressionResult

2016-11-08 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-395:


 Summary: Unit Test cases for package 
org.apache.carbondata.scan.expression.ExpressionResult
 Key: CARBONDATA-395
 URL: https://issues.apache.org/jira/browse/CARBONDATA-395
 Project: CarbonData
  Issue Type: Test
Reporter: SWATI RAO
Priority: Trivial






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-394) Carbon Loading data from files having invalid extensions or no extension

2016-11-08 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-394:


 Summary: Carbon Loading data from files having invalid extensions 
or no extension
 Key: CARBONDATA-394
 URL: https://issues.apache.org/jira/browse/CARBONDATA-394
 Project: CarbonData
  Issue Type: Bug
Reporter: SWATI RAO
Priority: Trivial


When I try to run the following queries :

LOAD DATA inpath 'hdfs://localhost:54310/user/hive/warehouse/file1.csv.csv' 
INTO table empdata options('DELIMITER'=',', 'FILEHEADER'='id, 
name','QUOTECHAR'='"');

LOAD DATA inpath 
'hdfs://localhost:54310/user/hive/warehouse/file2.csv.csv.csv.csv' INTO table 
empdata options('DELIMITER'=',', 'FILEHEADER'='id, name','QUOTECHAR'='"');

 LOAD DATA inpath 'hdfs://localhost:54310/user/hive/warehouse/file3.txttt' INTO 
table empdata options('DELIMITER'=',', 'FILEHEADER'='id, name','QUOTECHAR'='"');

LOAD DATA inpath 'hdfs://localhost:54310/user/hive/warehouse/file4' INTO table 
empdata options('DELIMITER'=',', 'FILEHEADER'='id, name','QUOTECHAR'='"');

LOAD DATA inpath 'hdfs://localhost:54310/user/hive/warehouse/file5.txt.bat.csv' 
INTO table empdata options('DELIMITER'=',', 'FILEHEADER'='id, 
name','QUOTECHAR'='"');

We should get Input File Errors, but the data is loaded successfully into the 
Carbon table.














--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CARBONDATA-378) Long datatype is not working

2016-11-07 Thread SWATI RAO (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-378?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15643479#comment-15643479
 ] 

SWATI RAO commented on CARBONDATA-378:
--

Okay. Thanks.

> Long datatype is not working
> 
>
> Key: CARBONDATA-378
> URL: https://issues.apache.org/jira/browse/CARBONDATA-378
> Project: CarbonData
>  Issue Type: Bug
>Reporter: SWATI RAO
>
> As it is mentioned in DataType.java file that long is supported datatype but 
> when I run query using long datatype its displays an error.
> e.g:
> create table testTable(id long, name String) stored by 'carbondata' ;
> Error: org.apache.spark.sql.AnalysisException: cannot recognize input near 
> 'long' ',' 'name' in column type; line 1 pos 29 (state=,code=0)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-378) Long datatype is not working

2016-11-04 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-378:


 Summary: Long datatype is not working
 Key: CARBONDATA-378
 URL: https://issues.apache.org/jira/browse/CARBONDATA-378
 Project: CarbonData
  Issue Type: Bug
Reporter: SWATI RAO


As it is mentioned in DataType.java file that long is supported datatype but 
when I run query using long datatype its displays an error.

e.g:
create table testTable(id long, name String) stored by 'carbondata' ;
Error: org.apache.spark.sql.AnalysisException: cannot recognize input near 
'long' ',' 'name' in column type; line 1 pos 29 (state=,code=0)




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CARBONDATA-370) Abnormal behaviour of datatype Int & bigInt

2016-11-03 Thread SWATI RAO (JIRA)
SWATI RAO created CARBONDATA-370:


 Summary: Abnormal behaviour of datatype Int & bigInt
 Key: CARBONDATA-370
 URL: https://issues.apache.org/jira/browse/CARBONDATA-370
 Project: CarbonData
  Issue Type: Bug
Reporter: SWATI RAO


The specified data-type range for Int & bigInt is:
INT –––> -2147483648 to 2147483647
BIGINT ––> -922337203685477580807 to 922337203685477580806
 but I found abnormal behaviour in these data-type . 
1.When I insert beyond range in Int it allows range even beyond bigInt. 
2.When I insert into bigInt it doesn’t accept all values following in its range.

 




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)