[GitHub] carbondata issue #1732: [CARBONDATA-1946] Exception thrown after alter data ...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1732 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2648/ ---
[GitHub] carbondata issue #1082: [CARBONDATA-1218] [GLOBAL SORT] In case of data-load...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1082 Build Success with Spark 2.2.0, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1246/ ---
[GitHub] carbondata issue #1082: [CARBONDATA-1218] [GLOBAL SORT] In case of data-load...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1082 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2470/ ---
[jira] [Created] (CARBONDATA-1963) Support S3 table with dictionary
Sangeeta Gulia created CARBONDATA-1963: -- Summary: Support S3 table with dictionary Key: CARBONDATA-1963 URL: https://issues.apache.org/jira/browse/CARBONDATA-1963 Project: CarbonData Issue Type: Task Reporter: Sangeeta Gulia Priority: Minor -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Assigned] (CARBONDATA-1960) Add example for creating a local table and load CSV data which is stored in S3.
[ https://issues.apache.org/jira/browse/CARBONDATA-1960?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jatin reassigned CARBONDATA-1960: - Assignee: Jatin > Add example for creating a local table and load CSV data which is stored in > S3. > --- > > Key: CARBONDATA-1960 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1960 > Project: CarbonData > Issue Type: Task >Reporter: Sangeeta Gulia >Assignee: Jatin >Priority: Trivial > -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Assigned] (CARBONDATA-1962) Support alter table add columns/drop columns on S3 table
[ https://issues.apache.org/jira/browse/CARBONDATA-1962?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jatin reassigned CARBONDATA-1962: - Assignee: Jatin > Support alter table add columns/drop columns on S3 table > > > Key: CARBONDATA-1962 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1962 > Project: CarbonData > Issue Type: Task >Reporter: Sangeeta Gulia >Assignee: Jatin >Priority: Minor > -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Assigned] (CARBONDATA-1961) Support data update/delete on S3 table
[ https://issues.apache.org/jira/browse/CARBONDATA-1961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jatin reassigned CARBONDATA-1961: - Assignee: Jatin > Support data update/delete on S3 table > -- > > Key: CARBONDATA-1961 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1961 > Project: CarbonData > Issue Type: Task >Reporter: Sangeeta Gulia >Assignee: Jatin >Priority: Minor > -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Assigned] (CARBONDATA-1959) Support compaction on S3 table
[ https://issues.apache.org/jira/browse/CARBONDATA-1959?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jatin reassigned CARBONDATA-1959: - Assignee: Jatin > Support compaction on S3 table > -- > > Key: CARBONDATA-1959 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1959 > Project: CarbonData > Issue Type: Task >Reporter: Sangeeta Gulia >Assignee: Jatin >Priority: Minor > -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata issue #1718: [CARBONDATA-1929][Validation]carbon property configu...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1718 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2647/ ---
[GitHub] carbondata issue #1732: [CARBONDATA-1946] Exception thrown after alter data ...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1732 Build Success with Spark 2.2.0, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1245/ ---
[GitHub] carbondata issue #1732: [CARBONDATA-1946] Exception thrown after alter data ...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1732 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2469/ ---
[jira] [Updated] (CARBONDATA-1827) Add Support to provide S3 Functionality in Carbondata
[ https://issues.apache.org/jira/browse/CARBONDATA-1827?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sangeeta Gulia updated CARBONDATA-1827: --- Issue Type: Task (was: New Feature) > Add Support to provide S3 Functionality in Carbondata > - > > Key: CARBONDATA-1827 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1827 > Project: CarbonData > Issue Type: Task > Components: core >Reporter: Sangeeta Gulia >Assignee: Jatin >Priority: Minor > Time Spent: 9h > Remaining Estimate: 0h > > Added Support to provide S3 Functionality in Carbondata. -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Assigned] (CARBONDATA-1958) CarbonSqlCliDriver not show result in case of select query
[ https://issues.apache.org/jira/browse/CARBONDATA-1958?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anubhav tarar reassigned CARBONDATA-1958: - Assignee: anubhav tarar > CarbonSqlCliDriver not show result in case of select query > -- > > Key: CARBONDATA-1958 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1958 > Project: CarbonData > Issue Type: Bug > Components: data-query >Affects Versions: 1.3.0 > Environment: sparksqlclidriver,hadoop2.7 >Reporter: anubhav tarar >Assignee: anubhav tarar > > 1.start the carbonsparksqlcli driver > go to bin folder of carbon project and execute command > ./carbon-spark-sql > 2.create carbon data table > spark-sql> create table sparktable(id int)stored by 'carbondata'; > 3.load data in carbontable > spark-sql> insert into sparktable values(1); > 4.query the data > 18/01/02 12:06:45 INFO TableInfo: main Table block size not specified for > default_sparktable. Therefore considering the default value 1024 MB > 18/01/02 12:06:45 INFO BlockletDataMap: main Time taken to load blocklet > datamap from file : > /home/anubhav/Documents/carbondata/carbondata/bin/carbonsqlclistore/default/sparktable/Fact/Part0/Segment_0/0_batchno0-0-1514874975013.carbonindexis > 19 > 18/01/02 12:06:45 INFO CarbonScanRDD: > Identified no.of.blocks: 0, > no.of.tasks: 0, > no.of.nodes: 0, > parallelism: 4 > > 18/01/02 12:06:45 INFO SparkContext: Starting job: processCmd at > CliDriver.java:376 > 18/01/02 12:06:45 INFO DAGScheduler: Job 4 finished: processCmd at > CliDriver.java:376, took 0.73 s > Time taken: 0.378 seconds > 18/01/02 12:06:45 INFO CliDriver: Time taken: 0.378 seconds > here no records get displayed -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Created] (CARBONDATA-1962) Support alter table add columns/drop columns on S3 table
Sangeeta Gulia created CARBONDATA-1962: -- Summary: Support alter table add columns/drop columns on S3 table Key: CARBONDATA-1962 URL: https://issues.apache.org/jira/browse/CARBONDATA-1962 Project: CarbonData Issue Type: Task Reporter: Sangeeta Gulia Priority: Minor -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Updated] (CARBONDATA-1952) Incorrect Result displays while applying delete query on table
[ https://issues.apache.org/jira/browse/CARBONDATA-1952?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kunal Kapoor updated CARBONDATA-1952: - Component/s: hive-integration > Incorrect Result displays while applying delete query on table > -- > > Key: CARBONDATA-1952 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1952 > Project: CarbonData > Issue Type: Bug > Components: hive-integration >Affects Versions: 1.3.0 > Environment: spark 2.1 >Reporter: Vandana Yadav >Priority: Minor > Attachments: 2000_UniqData.csv > > > Incorrect Result while applying delete query on the table. > Steps to reproduce: > 1)Create table: > CREATE TABLE uniqdata_delete (CUST_ID int,CUST_NAME > String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, > BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), > DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double, > INTEGER_COLUMN1 int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES > ("TABLE_BLOCKSIZE"= "256 MB") > 2)Load Data Into table: > LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into > table uniqdata_delete > OPTIONS('FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') > 3)Execute Query: > a) select count(*) from uniqdata_delete; > Output: > +---+--+ > | count(1) | > +---+--+ > | 2013 | > +---+--+ > 1 row selected (0.203 seconds) > b)delete from uniqdata_delete where CUST_ID NOT IN(9996,); > c) select count(*) from uniqdata_delete; > output: > +---+--+ > | count(1) | > +---+--+ > | 14| > +---+--+ > 1 row selected (0.22 seconds) > d) select * from uniqdata_delete; > output: > +--+--++++-+-+-+-+--+---+--+--+ > | cust_id |cust_name |active_emui_version | dob > | doj | bigint_column1 | bigint_column2 | > decimal_column1 | decimal_column2 |double_column1| > double_column2 | integer_column1 | > +--+--++++-+-+-+-+--+---+--+--+ > | NULL | || NULL > | NULL | NULL| NULL| NULL >| NULL| NULL | NULL > | NULL | > | NULL | || NULL > | NULL | 1233720368578 | NULL| NULL >| NULL| NULL | NULL > | NULL | > | NULL | || NULL > | NULL | NULL| -223372036854 | NULL >| NULL| NULL | NULL > | NULL | > | NULL | || NULL > | NULL | NULL| NULL| > 12345678901.123400 | NULL| NULL | > NULL | NULL | > | NULL | || NULL > | NULL | NULL| NULL| NULL >| 22345678901.123400 | NULL | NULL > | NULL | > | NULL | || NULL > | NULL | NULL| NULL| NULL >| NULL| 1.12345674897976E10 | NULL > | NULL | > | NULL | || NULL > | NULL | NULL| NULL| NULL >| NULL| NULL | > -1.12345674897976E10 | NULL | > | NULL | || NULL > | NULL | NULL| NULL| NULL >| NULL
[jira] [Updated] (CARBONDATA-1952) Incorrect Result displays while applying delete query on table
[ https://issues.apache.org/jira/browse/CARBONDATA-1952?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kunal Kapoor updated CARBONDATA-1952: - Component/s: (was: data-query) > Incorrect Result displays while applying delete query on table > -- > > Key: CARBONDATA-1952 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1952 > Project: CarbonData > Issue Type: Bug >Affects Versions: 1.3.0 > Environment: spark 2.1 >Reporter: Vandana Yadav >Priority: Minor > Attachments: 2000_UniqData.csv > > > Incorrect Result while applying delete query on the table. > Steps to reproduce: > 1)Create table: > CREATE TABLE uniqdata_delete (CUST_ID int,CUST_NAME > String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, > BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), > DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double, > INTEGER_COLUMN1 int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES > ("TABLE_BLOCKSIZE"= "256 MB") > 2)Load Data Into table: > LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into > table uniqdata_delete > OPTIONS('FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') > 3)Execute Query: > a) select count(*) from uniqdata_delete; > Output: > +---+--+ > | count(1) | > +---+--+ > | 2013 | > +---+--+ > 1 row selected (0.203 seconds) > b)delete from uniqdata_delete where CUST_ID NOT IN(9996,); > c) select count(*) from uniqdata_delete; > output: > +---+--+ > | count(1) | > +---+--+ > | 14| > +---+--+ > 1 row selected (0.22 seconds) > d) select * from uniqdata_delete; > output: > +--+--++++-+-+-+-+--+---+--+--+ > | cust_id |cust_name |active_emui_version | dob > | doj | bigint_column1 | bigint_column2 | > decimal_column1 | decimal_column2 |double_column1| > double_column2 | integer_column1 | > +--+--++++-+-+-+-+--+---+--+--+ > | NULL | || NULL > | NULL | NULL| NULL| NULL >| NULL| NULL | NULL > | NULL | > | NULL | || NULL > | NULL | 1233720368578 | NULL| NULL >| NULL| NULL | NULL > | NULL | > | NULL | || NULL > | NULL | NULL| -223372036854 | NULL >| NULL| NULL | NULL > | NULL | > | NULL | || NULL > | NULL | NULL| NULL| > 12345678901.123400 | NULL| NULL | > NULL | NULL | > | NULL | || NULL > | NULL | NULL| NULL| NULL >| 22345678901.123400 | NULL | NULL > | NULL | > | NULL | || NULL > | NULL | NULL| NULL| NULL >| NULL| 1.12345674897976E10 | NULL > | NULL | > | NULL | || NULL > | NULL | NULL| NULL| NULL >| NULL| NULL | > -1.12345674897976E10 | NULL | > | NULL | || NULL > | NULL | NULL| NULL| NULL >| NULL| NULL | NULL
[jira] [Updated] (CARBONDATA-1961) Support data update/delete on S3 table
[ https://issues.apache.org/jira/browse/CARBONDATA-1961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sangeeta Gulia updated CARBONDATA-1961: --- Priority: Minor (was: Major) > Support data update/delete on S3 table > -- > > Key: CARBONDATA-1961 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1961 > Project: CarbonData > Issue Type: Task >Reporter: Sangeeta Gulia >Priority: Minor > -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Created] (CARBONDATA-1961) Support data update/delete on S3 table
Sangeeta Gulia created CARBONDATA-1961: -- Summary: Support data update/delete on S3 table Key: CARBONDATA-1961 URL: https://issues.apache.org/jira/browse/CARBONDATA-1961 Project: CarbonData Issue Type: Task Reporter: Sangeeta Gulia -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Created] (CARBONDATA-1960) Add example for creating a local table and load CSV data which is stored in S3.
Sangeeta Gulia created CARBONDATA-1960: -- Summary: Add example for creating a local table and load CSV data which is stored in S3. Key: CARBONDATA-1960 URL: https://issues.apache.org/jira/browse/CARBONDATA-1960 Project: CarbonData Issue Type: Task Reporter: Sangeeta Gulia Priority: Trivial -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Updated] (CARBONDATA-1959) Support compaction on S3 table
[ https://issues.apache.org/jira/browse/CARBONDATA-1959?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sangeeta Gulia updated CARBONDATA-1959: --- Priority: Minor (was: Major) > Support compaction on S3 table > -- > > Key: CARBONDATA-1959 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1959 > Project: CarbonData > Issue Type: Task >Reporter: Sangeeta Gulia >Priority: Minor > -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Created] (CARBONDATA-1959) Support compaction on S3 table
Sangeeta Gulia created CARBONDATA-1959: -- Summary: Support compaction on S3 table Key: CARBONDATA-1959 URL: https://issues.apache.org/jira/browse/CARBONDATA-1959 Project: CarbonData Issue Type: Task Reporter: Sangeeta Gulia -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Created] (CARBONDATA-1958) CarbonSqlCliDriver not show result in case of select query
anubhav tarar created CARBONDATA-1958: - Summary: CarbonSqlCliDriver not show result in case of select query Key: CARBONDATA-1958 URL: https://issues.apache.org/jira/browse/CARBONDATA-1958 Project: CarbonData Issue Type: Bug Components: data-query Affects Versions: 1.3.0 Environment: sparksqlclidriver,hadoop2.7 Reporter: anubhav tarar 1.start the carbonsparksqlcli driver go to bin folder of carbon project and execute command ./carbon-spark-sql 2.create carbon data table spark-sql> create table sparktable(id int)stored by 'carbondata'; 3.load data in carbontable spark-sql> insert into sparktable values(1); 4.query the data 18/01/02 12:06:45 INFO TableInfo: main Table block size not specified for default_sparktable. Therefore considering the default value 1024 MB 18/01/02 12:06:45 INFO BlockletDataMap: main Time taken to load blocklet datamap from file : /home/anubhav/Documents/carbondata/carbondata/bin/carbonsqlclistore/default/sparktable/Fact/Part0/Segment_0/0_batchno0-0-1514874975013.carbonindexis 19 18/01/02 12:06:45 INFO CarbonScanRDD: Identified no.of.blocks: 0, no.of.tasks: 0, no.of.nodes: 0, parallelism: 4 18/01/02 12:06:45 INFO SparkContext: Starting job: processCmd at CliDriver.java:376 18/01/02 12:06:45 INFO DAGScheduler: Job 4 finished: processCmd at CliDriver.java:376, took 0.73 s Time taken: 0.378 seconds 18/01/02 12:06:45 INFO CliDriver: Time taken: 0.378 seconds here no records get displayed -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Created] (CARBONDATA-1957) create datamap query fails on table having dictionary_include
Geetika Gupta created CARBONDATA-1957: - Summary: create datamap query fails on table having dictionary_include Key: CARBONDATA-1957 URL: https://issues.apache.org/jira/browse/CARBONDATA-1957 Project: CarbonData Issue Type: Bug Components: data-query Affects Versions: 1.3.0 Environment: spark2.1 Reporter: Geetika Gupta Fix For: 1.3.0 Attachments: 2000_UniqData.csv I created a datamap using the following command: create datamap uniqdata_agg on table uniqdata using 'preaggregate' as select cust_id, cust_name,avg(decimal_column1) from uniqdata group by cust_id,cust_name; It throws the following error: Error: java.lang.Exception: DataLoad failure: (state=,code=0) Steps to reproduce: CREATE TABLE uniqdata(CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES('DICTIONARY_INCLUDE'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') Load command: LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/2000_UniqData.csv' into table uniqdata OPTIONS('DELIMITER'=',', 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') Create datamap commad: create datamap uniqdata_agg on table uniqdata using 'preaggregate' as select cust_id, cust_name,avg(decimal_column1) from uniqdata group by cust_id,cust_name; The above command throws the following exception: Error: java.lang.Exception: DataLoad failure: (state=,code=0) Here are the logs: 18/01/02 11:46:58 ERROR ParallelReadMergeSorterImpl: SafeParallelSorterPool:uniqdata_uniqdata_agg java.lang.IllegalArgumentException: requirement failed: Decimal precision 2922 exceeds max precision 38 at scala.Predef$.require(Predef.scala:224) at org.apache.spark.sql.types.Decimal.set(Decimal.scala:113) at org.apache.spark.sql.types.Decimal$.apply(Decimal.scala:426) at org.apache.spark.sql.types.Decimal.apply(Decimal.scala) at org.apache.spark.sql.catalyst.expressions.UnsafeRow.getDecimal(UnsafeRow.java:409) at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificSafeProjection.apply_0$(Unknown Source) at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificSafeProjection.apply(Unknown Source) at scala.collection.Iterator$$anon$11.next(Iterator.scala:409) at scala.collection.Iterator$$anon$11.next(Iterator.scala:409) at org.apache.carbondata.spark.rdd.LazyRddIterator.next(NewCarbonDataLoadRDD.scala:514) at org.apache.carbondata.spark.rdd.LazyRddIterator.next(NewCarbonDataLoadRDD.scala:477) at org.apache.carbondata.processing.loading.steps.InputProcessorStepImpl$InputProcessorIterator.getBatch(InputProcessorStepImpl.java:239) at org.apache.carbondata.processing.loading.steps.InputProcessorStepImpl$InputProcessorIterator.next(InputProcessorStepImpl.java:200) at org.apache.carbondata.processing.loading.steps.InputProcessorStepImpl$InputProcessorIterator.next(InputProcessorStepImpl.java:129) at org.apache.carbondata.processing.loading.steps.DataConverterProcessorStepImpl$1.next(DataConverterProcessorStepImpl.java:97) at org.apache.carbondata.processing.loading.steps.DataConverterProcessorStepImpl$1.next(DataConverterProcessorStepImpl.java:83) at org.apache.carbondata.processing.loading.sort.impl.ParallelReadMergeSorterImpl$SortIteratorThread.run(ParallelReadMergeSorterImpl.java:218) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 18/01/02 11:46:58 ERROR ForwardDictionaryCache: SafeParallelSorterPool:uniqdata_uniqdata_agg Error loading the dictionary: null 18/01/02 11:46:58 ERROR ForwardDictionaryCache: SafeParallelSorterPool:uniqdata_uniqdata_agg Error loading the dictionary: null 18/01/02 11:46:58 ERROR ForwardDictionaryCache: SafeParallelSorterPool:uniqdata_uniqdata_agg Error loading the dictionary: null 18/01/02 11:46:58 ERROR ForwardDictionaryCache: SafeParallelSorterPool:uniqdata_uniqdata_agg Error loading the dictionary: null 18/01/02 11:46:58 ERROR ParallelReadMergeSorterImpl: SafeParallelSorterPool:uniqdata_uniqdata_agg java.lang.IllegalArgumentException: requirement failed: Decimal precision 3128 exceeds max precision 38 at
[GitHub] carbondata pull request #1746: [TEST] Enable pre-aggregate tests as CI issue...
GitHub user kunal642 opened a pull request: https://github.com/apache/carbondata/pull/1746 [TEST] Enable pre-aggregate tests as CI issue is fixed Be sure to do all of the following checklist to help us incorporate your contribution quickly and easily: - [ ] Any interfaces changed? - [ ] Any backward compatibility impacted? - [ ] Document update required? - [ ] Testing done Please provide details on - Whether new unit test cases have been added or why no new tests are required? - How it is tested? Please attach test report. - Is it a performance related change? Please attach the performance test report. - Any additional information to help reviewers in testing this change. - [ ] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA. You can merge this pull request into a Git repository by running: $ git pull https://github.com/kunal642/carbondata enable_preagg_tests Alternatively you can review and apply these changes as the patch at: https://github.com/apache/carbondata/pull/1746.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #1746 commit 6e47cb226d5eefe45b6a8ec1c01cf84922e611e6 Author: kunal642 Date: 2018-01-02T06:30:09Z Enable pre-aggregate tests as CI issue is fixed ---
[GitHub] carbondata pull request #1745: [CARBONDATA-1955] Delta DataType calculation ...
Github user mohammadshahidkhan closed the pull request at: https://github.com/apache/carbondata/pull/1745 ---
[GitHub] carbondata issue #1745: [CARBONDATA-1955] Delta DataType calculation is inco...
Github user mohammadshahidkhan commented on the issue: https://github.com/apache/carbondata/pull/1745 Closed after discussion with Ravindra ---
[GitHub] carbondata issue #1702: [CARBONDATA-1896] Clean files operation improvement
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1702 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2646/ ---
[GitHub] carbondata issue #1104: [CARBONDATA-1239] Add validation for set command par...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1104 Build Success with Spark 2.2.0, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1244/ ---
[GitHub] carbondata issue #1104: [CARBONDATA-1239] Add validation for set command par...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1104 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2468/ ---
[GitHub] carbondata issue #1718: [CARBONDATA-1929][Validation]carbon property configu...
Github user manishgupta88 commented on the issue: https://github.com/apache/carbondata/pull/1718 retest sdv please ---
[jira] [Created] (CARBONDATA-1956) Select query with sum, count and avg throws exception for pre aggregate table
Geetika Gupta created CARBONDATA-1956: - Summary: Select query with sum, count and avg throws exception for pre aggregate table Key: CARBONDATA-1956 URL: https://issues.apache.org/jira/browse/CARBONDATA-1956 Project: CarbonData Issue Type: Bug Components: data-query Affects Versions: 1.3.0 Environment: spark2.1 Reporter: Geetika Gupta Fix For: 1.3.0 Attachments: 2000_UniqData.csv I create a datamap using the following command: create datamap uniqdata_agg_d on table uniqdata_29 using 'preaggregate' as select sum(decimal_column1), count(cust_id), avg(bigint_column1) from uniqdata_29 group by cust_id; The datamap creation was successfull, but when I tried the following query: select sum(decimal_column1), count(cust_id), avg(bigint_column1) from uniqdata_29 group by cust_id; It throws the following exception: Error: org.apache.spark.sql.AnalysisException: cannot resolve '(sum(uniqdata_29_uniqdata_agg_d.`uniqdata_29_bigint_column1_sum`) / sum(uniqdata_29_uniqdata_agg_d.`uniqdata_29_bigint_column1_count`))' due to data type mismatch: '(sum(uniqdata_29_uniqdata_agg_d.`uniqdata_29_bigint_column1_sum`) / sum(uniqdata_29_uniqdata_agg_d.`uniqdata_29_bigint_column1_count`))' requires (double or decimal) type, not bigint;; 'Aggregate [uniqdata_29_cust_id_count#244], [sum(uniqdata_29_decimal_column1_sum#243) AS sum(decimal_column1)#274, sum(cast(uniqdata_29_cust_id_count#244 as bigint)) AS count(cust_id)#276L, (sum(uniqdata_29_bigint_column1_sum#245L) / sum(uniqdata_29_bigint_column1_count#246L)) AS avg(bigint_column1)#279] +- Relation[uniqdata_29_decimal_column1_sum#243,uniqdata_29_cust_id_count#244,uniqdata_29_bigint_column1_sum#245L,uniqdata_29_bigint_column1_count#246L] CarbonDatasourceHadoopRelation [ Database name :28dec, Table name :uniqdata_29_uniqdata_agg_d, Schema :Some(StructType(StructField(uniqdata_29_decimal_column1_sum,DecimalType(30,10),true), StructField(uniqdata_29_cust_id_count,IntegerType,true), StructField(uniqdata_29_bigint_column1_sum,LongType,true), StructField(uniqdata_29_bigint_column1_count,LongType,true))) ] (state=,code=0) Steps for creation of maintable: CREATE TABLE uniqdata_29(CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 int) STORED BY 'org.apache.carbondata.format'; Load command: LOAD DATA INPATH 'hdfs://localhost:54311/Files/2000_UniqData.csv' into table uniqdata_29 OPTIONS('DELIMITER'=',', 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1'); Datamap creation command: create datamap uniqdata_agg_d on table uniqdata_29 using 'preaggregate' as select sum(decimal_column1), count(cust_id), avg(bigint_column1) from uniqdata_29 group by cust_id; Note: sum(decimal_column1), count(cust_id), avg(bigint_column1) from uniqdata_29 group by cust_id; executed successfully on maintable -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata issue #1104: [CARBONDATA-1239] Add validation for set command par...
Github user mohammadshahidkhan commented on the issue: https://github.com/apache/carbondata/pull/1104 retest this please ---
[GitHub] carbondata issue #1702: [CARBONDATA-1896] Clean files operation improvement
Github user dhatchayani commented on the issue: https://github.com/apache/carbondata/pull/1702 Retest sdv please ---
[GitHub] carbondata issue #1720: [CARBONDATA-1935]fix the backword compatibility issu...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1720 Build Success with Spark 2.2.0, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1243/ ---
[GitHub] carbondata issue #1720: [CARBONDATA-1935]fix the backword compatibility issu...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1720 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2467/ ---
[jira] [Closed] (CARBONDATA-1790) (Carbon1.3.0 - Streaming) Data load in Stream Segment fails if batch load is performed in between the streaming
[ https://issues.apache.org/jira/browse/CARBONDATA-1790?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ramakrishna S closed CARBONDATA-1790. - Resolution: Fixed This is is resolved in the B007 release. > (Carbon1.3.0 - Streaming) Data load in Stream Segment fails if batch load is > performed in between the streaming > --- > > Key: CARBONDATA-1790 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1790 > Project: CarbonData > Issue Type: Bug > Components: data-query >Affects Versions: 1.3.0 > Environment: 3 node ant cluster >Reporter: Ramakrishna S >Assignee: Bhavya Aggarwal > Labels: DFX > > Steps : > 1. Create a streaming table and do a batch load > 2. Set up the Streaming , so that it does streaming in chunk of 1000 records > 20 times > 3. Do another batch load on the table > 4. Do one more time streaming > +-++--+--+--++--+ > | Segment Id | Status | Load Start Time | Load End Time >| File Format | Merged To | > +-++--+--+--++--+ > | 2 | Success| 2017-11-21 21:42:36.77 | 2017-11-21 > 21:42:40.396 | COLUMNAR_V3 | NA | > | 1 | Streaming | 2017-11-21 21:40:46.2| NULL >| ROW_V1 | NA | > | 0 | Success| 2017-11-21 21:40:39.782 | 2017-11-21 > 21:40:43.168 | COLUMNAR_V3 | NA | > +-++--+--+--++--+ > *+Expected:+* Data should be loaded > *+Actual+* : Data load fiails > 1. One addition offset file is created(marked in bold) > -rw-r--r-- 2 root users 62 2017-11-21 21:40 > /user/hive/warehouse/Ram/default/stream_table5/.streaming/checkpoint/offsets/0 > -rw-r--r-- 2 root users 63 2017-11-21 21:40 > /user/hive/warehouse/Ram/default/stream_table5/.streaming/checkpoint/offsets/1 > -rw-r--r-- 2 root users 63 2017-11-21 21:42 > /user/hive/warehouse/Ram/default/stream_table5/.streaming/checkpoint/offsets/10 > -rw-r--r-- 2 root users 63 2017-11-21 21:40 > /user/hive/warehouse/Ram/default/stream_table5/.streaming/checkpoint/offsets/2 > -rw-r--r-- 2 root users 63 2017-11-21 21:41 > /user/hive/warehouse/Ram/default/stream_table5/.streaming/checkpoint/offsets/3 > -rw-r--r-- 2 root users 64 2017-11-21 21:41 > /user/hive/warehouse/Ram/default/stream_table5/.streaming/checkpoint/offsets/4 > -rw-r--r-- 2 root users 64 2017-11-21 21:41 > /user/hive/warehouse/Ram/default/stream_table5/.streaming/checkpoint/offsets/5 > -rw-r--r-- 2 root users 64 2017-11-21 21:41 > /user/hive/warehouse/Ram/default/stream_table5/.streaming/checkpoint/offsets/6 > -rw-r--r-- 2 root users 64 2017-11-21 21:41 > /user/hive/warehouse/Ram/default/stream_table5/.streaming/checkpoint/offsets/7 > -rw-r--r-- 2 root users 64 2017-11-21 21:41 > /user/hive/warehouse/Ram/default/stream_table5/.streaming/checkpoint/offsets/8 > *-rw-r--r-- 2 root users 63 2017-11-21 21:42 > /user/hive/warehouse/Ram/default/stream_table5/.streaming/checkpoint/offsets/9* > 2. Following error thrown: > === Streaming Query === > Identifier: [id = 3a5334bc-d471-4676-b6ce-f21105d491d1, runId = > b2be9f97-8141-46be-89db-9a0f98d13369] > Current Offsets: > {org.apache.spark.sql.execution.streaming.TextSocketSource@14c45193: 1000} > Current State: ACTIVE > Thread State: RUNNABLE > Logical Plan: > org.apache.spark.sql.execution.streaming.TextSocketSource@14c45193 > at > org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches(StreamExecution.scala:284) > at > org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:177) > Caused by: java.lang.RuntimeException: Offsets committed out of order: 20019 > followed by 1000 > at scala.sys.package$.error(package.scala:27) > at > org.apache.spark.sql.execution.streaming.TextSocketSource.commit(socket.scala:151) > at > org.apache.spark.sql.execution.streaming.StreamExecution$$anonfun$org$apache$spark$sql$execution$streaming$StreamExecution$$constructNextBatch$2$$anonfun$apply$mcV$sp$4.apply(StreamExecution.scala:421) > at > org.apache.spark.sql.execution.streaming.StreamExecution$$anonfun$org$apache$spark$sql$execution$streaming$StreamExecution$$constructNextBatch$2$$anonfun$apply$mcV$sp$4.apply(StreamExecution.scala:420) > at scala.collection.Iterator$class.foreach(Itera
[GitHub] carbondata issue #1720: [CARBONDATA-1935]fix the backword compatibility issu...
Github user akashrn5 commented on the issue: https://github.com/apache/carbondata/pull/1720 retest this please ---
[GitHub] carbondata issue #1743: [CARBONDATA-1954] HiveMetastore updated while droppi...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1743 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2645/ ---
[GitHub] carbondata issue #1743: [CARBONDATA-1954] HiveMetastore updated while droppi...
Github user rahulforallp commented on the issue: https://github.com/apache/carbondata/pull/1743 retest sdv please ---
[GitHub] carbondata issue #1738: [CARBONDATA-1932]Add version info for CarbonData
Github user zzcclp commented on the issue: https://github.com/apache/carbondata/pull/1738 @xubo245 , we don't package pom.xml file into release jar, so can't directly read version info from pom.xml ---
[GitHub] carbondata issue #1738: [CARBONDATA-1932]Add version info for CarbonData
Github user xubo245 commented on the issue: https://github.com/apache/carbondata/pull/1738 Can we obtain directly the CarbonData version from pom.xml ? ---
[GitHub] carbondata issue #1707: [CARBONDATA-1839] [DataLoad] Fix bugs and optimize i...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1707 Build Success with Spark 2.2.0, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1242/ ---
[GitHub] carbondata issue #1707: [CARBONDATA-1839] [DataLoad] Fix bugs and optimize i...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1707 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2466/ ---
[GitHub] carbondata issue #1707: [CARBONDATA-1839] [DataLoad] Fix bugs and optimize i...
Github user xuchuanyin commented on the issue: https://github.com/apache/carbondata/pull/1707 retest this please ---
[GitHub] carbondata issue #1742: [CARBONDATA-1953]Pre-aggregate Should inherit sort c...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1742 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2644/ ---
[GitHub] carbondata issue #1742: [CARBONDATA-1953]Pre-aggregate Should inherit sort c...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1742 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2643/ ---
[GitHub] carbondata issue #1742: [CARBONDATA-1953]Pre-aggregate Should inherit sort c...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1742 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2465/ ---
[GitHub] carbondata issue #1742: [CARBONDATA-1953]Pre-aggregate Should inherit sort c...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1742 Build Success with Spark 2.2.0, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1241/ ---