[jira] [Created] (CARBONDATA-862) USE_KETTLE option described in dml-operation-on-carbondata.md document doesn't work
Divya Gupta created CARBONDATA-862: -- Summary: USE_KETTLE option described in dml-operation-on-carbondata.md document doesn't work Key: CARBONDATA-862 URL: https://issues.apache.org/jira/browse/CARBONDATA-862 Project: CarbonData Issue Type: Bug Components: docs Affects Versions: 1.0.0-incubating Reporter: Divya Gupta Assignee: Divya Gupta Priority: Minor -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[jira] [Updated] (CARBONDATA-604) Use Kettle wrong option in Thrift server give wrong error message
[ https://issues.apache.org/jira/browse/CARBONDATA-604?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-604: --- Affects Version/s: 1.0.0-incubating > Use Kettle wrong option in Thrift server give wrong error message > - > > Key: CARBONDATA-604 > URL: https://issues.apache.org/jira/browse/CARBONDATA-604 > Project: CarbonData > Issue Type: Bug >Affects Versions: 1.0.0-incubating > Environment: Spark 1.6 >Reporter: Harmeet Singh >Assignee: Ravindra Pesala >Priority: Trivial > > I am using Thrift Server with following configuration: > -Duse.kettle=default > Here kettle value is wrong and when I run the thrift server, the thrift > server is running successfully. After that, I am going to start beeline. In > beeline, I am executing load command without any kettle option as below: > LOAD DATA INPATH 'hdfs://localhost:54310/csvs/3000_UniqData.csv' into table > uniqdata OPTIONS('DELIMITER'=',' , > 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1'); > After executing above command, I am getting Error: > java.lang.IllegalArgumentException: For input string: "default" > (state=,code=0) > This error, seems to be wrong, because I am not using any default value in > load query, and default is used with thrift. > So I have following Suggestions: > 1. Change the error message, while load query is executing. > 2. While starting thrift server with wrong value, thrift give us an error and > not start. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-665) Comparision Failure occurs when we execute the same query in hive and Carbondata
[ https://issues.apache.org/jira/browse/CARBONDATA-665?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-665: --- Environment: Spark 1.6 > Comparision Failure occurs when we execute the same query in hive and > Carbondata > > > Key: CARBONDATA-665 > URL: https://issues.apache.org/jira/browse/CARBONDATA-665 > Project: CarbonData > Issue Type: Bug >Affects Versions: 1.0.0-incubating > Environment: Spark 1.6 >Reporter: SWATI RAO > Attachments: Test_Data1.csv > > > Orderby is not working , so records are not coming in sequence as well there > is data difference and some values being stored as null > Data itself is stored incorrectly and is different from Hive > Spark version :1.6.2 > Create 1 query : create table Test_Boundary (c1_int int,c2_Bigint > Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string > string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY > 'org.apache.carbondata.format' > Load 1 Query : LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO > table Test_Boundary > OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='') > Create 2 query : create table Test_Boundary1 (c1_int int,c2_Bigint > Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string > string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY > 'org.apache.carbondata.format' > Load 2 query: LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO > table Test_Boundary1 > OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='') > Select Query : > select > c1_int,c2_Bigint,c3_Decimal,c4_double,c5_string,c6_Timestamp,c7_Datatype_Desc > from Test_Boundary where c2_bigint=c2_bigint -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-631) Select,Delete and Insert Query Failing for table created in 0.2 with data loaded in 1.0
[ https://issues.apache.org/jira/browse/CARBONDATA-631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-631: --- Affects Version/s: 1.0.0-incubating > Select,Delete and Insert Query Failing for table created in 0.2 with data > loaded in 1.0 > --- > > Key: CARBONDATA-631 > URL: https://issues.apache.org/jira/browse/CARBONDATA-631 > Project: CarbonData > Issue Type: Bug >Affects Versions: 1.0.0-incubating > Environment: Spark 1.6 >Reporter: Pallavi Singh >Assignee: kumar vishal > Fix For: NONE > > > Created table with the 0.2 jar: > CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION > string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 > bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 > decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 > int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES > ("TABLE_BLOCKSIZE"= "256 MB"); > then > LOAD DATA INPATH 'hdfs://localhost:54310/csv/2000_UniqData.csv' into table > uniqdata OPTIONS('DELIMITER'=',' , > 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1'); > Switched to 1.0 jar > LOAD DATA INPATH 'hdfs://localhost:54310/csv/2000_UniqData.csv' into table > uniqdata OPTIONS('DELIMITER'=',' , > 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1'); > After successful load : > select count(*) from uniqdata; > I get following error : > INFO 12-01 18:31:04,057 - Running query 'select count(*) from uniqdata' with > 81129cf3-fcd4-429d-9adf-d37d35cdf051 > INFO 12-01 18:31:04,058 - pool-27-thread-46 Query [SELECT COUNT(*) FROM > UNIQDATA] > INFO 12-01 18:31:04,060 - Parsing command: select count(*) from uniqdata > INFO 12-01 18:31:04,060 - Parse Completed > INFO 12-01 18:31:04,061 - Parsing command: select count(*) from uniqdata > INFO 12-01 18:31:04,061 - Parse Completed > INFO 12-01 18:31:04,061 - 27: get_table : db=12jan17 tbl=uniqdata > INFO 12-01 18:31:04,061 - ugi=pallaviip=unknown-ip-addr > cmd=get_table : db=12jan17 tbl=uniqdata > INFO 12-01 18:31:04,061 - 27: Opening raw store with implemenation > class:org.apache.hadoop.hive.metastore.ObjectStore > INFO 12-01 18:31:04,063 - ObjectStore, initialize called > INFO 12-01 18:31:04,068 - Reading in results for query > "org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is > closing > INFO 12-01 18:31:04,069 - Using direct SQL, underlying DB is DERBY > INFO 12-01 18:31:04,069 - Initialized ObjectStore > INFO 12-01 18:31:04,101 - pool-27-thread-46 Starting to optimize plan > ERROR 12-01 18:31:04,168 - pool-27-thread-46 Cannot convert12-01-2017 > 16:02:28 to Time/Long type valueUnparseable date: "12-01-2017 16:02:28" > ERROR 12-01 18:31:04,185 - pool-27-thread-46 Cannot convert12-01-2017 > 16:02:08 to Time/Long type valueUnparseable date: "12-01-2017 16:02:08" > ERROR 12-01 18:31:04,185 - pool-27-thread-46 Cannot convert12-01-2017 > 16:02:08 to Time/Long type valueUnparseable date: "12-01-2017 16:02:08" > ERROR 12-01 18:31:04,204 - pool-27-thread-46 Cannot convert12-01-2017 > 16:02:08 to Time/Long type valueUnparseable date: "12-01-2017 16:02:08" > ERROR 12-01 18:31:04,210 - Error executing query, currentState RUNNING, > org.apache.spark.sql.catalyst.errors.package$TreeNodeException: execute, tree: > CarbonDictionaryDecoder [CarbonDecoderRelation(Map(dob#280 -> dob#280, > double_column1#287 -> double_column1#287, decimal_column1#285 -> > decimal_column1#285, cust_id#282L -> cust_id#282L, integer_column1#289L -> > integer_column1#289L, decimal_column2#286 -> decimal_column2#286, > cust_name#278 -> cust_name#278, double_column2#288 -> double_column2#288, > active_emui_version#279 -> active_emui_version#279, bigint_column1#283L -> > bigint_column1#283L, bigint_column2#284L -> bigint_column2#284L, doj#281 -> > doj#281),CarbonDatasourceRelation(`12jan17`.`uniqdata`,None))], > ExcludeProfile(ArrayBuffer()), CarbonAliasDecoderRelation() > +- TungstenAggregate(key=[], > functions=[(count(1),mode=Final,isDistinct=false)], output=[_c0#750L]) >+- TungstenExchange SinglePartition, None > +- TungstenAggregate(key=[], > functions=[(count(1),mode=Partial,isDistinct=false)], output=[count#754L]) > +- CarbonScan CarbonRelation 12jan17, uniqdata, > CarbonMetaData(ArrayBuffer(cust_name, active_emui_version, dob, > doj),ArrayBuffer(cust_id, bigint_column1,
[jira] [Updated] (CARBONDATA-678) Corr function is not working for double datatype.
[ https://issues.apache.org/jira/browse/CARBONDATA-678?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-678: --- Affects Version/s: 1.0.0-incubating > Corr function is not working for double datatype. > - > > Key: CARBONDATA-678 > URL: https://issues.apache.org/jira/browse/CARBONDATA-678 > Project: CarbonData > Issue Type: Bug > Components: sql >Affects Versions: 1.0.0-incubating > Environment: Spark 1.6 >Reporter: SWATI RAO >Priority: Trivial > Fix For: 1.1.0-incubating > > Attachments: 2000_UniqData.csv, 3000_1_UniqData.csv, > 4000_UniqData.csv, 5000_UniqData.csv, 6000_UniqData.csv, 7000_UniqData.csv > > > We have created a table uniqdata_h and then we upload data in this table then > we perform select query having data-type double then it displays null value > but when we perform select query with integer or bigint data-type it displays > a value. > "Double" data-type is not working. > CREATE QUERY : > 0: jdbc:hive2://localhost:1> CREATE TABLE uniqdata_h (CUST_ID > int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ > timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 > decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, > Double_COLUMN2 double,INTEGER_COLUMN1 int) ROW FORMAT DELIMITED FIELDS > TERMINATED BY ','; > +-+--+ > | result | > +-+--+ > +-+--+ > No rows selected (4.249 seconds) > LOAD 1 : > 0: jdbc:hive2://localhost:1> load data local inpath > '/opt/Carbon/CarbonData/TestData/Data/uniqdata/2000_UniqData.csv' into table > uniqdata_h; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (1.176 seconds) > LOAD 2 : > 0: jdbc:hive2://localhost:1> load data local inpath > '/opt/Carbon/CarbonData/TestData/Data/uniqdata/4000_UniqData.csv' into table > uniqdata_h; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (0.485 seconds) > LOAD 3 : > 0: jdbc:hive2://localhost:1> load data local inpath > '/opt/Carbon/CarbonData/TestData/Data/uniqdata/5000_UniqData.csv' into table > uniqdata_h; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (0.365 seconds) > LOAD 4 : > 0: jdbc:hive2://localhost:1> load data local inpath > '/opt/Carbon/CarbonData/TestData/Data/uniqdata/6000_UniqData.csv' into table > uniqdata_h; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (0.594 seconds) > LOAD 5 : > 0: jdbc:hive2://localhost:1> load data local inpath > '/opt/Carbon/CarbonData/TestData/Data/uniqdata/7000_UniqData.csv' into table > uniqdata_h; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (0.753 seconds) > LOAD 6 : > 0: jdbc:hive2://localhost:1> load data local inpath > '/opt/Carbon/CarbonData/TestData/Data/uniqdata/3000_1_UniqData.csv' into > table uniqdata_h; > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (0.417 seconds) > SELECT QUERY : double data-type displays NULL value > 0: jdbc:hive2://localhost:1> select corr(Double_COLUMN1,Double_COLUMN1) > as a from uniqdata_h ; > +---+--+ > | a | > +---+--+ > | NULL | > +---+--+ > whereas select query with integer & bigdata type displays some value. > 0: jdbc:hive2://localhost:1> select corr(integer_COLUMN1,bigint_COLUMN1) > as a from uniqdata_h ; > +-+--+ > | a | > +-+--+ > | 1.0002 | > +-+--+ > 1 row selected (0.54 seconds) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-665) Comparision Failure occurs when we execute the same query in hive and Carbondata
[ https://issues.apache.org/jira/browse/CARBONDATA-665?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-665: --- Affects Version/s: 1.0.0-incubating > Comparision Failure occurs when we execute the same query in hive and > Carbondata > > > Key: CARBONDATA-665 > URL: https://issues.apache.org/jira/browse/CARBONDATA-665 > Project: CarbonData > Issue Type: Bug >Affects Versions: 1.0.0-incubating >Reporter: SWATI RAO > Attachments: Test_Data1.csv > > > Orderby is not working , so records are not coming in sequence as well there > is data difference and some values being stored as null > Data itself is stored incorrectly and is different from Hive > Spark version :1.6.2 > Create 1 query : create table Test_Boundary (c1_int int,c2_Bigint > Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string > string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY > 'org.apache.carbondata.format' > Load 1 Query : LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO > table Test_Boundary > OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='') > Create 2 query : create table Test_Boundary1 (c1_int int,c2_Bigint > Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string > string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY > 'org.apache.carbondata.format' > Load 2 query: LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO > table Test_Boundary1 > OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='') > Select Query : > select > c1_int,c2_Bigint,c3_Decimal,c4_double,c5_string,c6_Timestamp,c7_Datatype_Desc > from Test_Boundary where c2_bigint=c2_bigint -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-596) unable to crete filter resolver tree exception when using equals expression with carbon input format api
[ https://issues.apache.org/jira/browse/CARBONDATA-596?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-596: --- Environment: Spark 2.1 Spark 1.6 > unable to crete filter resolver tree exception when using equals expression > with carbon input format api > > > Key: CARBONDATA-596 > URL: https://issues.apache.org/jira/browse/CARBONDATA-596 > Project: CarbonData > Issue Type: Bug > Components: hadoop-integration >Affects Versions: 1.0.0-incubating > Environment: Spark 2.1 Spark 1.6 >Reporter: anubhav tarar >Assignee: Manish Gupta >Priority: Minor > > unable to crete filter resolver tree exception when using equals expression > Expression expression = new EqualToExpression(new ColumnExpression("c1", > DataType.STRING), > new LiteralExpression("a", DataType.STRING)); > CarbonInputFormat.setFilterPredicates(job.getConfiguration(), > expression); > List splits = carbonInputFormat.getSplits(job); > it throws java.io.IOException: Error while resolving filter expression -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-590) unusual behaviour of using carbonthrift server with spark 2.0
[ https://issues.apache.org/jira/browse/CARBONDATA-590?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-590: --- Environment: Spark 2.0 > unusual behaviour of using carbonthrift server with spark 2.0 > - > > Key: CARBONDATA-590 > URL: https://issues.apache.org/jira/browse/CARBONDATA-590 > Project: CarbonData > Issue Type: Bug > Components: build >Affects Versions: 1.0.0-incubating > Environment: Spark 2.0 >Reporter: anubhav tarar > > have a look at these logs > started thrift server > ./bin/spark-submit --conf spark.sql.hive.thriftServer.singleSession=true > --class org.apache.carbondata.spark.thriftserver.CarbonThriftServer > $SPARK_HOME/carbonlib/carbondata_2.11-1.0.0-incubating-SNAPSHOT-shade-hadoop2.2.0.jar > hdfs://localhost:54310/opt/carbonStore > logs > INFO 04-01 12:46:55,421 - Service:ThriftBinaryCLIService is started. > INFO 04-01 12:46:55,421 - Service:HiveServer2 is started. > started beeline > connected to thrift server > !connect jdbc:hive2://localhost:1 > executed the query > CREATE TABLE Bug212(int string)USING org.apache.spark.sql.CarbonSource > OPTIONS("bucketnumber"="1", "bucketcolumns"="String","tableName"="t100"); > Error: > org.apache.carbondata.spark.exception.MalformedCarbonCommandException: Table > default.t 100 can not be created without key columns. Please use > DICTIONARY_INCLUDE or DICTIONARY_EXCLUDE to set at least one key column if > all specified columns are numeric types (state=,code=0) > 2 minutes later > CREATE TABLE Bug211(int int)USING org.apache.spark.sql.CarbonSource > OPTIONS("bucketnumber"="1", "bucketcolumns"="String","tableName"="t 100"); > +-+--+ > | Result | > +-+--+ > +-+--+ > No rows selected (0.212 seconds) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-589) carbon spark shell is not working with spark 2.0
[ https://issues.apache.org/jira/browse/CARBONDATA-589?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-589: --- Environment: Spark 2.0 > carbon spark shell is not working with spark 2.0 > > > Key: CARBONDATA-589 > URL: https://issues.apache.org/jira/browse/CARBONDATA-589 > Project: CarbonData > Issue Type: Bug > Components: build >Affects Versions: 1.0.0-incubating > Environment: Spark 2.0 >Reporter: anubhav tarar >Priority: Minor > > carbon shell is not working with spark 2.0 version > here are the logs > ./carknoldus@knoldus:~/Desktop/open source/incubator-carbondata/bin$ > ./carbon-spark-shell > java.lang.ClassNotFoundException: org.apache.spark.repl.carbon.Main > at java.net.URLClassLoader.findClass(URLClassLoader.java:381) > at java.lang.ClassLoader.loadClass(ClassLoader.java:424) > at java.lang.ClassLoader.loadClass(ClassLoader.java:357) > at java.lang.Class.forName0(Native Method) > at java.lang.Class.forName(Class.java:348) > at org.apache.spark.util.Utils$.classForName(Utils.scala:225) > at > org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:686) > at > org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:185) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:210) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:124) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-615) Update query store wrong value for Date data type
[ https://issues.apache.org/jira/browse/CARBONDATA-615?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-615: --- Environment: Spark 1.6 Spark 2.1 (was: Spark 1.6) > Update query store wrong value for Date data type > - > > Key: CARBONDATA-615 > URL: https://issues.apache.org/jira/browse/CARBONDATA-615 > Project: CarbonData > Issue Type: Bug > Components: data-query >Affects Versions: 1.0.0-incubating > Environment: Spark 1.6 Spark 2.1 >Reporter: Anurag Srivastava >Assignee: ravikiran >Priority: Minor > Attachments: 2000_UniqData.csv, update_dob.png > > > I am trying to update DOB column with Date Data Type. It is storing a day > before date which I have mentioned for updating in DOB column. > *Create Table :* CREATE TABLE uniqdata (CUST_ID int,CUST_NAME > char(30),ACTIVE_EMUI_VERSION string, DOB Date, DOJ Date, BIGINT_COLUMN1 > bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 > decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double, INTEGER_COLUMN1 > int) STORED BY 'org.apache.carbondata.format'; > *Load Data :* LOAD DATA INPATH 'hdfs://localhost:54310/2000_UniqData.csv' > into table uniqdata OPTIONS ('DELIMITER'=',' > ,'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1','SINGLE_PASS'='true'); > *Update Query :* update uniqdata set (dob)=(to_date('2016-12-01')) where > cust_name = 'CUST_NAME_01999'; > *Expected Result :* It should update DOB column with *2016-12-01*. > *Actual Result :* It is updating DOB column with *2016-11-30*. > !https://issues.apache.org/jira/secure/attachment/12846515/update_dob.png! -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-615) Update query store wrong value for Date data type
[ https://issues.apache.org/jira/browse/CARBONDATA-615?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-615: --- Environment: Spark 1.6 > Update query store wrong value for Date data type > - > > Key: CARBONDATA-615 > URL: https://issues.apache.org/jira/browse/CARBONDATA-615 > Project: CarbonData > Issue Type: Bug > Components: data-query >Affects Versions: 1.0.0-incubating > Environment: Spark 1.6 >Reporter: Anurag Srivastava >Assignee: ravikiran >Priority: Minor > Attachments: 2000_UniqData.csv, update_dob.png > > > I am trying to update DOB column with Date Data Type. It is storing a day > before date which I have mentioned for updating in DOB column. > *Create Table :* CREATE TABLE uniqdata (CUST_ID int,CUST_NAME > char(30),ACTIVE_EMUI_VERSION string, DOB Date, DOJ Date, BIGINT_COLUMN1 > bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 > decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double, INTEGER_COLUMN1 > int) STORED BY 'org.apache.carbondata.format'; > *Load Data :* LOAD DATA INPATH 'hdfs://localhost:54310/2000_UniqData.csv' > into table uniqdata OPTIONS ('DELIMITER'=',' > ,'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1','SINGLE_PASS'='true'); > *Update Query :* update uniqdata set (dob)=(to_date('2016-12-01')) where > cust_name = 'CUST_NAME_01999'; > *Expected Result :* It should update DOB column with *2016-12-01*. > *Actual Result :* It is updating DOB column with *2016-11-30*. > !https://issues.apache.org/jira/secure/attachment/12846515/update_dob.png! -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-593) Select command seems to be not working on carbon-spark-shell . It throws a runtime error on select query after show method is invoked
[ https://issues.apache.org/jira/browse/CARBONDATA-593?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-593: --- Environment: Spark 1.6 > Select command seems to be not working on carbon-spark-shell . It throws a > runtime error on select query after show method is invoked > - > > Key: CARBONDATA-593 > URL: https://issues.apache.org/jira/browse/CARBONDATA-593 > Project: CarbonData > Issue Type: Bug > Components: sql >Affects Versions: 1.0.0-incubating > Environment: Spark 1.6 >Reporter: DEEPAK MEHRA >Priority: Minor > Attachments: employee.csv > > > Select command seems to be not working on carbon-spark-shell . It throws a > runtime error on select query after show method is invoked. It says > java.io.FileNotFoundException: File does not exist: > /home/hduser/software/spark-1.6.2-bin-hadoop2.6/carbon.store/mydb/demo/Fact/Part0/Segment_0 > Query Executed : > scala> cc.sql("create table demo(id int,name string,age int,city string) > stored by 'carbondata'").show > scala> cc.sql("LOAD DATA inpath > 'hdfs://hadoop-master:54311/data/employee.csv' INTO table demo"); > cc.sql("select * from demo").show > Result :It's throwing java.lang.RuntimeException : File Not Found > scala> cc.sql("select * from mydb.demo").show > INFO 04-01 16:29:22,157 - main Query [SELECT * FROM MYDB.DEMO] > INFO 04-01 16:29:22,160 - Parsing command: select * from mydb.demo > INFO 04-01 16:29:22,161 - Parse Completed > INFO 04-01 16:29:22,162 - Parsing command: select * from mydb.demo > INFO 04-01 16:29:22,163 - Parse Completed > INFO 04-01 16:29:22,163 - 0: get_table : db=mydb tbl=demo > INFO 04-01 16:29:22,163 - ugi=hduser ip=unknown-ip-addr cmd=get_table : > db=mydb tbl=demo > INFO 04-01 16:29:22,195 - main Starting to optimize plan > java.io.FileNotFoundException: File does not exist: > /home/hduser/software/spark-1.6.2-bin-hadoop2.6/carbon.store/mydb/demo/Fact/Part0/Segment_0 > at > org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1110) > at > org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1102) > at > org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) > at > org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1102) > at org.apache.hadoop.fs.FileSystem.resolvePath(FileSystem.java:747) > at > org.apache.hadoop.hdfs.DistributedFileSystem$15.(DistributedFileSystem.java:726) > at > org.apache.hadoop.hdfs.DistributedFileSystem.listLocatedStatus(DistributedFileSystem.java:717) > at > org.apache.hadoop.fs.FileSystem.listLocatedStatus(FileSystem.java:1780) > at > org.apache.carbondata.hadoop.CarbonInputFormat.getFileStatusOfSegments(CarbonInputFormat.java:559) > at > org.apache.carbondata.hadoop.CarbonInputFormat.listStatus(CarbonInputFormat.java:519) > at > org.apache.hadoop.mapreduce.lib.input.FileInputFormat.getSplits(FileInputFormat.java:340) > at > org.apache.carbondata.hadoop.CarbonInputFormat.getSplitsInternal(CarbonInputFormat.java:251) > at > org.apache.carbondata.hadoop.CarbonInputFormat.getTableBlockInfo(CarbonInputFormat.java:372) > at > org.apache.carbondata.hadoop.CarbonInputFormat.getSegmentAbstractIndexs(CarbonInputFormat.java:402) > at > org.apache.carbondata.hadoop.CarbonInputFormat.getDataBlocksOfSegment(CarbonInputFormat.java:325) > at > org.apache.carbondata.hadoop.CarbonInputFormat.getSplits(CarbonInputFormat.java:288) > at > org.apache.carbondata.hadoop.CarbonInputFormat.getSplits(CarbonInputFormat.java:237) > at > org.apache.carbondata.spark.rdd.CarbonScanRDD.getPartitions(CarbonScanRDD.scala:82) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237) > at scala.Option.getOrElse(Option.scala:120) > at org.apache.spark.rdd.RDD.partitions(RDD.scala:237) > at > org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237) > at scala.Option.getOrElse(Option.scala:120) > at org.apache.spark.rdd.RDD.partitions(RDD.scala:237) > at > org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:239) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:237) > at
[jira] [Updated] (CARBONDATA-604) Use Kettle wrong option in Thrift server give wrong error message
[ https://issues.apache.org/jira/browse/CARBONDATA-604?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-604: --- Environment: Spark 1.6 > Use Kettle wrong option in Thrift server give wrong error message > - > > Key: CARBONDATA-604 > URL: https://issues.apache.org/jira/browse/CARBONDATA-604 > Project: CarbonData > Issue Type: Bug > Environment: Spark 1.6 >Reporter: Harmeet Singh >Priority: Trivial > > I am using Thrift Server with following configuration: > -Duse.kettle=default > Here kettle value is wrong and when I run the thrift server, the thrift > server is running successfully. After that, I am going to start beeline. In > beeline, I am executing load command without any kettle option as below: > LOAD DATA INPATH 'hdfs://localhost:54310/csvs/3000_UniqData.csv' into table > uniqdata OPTIONS('DELIMITER'=',' , > 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1'); > After executing above command, I am getting Error: > java.lang.IllegalArgumentException: For input string: "default" > (state=,code=0) > This error, seems to be wrong, because I am not using any default value in > load query, and default is used with thrift. > So I have following Suggestions: > 1. Change the error message, while load query is executing. > 2. While starting thrift server with wrong value, thrift give us an error and > not start. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-619) Compaction API for Spark 2.1 : Issue in compaction type
[ https://issues.apache.org/jira/browse/CARBONDATA-619?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-619: --- Labels: 2.1.0 Spark (was: ) > Compaction API for Spark 2.1 : Issue in compaction type > --- > > Key: CARBONDATA-619 > URL: https://issues.apache.org/jira/browse/CARBONDATA-619 > Project: CarbonData > Issue Type: Bug > Components: data-query >Affects Versions: 1.0.0-incubating >Reporter: DEEPAK MEHRA >Assignee: ravikiran >Priority: Minor > Attachments: compaction_result.png, employee.csv > > > Compaction API for Spark 2.1 : Compaction is successful on the segment ids > irrespective of the compaction types which are( Minor and Major) i.e The > compaction was successful for type other than Minor/Major > I created a table with name carbon_table and made 5 loads into the table with > the threshold of (4,3) > cc.sql("create table carbon_table(id int,name string) stored by 'carbondata'") > cc.sql("LOAD DATA LOCAL INPATH 'hdfs://hadoop-master:54311/data/employee.csv' > INTO TABLE carbon_table options('FILEHEADER'='id,name')") > For instance > Compaction.compaction(spark,"default","carbon_table","xyz") > here > default is the database name > carbon_table is the table name > xyz is the compaction type > I invoked the above method to compact the segments and it was executed > successfully and compacted the segment ids. On show segment It produced the > result with success for compaction . I have attached the screen shot for the > same. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-619) Compaction API for Spark 2.1 : Issue in compaction type
[ https://issues.apache.org/jira/browse/CARBONDATA-619?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-619: --- Labels: (was: 2.1.0 Spark) > Compaction API for Spark 2.1 : Issue in compaction type > --- > > Key: CARBONDATA-619 > URL: https://issues.apache.org/jira/browse/CARBONDATA-619 > Project: CarbonData > Issue Type: Bug > Components: data-query >Affects Versions: 1.0.0-incubating >Reporter: DEEPAK MEHRA >Assignee: ravikiran >Priority: Minor > Attachments: compaction_result.png, employee.csv > > > Compaction API for Spark 2.1 : Compaction is successful on the segment ids > irrespective of the compaction types which are( Minor and Major) i.e The > compaction was successful for type other than Minor/Major > I created a table with name carbon_table and made 5 loads into the table with > the threshold of (4,3) > cc.sql("create table carbon_table(id int,name string) stored by 'carbondata'") > cc.sql("LOAD DATA LOCAL INPATH 'hdfs://hadoop-master:54311/data/employee.csv' > INTO TABLE carbon_table options('FILEHEADER'='id,name')") > For instance > Compaction.compaction(spark,"default","carbon_table","xyz") > here > default is the database name > carbon_table is the table name > xyz is the compaction type > I invoked the above method to compact the segments and it was executed > successfully and compacted the segment ids. On show segment It produced the > result with success for compaction . I have attached the screen shot for the > same. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (CARBONDATA-619) Compaction API for Spark 2.1 : Issue in compaction type
[ https://issues.apache.org/jira/browse/CARBONDATA-619?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Divya Gupta updated CARBONDATA-619: --- Environment: Spark 2.1 > Compaction API for Spark 2.1 : Issue in compaction type > --- > > Key: CARBONDATA-619 > URL: https://issues.apache.org/jira/browse/CARBONDATA-619 > Project: CarbonData > Issue Type: Bug > Components: data-query >Affects Versions: 1.0.0-incubating > Environment: Spark 2.1 >Reporter: DEEPAK MEHRA >Assignee: ravikiran >Priority: Minor > Attachments: compaction_result.png, employee.csv > > > Compaction API for Spark 2.1 : Compaction is successful on the segment ids > irrespective of the compaction types which are( Minor and Major) i.e The > compaction was successful for type other than Minor/Major > I created a table with name carbon_table and made 5 loads into the table with > the threshold of (4,3) > cc.sql("create table carbon_table(id int,name string) stored by 'carbondata'") > cc.sql("LOAD DATA LOCAL INPATH 'hdfs://hadoop-master:54311/data/employee.csv' > INTO TABLE carbon_table options('FILEHEADER'='id,name')") > For instance > Compaction.compaction(spark,"default","carbon_table","xyz") > here > default is the database name > carbon_table is the table name > xyz is the compaction type > I invoked the above method to compact the segments and it was executed > successfully and compacted the segment ids. On show segment It produced the > result with success for compaction . I have attached the screen shot for the > same. -- This message was sent by Atlassian JIRA (v6.3.4#6332)