[GitHub] carbondata issue #1605: [WIP] added support to compact segments in pre-agg t...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1605 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1667/ ---
[GitHub] carbondata issue #1559: [CARBONDATA-1805][Dictionary] Optimize pruning for d...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1559 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1666/ ---
[GitHub] carbondata issue #1593: [CARBONDATA-1837] Reusing origin row to reduce memor...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1593 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1665/ ---
[jira] [Created] (CARBONDATA-1846) Incorrect output on presto CLI while executing IN operator with multiple load
Vandana Yadav created CARBONDATA-1846: - Summary: Incorrect output on presto CLI while executing IN operator with multiple load Key: CARBONDATA-1846 URL: https://issues.apache.org/jira/browse/CARBONDATA-1846 Project: CarbonData Issue Type: Bug Components: presto-integration Affects Versions: 1.3.0 Environment: spark 2.1 Reporter: Vandana Yadav Priority: Minor Attachments: 2000_UniqData.csv, 3000_UniqData.csv, 4000_UniqData.csv, 5000_UniqData.csv, 6000_UniqData.csv, 7000_UniqData.csv Incorrect output on presto CLI while executing IN operator with multiple load Steps to reproduce: On beeline: 1) Create a table: CREATE TABLE uniqdata_time (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB") 2)Load Data with multiple segments in it: a) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/2000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') b) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/3000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') c) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/4000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') d) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/5000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') e) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/6000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') f) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/7000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') 3) Execute Query: select DOB from UNIQDATA_time where DOB in (cast('2015-10-04 01:00:03' as timestamp),cast('2015-10-07' as timestamp),cast('2015-10-07 01:00:03' as timestamp)); Output on Beeline: ++--+ | DOB | ++--+ | 2015-10-04 01:00:03.0 | | 2015-10-07 01:00:03.0 | ++--+ 2 rows selected (0.351 seconds) 4) Output on Presto CLI: a) Expected Output: similar to the result on Beeline. b) Actual Output : DOB - (0 rows) -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Created] (CARBONDATA-1845) Incorrect output on presto CLI while executing OR operator with multiple load
Vandana Yadav created CARBONDATA-1845: - Summary: Incorrect output on presto CLI while executing OR operator with multiple load Key: CARBONDATA-1845 URL: https://issues.apache.org/jira/browse/CARBONDATA-1845 Project: CarbonData Issue Type: Bug Components: presto-integration Affects Versions: 1.3.0 Environment: spark 2.1 Reporter: Vandana Yadav Priority: Minor Attachments: 2000_UniqData.csv, 3000_UniqData.csv, 4000_UniqData.csv, 5000_UniqData.csv, 6000_UniqData.csv, 7000_UniqData;delm.csv Incorrect output on presto CLI while executing OR operator with multiple load Steps to reproduce: On beeline: 1) Create a table: CREATE TABLE uniqdata_time (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB") 2)Load Data with multiple segments in it: a) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/2000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') b) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/3000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') c) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/4000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') d) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/5000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') e) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/6000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') f) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/7000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') 3) Execute Query: select * from uniqdata_time where DOB = cast('2015-10-04 01:00:03' as timestamp) or DOB = cast('2015-10-07 01:00:03' as timestamp) Output on beeline: --+-+-+--+---+--+--+ | CUST_ID |CUST_NAME |ACTIVE_EMUI_VERSION | DOB | DOJ | BIGINT_COLUMN1 | BIGINT_COLUMN2 | DECIMAL_COLUMN1 | DECIMAL_COLUMN2 |Double_COLUMN1| Double_COLUMN2 | INTEGER_COLUMN1 | +--+--++++-+-+-+-+--+---+--+--+ | 25712| CUST_NAME_16712 | ACTIVE_EMUI_VERSION_16712 | 2015-10-04 01:00:03.0 | 2015-10-04 02:00:03.0 | 123372053566| -223372020142 | 12345695613.123400 | 22345695613.123400 | 1.12345674897976E10 | -1.12345674897976E10 | 16713| | 25715| CUST_NAME_16715 | ACTIVE_EMUI_VERSION_16715 | 2015-10-07 01:00:03.0 | 2015-10-07 02:00:03.0 | 123372053569| -223372020139 | 12345695616.123400 | 22345695616.123400 | 1.12345674897976E10 | -1.12345674897976E10 | 16716 4) Output On Presto CLI: a)Expected Output on Presto: Similar to the output on beeline b) Actual output on Presto: cust_id | cust_name | active_emui_version | dob | doj | bigint_column1 | bigint_column2 | decimal_column1 | decimal_column2 | double_column1 |
[GitHub] carbondata issue #1508: [CARBONDATA-1738] [PreAgg] Block direct insert/load ...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1508 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2055/ ---
[GitHub] carbondata issue #1508: [CARBONDATA-1738] [PreAgg] Block direct insert/load ...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1508 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1664/ ---
[GitHub] carbondata issue #1605: [WIP] added support to compact segments in pre-agg t...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1605 SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2054/ ---
[GitHub] carbondata issue #1559: [CARBONDATA-1805][Dictionary] Optimize pruning for d...
Github user xuchuanyin commented on the issue: https://github.com/apache/carbondata/pull/1559 retest this please ---
[GitHub] carbondata issue #1593: [CARBONDATA-1837] Reusing origin row to reduce memor...
Github user xuchuanyin commented on the issue: https://github.com/apache/carbondata/pull/1593 retest this please ---
[jira] [Resolved] (CARBONDATA-1844) Support specify tablePath when creating table
[ https://issues.apache.org/jira/browse/CARBONDATA-1844?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ravindra Pesala resolved CARBONDATA-1844. - Resolution: Fixed > Support specify tablePath when creating table > - > > Key: CARBONDATA-1844 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1844 > Project: CarbonData > Issue Type: New Feature >Reporter: Jacky Li >Assignee: Jacky Li > Fix For: 1.3.0 > > Time Spent: 6.5h > Remaining Estimate: 0h > > User should be able to specify table path when creating table -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata pull request #1603: [CARBONDATA-1844] Add tablePath support when ...
Github user asfgit closed the pull request at: https://github.com/apache/carbondata/pull/1603 ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1603 LGTM ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1603 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1663/ ---
[GitHub] carbondata issue #1508: [CARBONDATA-1738] [PreAgg] Block direct insert/load ...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1508 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2053/ ---
[GitHub] carbondata pull request #1583: [CARBONDATA-1822][Spark-Integration] Support ...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1583#discussion_r154566404 --- Diff: integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/RefreshCarbonTableCommand.scala --- @@ -0,0 +1,216 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + *http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.spark.sql.execution.command.management + +import java.util + +import scala.collection.JavaConverters._ + +import org.apache.spark.sql._ +import org.apache.spark.sql.execution.command.MetadataCommand +import org.apache.spark.sql.execution.command.table.CarbonCreateTableCommand +import org.apache.spark.sql.util.CarbonException + +import org.apache.carbondata.common.logging.{LogService, LogServiceFactory} +import org.apache.carbondata.core.constants.CarbonCommonConstants +import org.apache.carbondata.core.datastore.impl.FileFactory +import org.apache.carbondata.core.metadata.{AbsoluteTableIdentifier, CarbonTableIdentifier} +import org.apache.carbondata.core.metadata.schema.table.{DataMapSchema, TableInfo} +import org.apache.carbondata.core.metadata.schema.table.column.ColumnSchema +import org.apache.carbondata.core.util.CarbonProperties +import org.apache.carbondata.core.util.path.CarbonStorePath +import org.apache.carbondata.events.{OperationContext, OperationListenerBus, RefreshTablePostExecutionEvent, RefreshTablePreExecutionEvent} + +/** + * Command to register carbon table from existing carbon table data + */ +case class RefreshCarbonTableCommand( +dbName: Option[String], +tableName: String) + extends MetadataCommand { + val LOGGER: LogService = +LogServiceFactory.getLogService(this.getClass.getName) + + override def processMetadata(sparkSession: SparkSession): Seq[Row] = { +val metaStore = CarbonEnv.getInstance(sparkSession).carbonMetastore +val databaseName = GetDB.getDatabaseName(dbName, sparkSession) +val databaseLocation = GetDB.getDatabaseLocation(databaseName, sparkSession, + CarbonProperties.getStorePath) +// Steps +// 1. get table path +// 2. perform the below steps +// 2.1 check if the table already register with hive then ignore and continue with the next +// schema +// 2.2 register the table with the hive check if the table being registered has aggregate table +// then do the below steps +// 2.2.1 validate that all the aggregate tables are copied at the store location. +// 2.2.2 Register the aggregate tables +val tablePath = databaseLocation + CarbonCommonConstants.FILE_SEPARATOR + tableName +val absoluteTableIdentifier = AbsoluteTableIdentifier.from(tablePath, databaseName, tableName) +// 2.1 check if the table already register with hive then ignore and continue with the next +// schema +if (!sparkSession.sessionState.catalog.listTables(databaseName) + .exists(_.table.equalsIgnoreCase(tableName))) { + val carbonTablePath = CarbonStorePath.getCarbonTablePath(absoluteTableIdentifier) + // check the existence of the schema file to know its a carbon table + val schemaFilePath = carbonTablePath.getSchemaFilePath + // if schema file does not exist then the table will either non carbon table or stale + // carbon table + if (FileFactory.isFileExist(schemaFilePath, FileFactory.getFileType(schemaFilePath))) { +// read TableInfo +val tableInfo = metaStore.getWrapperTableInfo(absoluteTableIdentifier)(sparkSession) +// 2.2 register the table with the hive check if the table being registered has +// aggregate table then do the below steps +// 2.2.1 validate that all the aggregate tables are copied at the store location. +val dataMapSchemaList = tableInfo.getDataMapSchemaList +if (null != dataMapSchemaList && dataMapSchemaList.size() != 0) { + // validate all the
[GitHub] carbondata issue #1601: [CARBONDATA-1787] Validation for table properties in...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1601 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1662/ ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1603 retest this please ---
[GitHub] carbondata issue #1601: [CARBONDATA-1787] Validation for table properties in...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1601 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2052/ ---
[GitHub] carbondata issue #1605: [WIP] added support to compact segments in pre-agg t...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1605 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1661/ ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1603 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1660/ ---
[jira] [Resolved] (CARBONDATA-1064) Fix NullPointerException in Not InExpression
[ https://issues.apache.org/jira/browse/CARBONDATA-1064?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kunal Kapoor resolved CARBONDATA-1064. -- Resolution: Fixed > Fix NullPointerException in Not InExpression > > > Key: CARBONDATA-1064 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1064 > Project: CarbonData > Issue Type: Bug >Reporter: Kunal Kapoor >Assignee: Kunal Kapoor >Priority: Minor > Time Spent: 2h > Remaining Estimate: 0h > > select * from expression_test where id not in (1,2,'', NULL, ' ') > In the above query Right expression value is null in NotInExpression which > cased throws NullPointerException -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata issue #1605: [WIP] added support to compact segments in pre-agg t...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1605 SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2051/ ---
[GitHub] carbondata issue #1605: [WIP] added support to compact segments in pre-agg t...
Github user kunal642 commented on the issue: https://github.com/apache/carbondata/pull/1605 @jackylk Can you please start the first level review. ---
[GitHub] carbondata pull request #1605: [WIP] added support to compact segments in pr...
GitHub user kunal642 opened a pull request: https://github.com/apache/carbondata/pull/1605 [WIP] added support to compact segments in pre-agg table Be sure to do all of the following checklist to help us incorporate your contribution quickly and easily: - [ ] Any interfaces changed? - [ ] Any backward compatibility impacted? - [ ] Document update required? - [ ] Testing done Please provide details on - Whether new unit test cases have been added or why no new tests are required? - How it is tested? Please attach test report. - Is it a performance related change? Please attach the performance test report. - Any additional information to help reviewers in testing this change. - [ ] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA. You can merge this pull request into a Git repository by running: $ git pull https://github.com/kunal642/carbondata preagg_compaction_support Alternatively you can review and apply these changes as the patch at: https://github.com/apache/carbondata/pull/1605.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #1605 commit cc58e9d2d26a99465b2daa3af1a24777581718cf Author: kunal642Date: 2017-11-22T14:03:37Z added support to compact segments in pre-agg table ---
[jira] [Closed] (CARBONDATA-1821) Incorrect headings in documentation
[ https://issues.apache.org/jira/browse/CARBONDATA-1821?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vandana Yadav closed CARBONDATA-1821. - Resolved > Incorrect headings in documentation > --- > > Key: CARBONDATA-1821 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1821 > Project: CarbonData > Issue Type: Improvement > Components: docs >Affects Versions: 1.3.0 >Reporter: Vandana Yadav >Assignee: Vandana Yadav >Priority: Trivial > Fix For: 1.3.0 > > Attachments: headings.png > > Time Spent: 2.5h > Remaining Estimate: 0h > > Incorrect headings in documentation -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1603 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2050/ ---
[jira] [Resolved] (CARBONDATA-1733) While load is in progress, Show segments is throwing NPE
[ https://issues.apache.org/jira/browse/CARBONDATA-1733?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] dhatchayani resolved CARBONDATA-1733. - Resolution: Fixed Fix Version/s: 1.3.0 > While load is in progress, Show segments is throwing NPE > > > Key: CARBONDATA-1733 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1733 > Project: CarbonData > Issue Type: Bug >Reporter: dhatchayani >Assignee: dhatchayani > Fix For: 1.3.0 > > Time Spent: 1.5h > Remaining Estimate: 0h > > *Scenario:* > Concurrent execution of load and show segments. > Show segments is thorwing NullPointerException -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Closed] (CARBONDATA-1734) Ignore empty line while reading CSV
[ https://issues.apache.org/jira/browse/CARBONDATA-1734?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] dhatchayani closed CARBONDATA-1734. --- Resolution: Fixed Fix Version/s: 1.3.0 > Ignore empty line while reading CSV > --- > > Key: CARBONDATA-1734 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1734 > Project: CarbonData > Issue Type: Improvement >Reporter: dhatchayani >Assignee: dhatchayani >Priority: Minor > Fix For: 1.3.0 > > Time Spent: 5h > Remaining Estimate: 0h > > Ignore empty line while reading CSV file in LOAD -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata pull request #1604: [CARBONDATA-1843] Block CTAS and external tab...
Github user asfgit closed the pull request at: https://github.com/apache/carbondata/pull/1604 ---
[GitHub] carbondata issue #1604: [CARBONDATA-1843] Block CTAS and external table feat...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1604 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1659/ ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1603 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1658/ ---
[GitHub] carbondata issue #1604: [CARBONDATA-1843] Block CTAS and external table feat...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1604 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2049/ ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1603 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2048/ ---
[GitHub] carbondata pull request #1603: [CARBONDATA-1844] Add tablePath support when ...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1603#discussion_r154552813 --- Diff: integration/spark-common-test/src/test/scala/org/apache/carbondata/spark/testsuite/dataload/TestLoadDataFrame.scala --- @@ -222,6 +223,26 @@ test("test the boolean data type"){ ) } + test("test datasource table with specified table path") { --- End diff -- This is new testcase for datasource table ---
[GitHub] carbondata pull request #1603: [CARBONDATA-1844] Add tablePath support when ...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1603#discussion_r154552516 --- Diff: integration/spark-common-test/src/test/scala/org/apache/carbondata/spark/testsuite/dataload/TestLoadDataWithHiveSyntaxDefaultFormat.scala --- @@ -687,6 +688,47 @@ class TestLoadDataWithHiveSyntaxDefaultFormat extends QueryTest with BeforeAndAf checkAnswer(sql("select salary from double_test limit 1"),Row(7.756787654567891E23)) } + test("test table with specified table path") { +val path = "./source" +sql("drop table if exists table_path_test") +sql( + "CREATE table table_path_test (empno string, salary double) STORED BY 'carbondata' " + + s"LOCATION '$path'" --- End diff -- changed to use LOCATION syntax ---
[jira] [Closed] (CARBONDATA-1842) Fix 'wrong argument number' error of class Cast for Spark 2.2 when pattern matching
[ https://issues.apache.org/jira/browse/CARBONDATA-1842?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zhichao Zhang closed CARBONDATA-1842. -- Resolution: Duplicate resolved by PR #1508 > Fix 'wrong argument number' error of class Cast for Spark 2.2 when pattern > matching > --- > > Key: CARBONDATA-1842 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1842 > Project: CarbonData > Issue Type: Bug > Components: spark-integration >Affects Versions: 1.3.0 >Reporter: Zhichao Zhang >Assignee: Zhichao Zhang >Priority: Minor > Fix For: 1.3.0 > > Time Spent: 1.5h > Remaining Estimate: 0h > > The constructor of class org.apache.spark.sql.catalyst.expressions.Cast has > two arguments in Spark 2.1, but it needs three arguments in Spark 2.2. -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata pull request #1603: [CARBONDATA-1844] Add tablePath support when ...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1603#discussion_r154551403 --- Diff: core/src/main/java/org/apache/carbondata/core/metadata/schema/table/TableInfo.java --- @@ -34,6 +34,8 @@ import org.apache.carbondata.core.metadata.schema.table.column.ColumnSchema; import org.apache.carbondata.core.metadata.schema.table.column.ParentColumnTableRelation; +import org.apache.http.annotation.Obsolete; --- End diff -- fixed ---
[GitHub] carbondata pull request #1603: [CARBONDATA-1844] Add tablePath support when ...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1603#discussion_r154551395 --- Diff: core/src/main/java/org/apache/carbondata/core/metadata/schema/table/TableInfo.java --- @@ -68,17 +70,10 @@ */ private long lastUpdatedTime; - /** - * metadata file path (check if it is really required ) - */ - private String metaDataFilepath; - - /** - * store location - */ + @Obsolete --- End diff -- fixed, it is not obsoleted ---
[GitHub] carbondata issue #1569: [CARBONDATA-1752] There are some scalastyle error sh...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1569 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1657/ ---
[GitHub] carbondata issue #1569: [CARBONDATA-1752] There are some scalastyle error sh...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1569 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2047/ ---
[GitHub] carbondata pull request #1604: [CARBONDATA-1843] Block CTAS and external tab...
Github user QiangCai commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1604#discussion_r154548820 --- Diff: integration/spark2/src/main/scala/org/apache/spark/sql/CarbonSession.scala --- @@ -170,8 +181,15 @@ object CarbonSession { } options.foreach { case (k, v) => session.sessionState.conf.setConfString(k, v) } SparkSession.setDefaultSession(session) -CommonUtil.cleanInProgressSegments( - carbonProperties.getProperty(CarbonCommonConstants.STORE_LOCATION), sparkContext) +try { + CommonUtil.cleanInProgressSegments( + carbonProperties.getProperty(CarbonCommonConstants.STORE_LOCATION), sparkContext) +} catch { + case e: Throwable => +// catch all exceptions to avoid CarbonSession initialization failure + LogServiceFactory.getLogService(this.getClass.getCanonicalName) + .error(e, "Failed to clean in progress segments") --- End diff -- indentation is incorrect ---
[GitHub] carbondata pull request #1603: [CARBONDATA-1844] Add tablePath support when ...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1603#discussion_r154530400 --- Diff: integration/spark-common-test/src/test/scala/org/apache/carbondata/spark/testsuite/dataload/TestLoadDataWithHiveSyntaxDefaultFormat.scala --- @@ -687,6 +688,48 @@ class TestLoadDataWithHiveSyntaxDefaultFormat extends QueryTest with BeforeAndAf checkAnswer(sql("select salary from double_test limit 1"),Row(7.756787654567891E23)) } + test("test table with specified table path") { +val path = "./source" +sql("drop table if exists table_path_test") +sql( + "CREATE table table_path_test (empno string, salary double) STORED BY 'carbondata' " + + s"TBLPROPERTIES('table_path'='$path')" --- End diff -- ok ---
[GitHub] carbondata pull request #1603: [CARBONDATA-1844] Add tablePath support when ...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1603#discussion_r154528674 --- Diff: integration/spark-common-test/src/test/scala/org/apache/carbondata/spark/testsuite/dataload/TestLoadDataWithHiveSyntaxDefaultFormat.scala --- @@ -687,6 +688,48 @@ class TestLoadDataWithHiveSyntaxDefaultFormat extends QueryTest with BeforeAndAf checkAnswer(sql("select salary from double_test limit 1"),Row(7.756787654567891E23)) } + test("test table with specified table path") { +val path = "./source" +sql("drop table if exists table_path_test") +sql( + "CREATE table table_path_test (empno string, salary double) STORED BY 'carbondata' " + + s"TBLPROPERTIES('table_path'='$path')" --- End diff -- We should not take path from tableproperties . Better use standard way of specifying the path. For both spark datasource and hive ddl uses [LOCATION path] to take tablePath. Better use the same. ---
[GitHub] carbondata pull request #1603: [CARBONDATA-1844] Add tablePath support when ...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1603#discussion_r154528596 --- Diff: core/src/main/java/org/apache/carbondata/core/constants/CarbonCommonConstants.java --- @@ -854,6 +854,7 @@ public static final String DICTIONARY_EXCLUDE = "dictionary_exclude"; public static final String DICTIONARY_INCLUDE = "dictionary_include"; public static final String SORT_COLUMNS = "sort_columns"; + public static final String TABLE_PATH = "table_path"; --- End diff -- I think this will be confusion if we use 2 different words `tablePath` and `table_path`, better use only one `tablePath`. ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1603 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1656/ ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1603 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2046/ ---
[GitHub] carbondata issue #1604: [CARBONDATA-1843] Block CTAS and external table feat...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1604 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2045/ ---
[GitHub] carbondata pull request #1603: [CARBONDATA-1844] Add tablePath support when ...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1603#discussion_r154527728 --- Diff: core/src/main/java/org/apache/carbondata/core/metadata/schema/table/TableInfo.java --- @@ -34,6 +34,8 @@ import org.apache.carbondata.core.metadata.schema.table.column.ColumnSchema; import org.apache.carbondata.core.metadata.schema.table.column.ParentColumnTableRelation; +import org.apache.http.annotation.Obsolete; --- End diff -- Better use java.lang.Deprecated ---
[GitHub] carbondata pull request #1603: [CARBONDATA-1844] Add tablePath support when ...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1603#discussion_r154527695 --- Diff: core/src/main/java/org/apache/carbondata/core/metadata/schema/table/TableInfo.java --- @@ -68,17 +70,10 @@ */ private long lastUpdatedTime; - /** - * metadata file path (check if it is really required ) - */ - private String metaDataFilepath; - - /** - * store location - */ + @Obsolete --- End diff -- Why it is obsolete ? is there any other way user can get tablePath? ---
[GitHub] carbondata issue #1604: [CARBONDATA-1843] Block CTAS and external table feat...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1604 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2044/ ---
[GitHub] carbondata issue #1604: [CARBONDATA-1843] Block CTAS and external table feat...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1604 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1655/ ---
[GitHub] carbondata pull request #1603: [CARBONDATA-1844] Add tablePath support when ...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1603#discussion_r154527161 --- Diff: integration/spark-common-test/src/test/scala/org/apache/carbondata/spark/testsuite/dataload/TestLoadDataWithHiveSyntaxDefaultFormat.scala --- @@ -687,6 +688,48 @@ class TestLoadDataWithHiveSyntaxDefaultFormat extends QueryTest with BeforeAndAf checkAnswer(sql("select salary from double_test limit 1"),Row(7.756787654567891E23)) } + test("test table with specified table path") { --- End diff -- these are new testcase ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1603 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1654/ ---
[GitHub] carbondata pull request #1604: [CARBONDATA-1843] Block CTAS and external tab...
Github user jackylk commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1604#discussion_r154526536 --- Diff: integration/spark2/src/main/scala/org/apache/spark/sql/CarbonSession.scala --- @@ -58,7 +59,12 @@ class CarbonSession(@transient val sc: SparkContext, */ @transient override lazy val sharedState: SharedState = { -existingSharedState.getOrElse(new SharedState(sparkContext)) +if (existingSharedState.isDefined) { --- End diff -- ok, fixed ---
[GitHub] carbondata pull request #1604: [CARBONDATA-1843] Block CTAS and external tab...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1604#discussion_r154526332 --- Diff: integration/spark2/src/main/scala/org/apache/spark/sql/CarbonSession.scala --- @@ -58,7 +59,12 @@ class CarbonSession(@transient val sc: SparkContext, */ @transient override lazy val sharedState: SharedState = { -existingSharedState.getOrElse(new SharedState(sparkContext)) +if (existingSharedState.isDefined) { --- End diff -- Better use `match {case}` ---
[GitHub] carbondata pull request #1602: [CARBONDATA-1843] Add configuration to enable...
Github user jackylk closed the pull request at: https://github.com/apache/carbondata/pull/1602 ---
[GitHub] carbondata pull request #1604: Basic2
GitHub user jackylk opened a pull request: https://github.com/apache/carbondata/pull/1604 Basic2 1. Block 'external' syntax 2. Block 'CTAS' syntax - [X] Any interfaces changed? No - [X] Any backward compatibility impacted? No - [X] Document update required? No - [X] Testing done Added - [X] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA. NA You can merge this pull request into a Git repository by running: $ git pull https://github.com/jackylk/incubator-carbondata basic2 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/carbondata/pull/1604.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #1604 commit 4c1855b158a1f4e5840a3ed6ec8fd8c83c8bbb1f Author: Jacky LiDate: 2017-12-02T14:46:56Z add config commit 0ecb111d933e33b91db5dcaeb9ea4980af3e29eb Author: Jacky Li Date: 2017-12-03T14:56:16Z change block ---
[jira] [Updated] (CARBONDATA-1843) Block CTAS and external table syntax
[ https://issues.apache.org/jira/browse/CARBONDATA-1843?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jacky Li updated CARBONDATA-1843: - Description: 1. Block 'external' syntax 2. Block 'CTAS' syntax was: 1. Add configuration for support dictionary and complex type 2. Block 'external', 'CTAS' syntax 3. Some other minor fix to catch exceptions Summary: Block CTAS and external table syntax (was: Add configuration to enable features to improve usability) > Block CTAS and external table syntax > > > Key: CARBONDATA-1843 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1843 > Project: CarbonData > Issue Type: Improvement >Reporter: Jacky Li >Assignee: Jacky Li >Priority: Minor > Fix For: 1.3.0 > > Time Spent: 1h 20m > Remaining Estimate: 0h > > 1. Block 'external' syntax > 2. Block 'CTAS' syntax -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1603 SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2041/ ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1603 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1653/ ---
[GitHub] carbondata pull request #1602: [CARBONDATA-1843] Add configuration to enable...
Github user ravipesala commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1602#discussion_r154523178 --- Diff: integration/spark2/src/main/scala/org/apache/spark/sql/parser/CarbonSparkSqlParser.scala --- @@ -18,6 +18,7 @@ package org.apache.spark.sql.parser import scala.collection.mutable +import org.antlr.v4.runtime.tree.TerminalNode --- End diff -- No, I mean these type of validations to move to listeners, actually we no need to add these validations to carbon layer. If other projects who want to block these operations can add their listener and just register it. I feel these configurations are not required. ---
[GitHub] carbondata issue #1571: [CARBONDATA-1811] [CTAS] Use StructType as schema wh...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1571 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1652/ ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1603 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1651/ ---
[GitHub] carbondata issue #1571: [CARBONDATA-1811] [CTAS] Use StructType as schema wh...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1571 SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2040/ ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user jackylk commented on the issue: https://github.com/apache/carbondata/pull/1603 retest this please ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1603 SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2039/ ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1603 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1649/ ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user ravipesala commented on the issue: https://github.com/apache/carbondata/pull/1603 SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2038/ ---
[GitHub] carbondata issue #1603: [CARBONDATA-1844] Add tablePath support when creatin...
Github user CarbonDataQA commented on the issue: https://github.com/apache/carbondata/pull/1603 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1648/ ---