[GitHub] carbondata issue #1759: [CARBONDATA-1980][PARTITION]Fix adding partition inf...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1759
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2694/



---


[GitHub] carbondata issue #1760: [CARBONDATA-1979] ][IMPLICIT COLUMN] Modified implic...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1760
  
Build Success with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1306/



---


[GitHub] carbondata issue #1740: [CARBONDATA-1949, CARBONDATA-1950] Fixed bug related...

2018-01-03 Thread geetikagupta16
Github user geetikagupta16 commented on the issue:

https://github.com/apache/carbondata/pull/1740
  
@manishgupta88 I have added the test case. Please check


---


[jira] [Created] (CARBONDATA-1982) Loading data into partition table with invalid partition column should throw proper exception

2018-01-03 Thread Geetika Gupta (JIRA)
Geetika Gupta created CARBONDATA-1982:
-

 Summary: Loading data into partition table with invalid partition 
column should throw proper exception
 Key: CARBONDATA-1982
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1982
 Project: CarbonData
  Issue Type: Bug
  Components: data-query
Affects Versions: 1.3.0
 Environment: spark2.1
Reporter: Geetika Gupta
Priority: Minor
 Fix For: 1.3.0
 Attachments: 2000_UniqData.csv

I created a partitioned table using:
CREATE TABLE uniqdata_int_dec(CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB 
timestamp,
DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,
DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,
INTEGER_COLUMN1 int) Partitioned by (cust_id int, decimal_column1 
decimal(30,10)) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES 
("TABLE_BLOCKSIZE"= "256 MB")

Load data command:
LOAD DATA INPATH 'hdfs://localhost:54311/2000_UniqData.csv' into table 
uniqdata_int_dec partition(cust_id123='1', abc='12345678901.1234') OPTIONS 
('FILEHEADER'='CUST_ID,CUST_NAME ,ACTIVE_EMUI_VERSION,DOB,DOJ, 
BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1, 
Double_COLUMN2,INTEGER_COLUMN1','BAD_RECORDS_ACTION'='FORCE');

OUTPUT:
0: jdbc:hive2://localhost:1> LOAD DATA INPATH 
'hdfs://localhost:54311/Files/2000_UniqData.csv' into table uniqdata_int_dec 
partition(cust_id123='1', decimal_column1='12345678901.1234') OPTIONS 
('FILEHEADER'='CUST_ID,CUST_NAME ,ACTIVE_EMUI_VERSION,DOB,DOJ, 
BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1, 
Double_COLUMN2,INTEGER_COLUMN1','BAD_RECORDS_ACTION'='FORCE');
Error: java.lang.IndexOutOfBoundsException: Index: 1, Size: 1 (state=,code=0)

The above command throws java.lang.IndexOutOfBoundsException: Index: 1, Size: 1 
whereas it should throw a proper exception like invalid column expression for 
partition load command.




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] carbondata pull request #1752: [CARBONDATA-1972][PARTITION] Fix compaction a...

2018-01-03 Thread jackylk
Github user jackylk commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1752#discussion_r159594673
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/mutation/CarbonProjectForDeleteCommand.scala
 ---
@@ -77,7 +77,7 @@ private[sql] case class CarbonProjectForDeleteCommand(
 dataRdd,
 timestamp,
 isUpdateOperation = false,
-executorErrors)) {
+executorErrors)._1) {
--- End diff --

It seems always true


---


[GitHub] carbondata pull request #1752: [CARBONDATA-1972][PARTITION] Fix compaction a...

2018-01-03 Thread jackylk
Github user jackylk commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1752#discussion_r159594689
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/execution/datasources/CarbonFileFormat.scala
 ---
@@ -107,11 +107,19 @@ with Serializable {
 CarbonTableOutputFormat.setOverwrite(conf, 
options("overwrite").toBoolean)
 // Set the update timestamp if user sets in case of update query. It 
needs to be updated
 // in load status update time
-val updateTimeStamp = options.getOrElse("updatetimestamp", null)
-if (updateTimeStamp != null) {
-  conf.set(CarbonTableOutputFormat.UPADTE_TIMESTAMP, updateTimeStamp)
-  model.setFactTimeStamp(updateTimeStamp.toLong)
+val updateTimeStamp = options.get("updatetimestamp")
+if (updateTimeStamp.isDefined) {
+  conf.set(CarbonTableOutputFormat.UPADTE_TIMESTAMP, 
updateTimeStamp.get)
+  model.setFactTimeStamp(updateTimeStamp.get.toLong)
 }
+// In case of update query there is chance to remove the older 
segments, so here we can set
+// the to be deleted segments to mark as delete while updating 
tablestatus
+val segemntsTobeDeleted = options.get("segmentsToBeDeleted")
+if (segemntsTobeDeleted.isDefined) {
+  conf.set(CarbonTableOutputFormat.SEGMENTS_TO_BE_DELETED, 
segemntsTobeDeleted.get)
+}
+
+
--- End diff --

remove empty line


---


[GitHub] carbondata pull request #1752: [CARBONDATA-1972][PARTITION] Fix compaction a...

2018-01-03 Thread jackylk
Github user jackylk commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1752#discussion_r159594614
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/mutation/DeleteExecution.scala
 ---
@@ -58,14 +58,15 @@ object DeleteExecution {
   dataRdd: RDD[Row],
   timestamp: String,
   isUpdateOperation: Boolean,
-  executorErrors: ExecutionErrors): Boolean = {
+  executorErrors: ExecutionErrors): (Boolean, Seq[String]) = {
--- End diff --

It seems the first return value is always true
And it is hard to understand this function, can you add comment for it


---


[GitHub] carbondata issue #1755: [CARBONDATA-1976][PARTITION] Support combination of ...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1755
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2535/



---


[GitHub] carbondata pull request #1755: [CARBONDATA-1976][PARTITION] Support combinat...

2018-01-03 Thread jackylk
Github user jackylk commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1755#discussion_r159591690
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala
 ---
@@ -611,7 +611,7 @@ case class CarbonLoadDataCommand(
   val output = if (partition.nonEmpty) {
 catalogTable.schema.map{ attr =>
   attributes.find(_.name.equalsIgnoreCase(attr.name)).get
-}.filter(attr => partition.get(attr.name).isEmpty)
+}.filter(attr => partition.getOrElse(attr.name, None).isEmpty)
--- End diff --

It seems the same as before?


---


[GitHub] carbondata pull request #1755: [CARBONDATA-1976][PARTITION] Support combinat...

2018-01-03 Thread jackylk
Github user jackylk commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1755#discussion_r159592167
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala
 ---
@@ -702,63 +702,71 @@ case class CarbonLoadDataCommand(
   sparkSession: SparkSession,
   table: CarbonTable,
   logicalPlan: LogicalPlan): Unit = {
-sparkSession.sessionState.catalog.listPartitions(
+val existingPartitions = 
sparkSession.sessionState.catalog.listPartitions(
   TableIdentifier(table.getTableName, Some(table.getDatabaseName)),
-  Some(partition.map(f => (f._1, f._2.get
-val partitionNames = partition.map(k => k._1 + "=" + k._2.get).toSet
+  Some(partition.filter(_._2.isDefined).map(f => (f._1, f._2.get
+val partitionNames = existingPartitions.toList.flatMap { partition =>
+  partition.spec.seq.map{case (column, value) => column + "=" + value}
+}.toSet
 val uniqueId = System.currentTimeMillis().toString
 val segments = new SegmentStatusManager(
   
table.getAbsoluteTableIdentifier).getValidAndInvalidSegments.getValidSegments
-try {
-  // First drop the partitions from partition mapper files of each 
segment
-  new CarbonDropPartitionRDD(
-sparkSession.sparkContext,
-table.getTablePath,
-segments.asScala,
-partitionNames.toSeq,
-uniqueId).collect()
-} catch {
-  case e: Exception =>
-// roll back the drop partitions from carbon store
-new CarbonDropPartitionCommitRDD(
+// If any existing partitions need to be overwritten then drop from 
partitionmap
+if (partitionNames.nonEmpty) {
+  try {
+// First drop the partitions from partition mapper files of each 
segment
+new CarbonDropPartitionRDD(
   sparkSession.sparkContext,
   table.getTablePath,
   segments.asScala,
-  false,
+  partitionNames.toSeq,
   uniqueId).collect()
-throw e
-}
+  } catch {
+case e: Exception =>
+  // roll back the drop partitions from carbon store
+  new CarbonDropPartitionCommitRDD(
+sparkSession.sparkContext,
+table.getTablePath,
+segments.asScala,
+false,
+uniqueId).collect()
+  throw e
+  }
 
-try {
+  try {
+Dataset.ofRows(sparkSession, logicalPlan)
+  } catch {
+case e: Exception =>
+  // roll back the drop partitions from carbon store
+  new CarbonDropPartitionCommitRDD(
+sparkSession.sparkContext,
+table.getTablePath,
+segments.asScala,
+false,
+uniqueId).collect()
+  throw e
+  }
+  // Commit the removed partitions in carbon store.
+  new CarbonDropPartitionCommitRDD(
+sparkSession.sparkContext,
+table.getTablePath,
+segments.asScala,
+true,
+uniqueId).collect()
+  // Update the loadstatus with update time to clear cache from driver.
+  val segmentSet = new util.HashSet[String](new 
SegmentStatusManager(table
+
.getAbsoluteTableIdentifier).getValidAndInvalidSegments.getValidSegments)
--- End diff --

make these two lines more readable


---


[GitHub] carbondata pull request #1755: [CARBONDATA-1976][PARTITION] Support combinat...

2018-01-03 Thread jackylk
Github user jackylk commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1755#discussion_r159591992
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala
 ---
@@ -702,63 +702,71 @@ case class CarbonLoadDataCommand(
   sparkSession: SparkSession,
   table: CarbonTable,
   logicalPlan: LogicalPlan): Unit = {
-sparkSession.sessionState.catalog.listPartitions(
+val existingPartitions = 
sparkSession.sessionState.catalog.listPartitions(
   TableIdentifier(table.getTableName, Some(table.getDatabaseName)),
-  Some(partition.map(f => (f._1, f._2.get
-val partitionNames = partition.map(k => k._1 + "=" + k._2.get).toSet
+  Some(partition.filter(_._2.isDefined).map(f => (f._1, f._2.get
+val partitionNames = existingPartitions.toList.flatMap { partition =>
+  partition.spec.seq.map{case (column, value) => column + "=" + value}
+}.toSet
 val uniqueId = System.currentTimeMillis().toString
 val segments = new SegmentStatusManager(
   
table.getAbsoluteTableIdentifier).getValidAndInvalidSegments.getValidSegments
-try {
-  // First drop the partitions from partition mapper files of each 
segment
-  new CarbonDropPartitionRDD(
-sparkSession.sparkContext,
-table.getTablePath,
-segments.asScala,
-partitionNames.toSeq,
-uniqueId).collect()
-} catch {
-  case e: Exception =>
-// roll back the drop partitions from carbon store
-new CarbonDropPartitionCommitRDD(
+// If any existing partitions need to be overwritten then drop from 
partitionmap
+if (partitionNames.nonEmpty) {
+  try {
+// First drop the partitions from partition mapper files of each 
segment
+new CarbonDropPartitionRDD(
   sparkSession.sparkContext,
   table.getTablePath,
   segments.asScala,
-  false,
+  partitionNames.toSeq,
   uniqueId).collect()
-throw e
-}
+  } catch {
+case e: Exception =>
+  // roll back the drop partitions from carbon store
+  new CarbonDropPartitionCommitRDD(
+sparkSession.sparkContext,
+table.getTablePath,
+segments.asScala,
+false,
+uniqueId).collect()
+  throw e
+  }
 
-try {
+  try {
+Dataset.ofRows(sparkSession, logicalPlan)
+  } catch {
+case e: Exception =>
+  // roll back the drop partitions from carbon store
+  new CarbonDropPartitionCommitRDD(
+sparkSession.sparkContext,
+table.getTablePath,
+segments.asScala,
+false,
+uniqueId).collect()
+  throw e
+  }
+  // Commit the removed partitions in carbon store.
+  new CarbonDropPartitionCommitRDD(
+sparkSession.sparkContext,
+table.getTablePath,
+segments.asScala,
+true,
--- End diff --

provide variable name to improve readability


---


[GitHub] carbondata pull request #1755: [CARBONDATA-1976][PARTITION] Support combinat...

2018-01-03 Thread jackylk
Github user jackylk commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1755#discussion_r159591980
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala
 ---
@@ -702,63 +702,71 @@ case class CarbonLoadDataCommand(
   sparkSession: SparkSession,
   table: CarbonTable,
   logicalPlan: LogicalPlan): Unit = {
-sparkSession.sessionState.catalog.listPartitions(
+val existingPartitions = 
sparkSession.sessionState.catalog.listPartitions(
   TableIdentifier(table.getTableName, Some(table.getDatabaseName)),
-  Some(partition.map(f => (f._1, f._2.get
-val partitionNames = partition.map(k => k._1 + "=" + k._2.get).toSet
+  Some(partition.filter(_._2.isDefined).map(f => (f._1, f._2.get
+val partitionNames = existingPartitions.toList.flatMap { partition =>
+  partition.spec.seq.map{case (column, value) => column + "=" + value}
+}.toSet
 val uniqueId = System.currentTimeMillis().toString
 val segments = new SegmentStatusManager(
   
table.getAbsoluteTableIdentifier).getValidAndInvalidSegments.getValidSegments
-try {
-  // First drop the partitions from partition mapper files of each 
segment
-  new CarbonDropPartitionRDD(
-sparkSession.sparkContext,
-table.getTablePath,
-segments.asScala,
-partitionNames.toSeq,
-uniqueId).collect()
-} catch {
-  case e: Exception =>
-// roll back the drop partitions from carbon store
-new CarbonDropPartitionCommitRDD(
+// If any existing partitions need to be overwritten then drop from 
partitionmap
+if (partitionNames.nonEmpty) {
+  try {
+// First drop the partitions from partition mapper files of each 
segment
+new CarbonDropPartitionRDD(
   sparkSession.sparkContext,
   table.getTablePath,
   segments.asScala,
-  false,
+  partitionNames.toSeq,
   uniqueId).collect()
-throw e
-}
+  } catch {
+case e: Exception =>
+  // roll back the drop partitions from carbon store
+  new CarbonDropPartitionCommitRDD(
+sparkSession.sparkContext,
+table.getTablePath,
+segments.asScala,
+false,
+uniqueId).collect()
+  throw e
+  }
 
-try {
+  try {
+Dataset.ofRows(sparkSession, logicalPlan)
+  } catch {
+case e: Exception =>
+  // roll back the drop partitions from carbon store
+  new CarbonDropPartitionCommitRDD(
+sparkSession.sparkContext,
+table.getTablePath,
+segments.asScala,
+false,
--- End diff --

provide variable name to improve readability


---


[GitHub] carbondata pull request #1755: [CARBONDATA-1976][PARTITION] Support combinat...

2018-01-03 Thread jackylk
Github user jackylk commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1755#discussion_r159591969
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala
 ---
@@ -702,63 +702,71 @@ case class CarbonLoadDataCommand(
   sparkSession: SparkSession,
   table: CarbonTable,
   logicalPlan: LogicalPlan): Unit = {
-sparkSession.sessionState.catalog.listPartitions(
+val existingPartitions = 
sparkSession.sessionState.catalog.listPartitions(
   TableIdentifier(table.getTableName, Some(table.getDatabaseName)),
-  Some(partition.map(f => (f._1, f._2.get
-val partitionNames = partition.map(k => k._1 + "=" + k._2.get).toSet
+  Some(partition.filter(_._2.isDefined).map(f => (f._1, f._2.get
+val partitionNames = existingPartitions.toList.flatMap { partition =>
+  partition.spec.seq.map{case (column, value) => column + "=" + value}
+}.toSet
 val uniqueId = System.currentTimeMillis().toString
 val segments = new SegmentStatusManager(
   
table.getAbsoluteTableIdentifier).getValidAndInvalidSegments.getValidSegments
-try {
-  // First drop the partitions from partition mapper files of each 
segment
-  new CarbonDropPartitionRDD(
-sparkSession.sparkContext,
-table.getTablePath,
-segments.asScala,
-partitionNames.toSeq,
-uniqueId).collect()
-} catch {
-  case e: Exception =>
-// roll back the drop partitions from carbon store
-new CarbonDropPartitionCommitRDD(
+// If any existing partitions need to be overwritten then drop from 
partitionmap
+if (partitionNames.nonEmpty) {
+  try {
+// First drop the partitions from partition mapper files of each 
segment
+new CarbonDropPartitionRDD(
   sparkSession.sparkContext,
   table.getTablePath,
   segments.asScala,
-  false,
+  partitionNames.toSeq,
   uniqueId).collect()
-throw e
-}
+  } catch {
+case e: Exception =>
+  // roll back the drop partitions from carbon store
+  new CarbonDropPartitionCommitRDD(
+sparkSession.sparkContext,
+table.getTablePath,
+segments.asScala,
+false,
--- End diff --

provide variable name to improve readability


---


[jira] [Commented] (CARBONDATA-1936) Bad Record logger is not working properly in Carbon Partition

2018-01-03 Thread Ravindra Pesala (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-1936?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16310903#comment-16310903
 ] 

Ravindra Pesala commented on CARBONDATA-1936:
-

: CREATE  TABLE IF NOT EXISTS emp1 (emp_no int,ename string,job string,mgr_id 
int,date_of_joining string,salary int,bonus int) partitioned by (dept_no int) 
STORED BY 'org.apache.carbondata.format' ;
2: LOAD DATA INPATH 'hdfs://hacluster/user/test/emp.csv' overwrite INTO TABLE 
emp1 OPTIONS('DELIMITER'=',', 'QUOTECHAR'= '\');
3: Observe that load is failing
4: LOAD DATA INPATH 'hdfs://hacluster/user/test/emp.csv' overwrite INTO TABLE 
emp1 OPTIONS('DELIMITER'=',', 'QUOTECHAR'= '\','BAD_RECORDS_ACTION'='FORCE');
5: Observe that load is success when 'BAD_RECORDS_ACTION'='FORCE' is given in 
load command
6: Also observe that same load is suucess in a table without partition.

> Bad Record logger is not working properly in Carbon Partition
> -
>
> Key: CARBONDATA-1936
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1936
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Ravindra Pesala
>Priority: Minor
> Attachments: test.csv.csv
>
>  Time Spent: 4h
>  Remaining Estimate: 0h
>
> Bad records are not logging and the load is always success irrespective of 
> bad records are present. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1936) Bad Record logger is not working properly in Carbon Partition

2018-01-03 Thread Ravindra Pesala (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1936?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravindra Pesala updated CARBONDATA-1936:

Description: 
Bad records are not logging and the load is always success irrespective of bad 
records are present. 

: CREATE  TABLE IF NOT EXISTS emp1 (emp_no int,ename string,job string,mgr_id 
int,date_of_joining string,salary int,bonus int) partitioned by (dept_no int) 
STORED BY 'org.apache.carbondata.format' ;
2: LOAD DATA INPATH 'hdfs://hacluster/user/test/emp.csv' overwrite INTO TABLE 
emp1 OPTIONS('DELIMITER'=',', 'QUOTECHAR'= '\');
3: Observe that load is failing
4: LOAD DATA INPATH 'hdfs://hacluster/user/test/emp.csv' overwrite INTO TABLE 
emp1 OPTIONS('DELIMITER'=',', 'QUOTECHAR'= '\','BAD_RECORDS_ACTION'='FORCE');
5: Observe that load is success when 'BAD_RECORDS_ACTION'='FORCE' is given in 
load command
6: Also observe that same load is suucess in a table without partition.

  was:Bad records are not logging and the load is always success irrespective 
of bad records are present. 


> Bad Record logger is not working properly in Carbon Partition
> -
>
> Key: CARBONDATA-1936
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1936
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Ravindra Pesala
>Priority: Minor
> Attachments: test.csv.csv
>
>  Time Spent: 4h
>  Remaining Estimate: 0h
>
> Bad records are not logging and the load is always success irrespective of 
> bad records are present. 
> : CREATE  TABLE IF NOT EXISTS emp1 (emp_no int,ename string,job string,mgr_id 
> int,date_of_joining string,salary int,bonus int) partitioned by (dept_no int) 
> STORED BY 'org.apache.carbondata.format' ;
> 2: LOAD DATA INPATH 'hdfs://hacluster/user/test/emp.csv' overwrite INTO TABLE 
> emp1 OPTIONS('DELIMITER'=',', 'QUOTECHAR'= '\');
> 3: Observe that load is failing
> 4: LOAD DATA INPATH 'hdfs://hacluster/user/test/emp.csv' overwrite INTO TABLE 
> emp1 OPTIONS('DELIMITER'=',', 'QUOTECHAR'= '\','BAD_RECORDS_ACTION'='FORCE');
> 5: Observe that load is success when 'BAD_RECORDS_ACTION'='FORCE' is given in 
> load command
> 6: Also observe that same load is suucess in a table without partition.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Issue Comment Deleted] (CARBONDATA-1936) Bad Record logger is not working properly in Carbon Partition

2018-01-03 Thread Ravindra Pesala (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1936?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravindra Pesala updated CARBONDATA-1936:

Comment: was deleted

(was: : CREATE  TABLE IF NOT EXISTS emp1 (emp_no int,ename string,job 
string,mgr_id int,date_of_joining string,salary int,bonus int) partitioned by 
(dept_no int) STORED BY 'org.apache.carbondata.format' ;
2: LOAD DATA INPATH 'hdfs://hacluster/user/test/emp.csv' overwrite INTO TABLE 
emp1 OPTIONS('DELIMITER'=',', 'QUOTECHAR'= '\');
3: Observe that load is failing
4: LOAD DATA INPATH 'hdfs://hacluster/user/test/emp.csv' overwrite INTO TABLE 
emp1 OPTIONS('DELIMITER'=',', 'QUOTECHAR'= '\','BAD_RECORDS_ACTION'='FORCE');
5: Observe that load is success when 'BAD_RECORDS_ACTION'='FORCE' is given in 
load command
6: Also observe that same load is suucess in a table without partition.)

> Bad Record logger is not working properly in Carbon Partition
> -
>
> Key: CARBONDATA-1936
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1936
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Ravindra Pesala
>Priority: Minor
> Attachments: test.csv.csv
>
>  Time Spent: 4h
>  Remaining Estimate: 0h
>
> Bad records are not logging and the load is always success irrespective of 
> bad records are present. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (CARBONDATA-1936) Bad Record logger is not working properly in Carbon Partition

2018-01-03 Thread Ravindra Pesala (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1936?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravindra Pesala updated CARBONDATA-1936:

Attachment: test.csv.csv

> Bad Record logger is not working properly in Carbon Partition
> -
>
> Key: CARBONDATA-1936
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1936
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Ravindra Pesala
>Priority: Minor
> Attachments: test.csv.csv
>
>  Time Spent: 4h
>  Remaining Estimate: 0h
>
> Bad records are not logging and the load is always success irrespective of 
> bad records are present. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] carbondata pull request #1755: [CARBONDATA-1976][PARTITION] Support combinat...

2018-01-03 Thread jackylk
Github user jackylk commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1755#discussion_r159591267
  
--- Diff: 
integration/spark-common-test/src/test/scala/org/apache/carbondata/spark/testsuite/standardpartition/StandardPartitionTableLoadingTestCase.scala
 ---
@@ -28,7 +31,7 @@ import org.apache.carbondata.core.util.CarbonProperties
 import org.apache.carbondata.core.util.path.CarbonTablePath
 
 class StandardPartitionTableLoadingTestCase extends QueryTest with 
BeforeAndAfterAll {
-
+  var executorService: ExecutorService = _
--- End diff --

move to within the testcase as only that testcase uses it


---


[GitHub] carbondata issue #1678: [CARBONDATA-1903] Fix code issues in carbondata

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1678
  
Build Failed with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1305/



---


[GitHub] carbondata issue #1752: [CARBONDATA-1972][PARTITION] Fix compaction after up...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1752
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2693/



---


[GitHub] carbondata issue #1711: [CARBONDATA-1754][BugFix] Fixed issue occuring on co...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1711
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2534/



---


[jira] [Closed] (CARBONDATA-1966) SET command for carbon.properties.filepath is not setting the property

2018-01-03 Thread Geetika Gupta (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1966?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Geetika Gupta closed CARBONDATA-1966.
-
Resolution: Fixed

> SET command for carbon.properties.filepath is not setting the property
> --
>
> Key: CARBONDATA-1966
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1966
> Project: CarbonData
>  Issue Type: Bug
>  Components: data-query
>Affects Versions: 1.3.0
> Environment: spark2.1
>Reporter: Geetika Gupta
> Fix For: 1.3.0
>
>
> I executed the following command:
> set carbon.properties.filepath=/home
> It is executed successfully, but it's value is not updated on Spark UI in 
> spark properties.
> It shows the following output:
> jdbc:hive2://localhost:1> set carbon.properties.filepath;
> +-++--+
> | key | value  |
> +-++--+
> | carbon.properties.filepath  | /home  |
> +-++--+
> 1 row selected (0.029 seconds)



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (CARBONDATA-1966) SET command for carbon.properties.filepath is not setting the property

2018-01-03 Thread Geetika Gupta (JIRA)

[ 
https://issues.apache.org/jira/browse/CARBONDATA-1966?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16310873#comment-16310873
 ] 

Geetika Gupta commented on CARBONDATA-1966:
---

This JIRA is invalid as carbon.properties.filepath is not a dynamically 
configurable property

> SET command for carbon.properties.filepath is not setting the property
> --
>
> Key: CARBONDATA-1966
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1966
> Project: CarbonData
>  Issue Type: Bug
>  Components: data-query
>Affects Versions: 1.3.0
> Environment: spark2.1
>Reporter: Geetika Gupta
> Fix For: 1.3.0
>
>
> I executed the following command:
> set carbon.properties.filepath=/home
> It is executed successfully, but it's value is not updated on Spark UI in 
> spark properties.
> It shows the following output:
> jdbc:hive2://localhost:1> set carbon.properties.filepath;
> +-++--+
> | key | value  |
> +-++--+
> | carbon.properties.filepath  | /home  |
> +-++--+
> 1 row selected (0.029 seconds)



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] carbondata pull request #1747: [Compatibility] Added changes for backward co...

2018-01-03 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1747#discussion_r159589331
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/util/path/CarbonTablePath.java ---
@@ -231,7 +231,22 @@ public String getSchemaFilePath() {
* @return schema file path
*/
   public static String getSchemaFilePath(String tablePath) {
-return tablePath + File.separator + METADATA_DIR + File.separator + 
SCHEMA_FILE;
+return getActualSchemaFilePath(tablePath);
+  }
+
+  private static String getActualSchemaFilePath(String tablePath) {
+String metaPath = tablePath + File.separator + METADATA_DIR;
--- End diff --

use CarbonCommonConstants.FILE_SEPARATOR instead of File.separator


---


[GitHub] carbondata pull request #1747: [Compatibility] Added changes for backward co...

2018-01-03 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1747#discussion_r159589266
  
--- Diff: 
integration/spark-common/src/main/scala/org/apache/carbondata/spark/util/CarbonScalaUtil.scala
 ---
@@ -291,4 +291,52 @@ object CarbonScalaUtil {
   })
 otherFields
   }
+
+  /**
+   * If the table is from an old store then the table parameters are in 
lowercase. In the current
+   * code we are reading the parameters as camel case.
+   * This method will convert all the schema parts to camel case
+   *
+   * @param parameters
+   * @return
+   */
+  def getDeserializedParameters(parameters: Map[String, String]): 
Map[String, String] = {
--- End diff --

there is no calling point for this method


---


[GitHub] carbondata pull request #1747: [Compatibility] Added changes for backward co...

2018-01-03 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1747#discussion_r159589250
  
--- Diff: 
integration/spark-common/src/main/scala/org/apache/carbondata/spark/util/CarbonScalaUtil.scala
 ---
@@ -291,4 +291,52 @@ object CarbonScalaUtil {
   })
 otherFields
   }
+
+  /**
+   * If the table is from an old store then the table parameters are in 
lowercase. In the current
+   * code we are reading the parameters as camel case.
+   * This method will convert all the schema parts to camel case
+   *
+   * @param parameters
+   * @return
+   */
+  def getDeserializedParameters(parameters: Map[String, String]): 
Map[String, String] = {
+val keyParts = 
parameters.getOrElse("spark.sql.sources.options.keys.numparts", "0").toInt
+if (keyParts == 0) {
+  parameters
+} else {
+  var keyStr = ""
+  for(i <- 0 until keyParts) {
+keyStr += parameters(s"spark.sql.sources.options.keys.part.$i")
+  }
+  val finalProperties = scala.collection.mutable.Map.empty[String, 
String]
+  keyStr.split(",") foreach {
--- End diff --

Please check the logic here.  Keystr is not appended explicitly with , but 
splitting is done based on comma


---


[GitHub] carbondata pull request #1732: [CARBONDATA-1946] Exception thrown after alte...

2018-01-03 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1732#discussion_r159589346
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/scan/result/vector/CarbonColumnVector.java
 ---
@@ -67,6 +67,20 @@
 
   DataType getType();
 
+  /**
+   * Method to be used for getting the restructured data type. This method 
will used for
+   * retrieving the data after change in data type restructure operation
+   *
+   * @return
+   */
+  DataType getRestructuredType();
--- End diff --

Better rename as BlockDataType to give more meaning ful


---


[GitHub] carbondata pull request #1759: [CARBONDATA-1980][PARTITION]Fix adding partit...

2018-01-03 Thread jackylk
Github user jackylk commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1759#discussion_r159589011
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/RefreshCarbonTableCommand.scala
 ---
@@ -91,6 +95,11 @@ case class RefreshCarbonTableCommand(
   registerAggregates(databaseName, dataMapSchemaList)(sparkSession)
 }
 registerTableWithHive(databaseName, tableName, 
tableInfo)(sparkSession)
+// Register partitions to hive metastore in case of hive 
partitioning carbon table
+if (tableInfo.getFactTable.getPartitionInfo != null &&
+tableInfo.getFactTable.getPartitionInfo.getPartitionType == 
PartitionType.NATIVE_HIVE) {
+  registerAllPartitionsToHive(tablePath, databaseName, tableName, 
sparkSession)
--- End diff --

I think passing AbsolutionTableIdentifier is better, it has all 
information. It is created in line 65


---


[jira] [Updated] (CARBONDATA-1956) Select query with sum, count and avg throws exception for pre aggregate table

2018-01-03 Thread Geetika Gupta (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1956?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Geetika Gupta updated CARBONDATA-1956:
--
Description: 
I create a datamap using the following command:

create datamap uniqdata_agg_d on table uniqdata_29 using 'preaggregate' as 
select sum(decimal_column1), count(cust_id), avg(bigint_column1) from 
uniqdata_29 group by cust_id;

The datamap creation was successfull, but when I tried the following query:
select sum(decimal_column1), count(cust_id), avg(bigint_column1) from 
uniqdata_29 group by cust_id;

It throws the following exception:
Error: org.apache.spark.sql.AnalysisException: cannot resolve 
'(sum(uniqdata_29_uniqdata_agg_d.`uniqdata_29_bigint_column1_sum`) / 
sum(uniqdata_29_uniqdata_agg_d.`uniqdata_29_bigint_column1_count`))' due to 
data type mismatch: 
'(sum(uniqdata_29_uniqdata_agg_d.`uniqdata_29_bigint_column1_sum`) / 
sum(uniqdata_29_uniqdata_agg_d.`uniqdata_29_bigint_column1_count`))' requires 
(double or decimal) type, not bigint;;
'Aggregate [uniqdata_29_cust_id_count#244], 
[sum(uniqdata_29_decimal_column1_sum#243) AS sum(decimal_column1)#274, 
sum(cast(uniqdata_29_cust_id_count#244 as bigint)) AS count(cust_id)#276L, 
(sum(uniqdata_29_bigint_column1_sum#245L) / 
sum(uniqdata_29_bigint_column1_count#246L)) AS avg(bigint_column1)#279]
+- 
Relation[uniqdata_29_decimal_column1_sum#243,uniqdata_29_cust_id_count#244,uniqdata_29_bigint_column1_sum#245L,uniqdata_29_bigint_column1_count#246L]
 CarbonDatasourceHadoopRelation [ Database name :28dec, Table name 
:uniqdata_29_uniqdata_agg_d, Schema 
:Some(StructType(StructField(uniqdata_29_decimal_column1_sum,DecimalType(30,10),true),
 StructField(uniqdata_29_cust_id_count,IntegerType,true), 
StructField(uniqdata_29_bigint_column1_sum,LongType,true), 
StructField(uniqdata_29_bigint_column1_count,LongType,true))) ] (state=,code=0)

Steps for creation of maintable:
CREATE TABLE uniqdata_29(CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 
int) STORED BY 'org.apache.carbondata.format';

Load command:
LOAD DATA INPATH 'hdfs://localhost:54311/Files/2000_UniqData.csv' into table 
uniqdata_29 OPTIONS('DELIMITER'=',', 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');

Datamap creation command:
create datamap uniqdata_agg_d on table uniqdata_29 using 'preaggregate' as 
select sum(decimal_column1), count(cust_id), avg(bigint_column1) from 
uniqdata_29 group by cust_id;

Note: select sum(decimal_column1), count(cust_id), avg(bigint_column1) from 
uniqdata_29 group by cust_id; executed successfully on maintable



  was:
I create a datamap using the following command:

create datamap uniqdata_agg_d on table uniqdata_29 using 'preaggregate' as 
select sum(decimal_column1), count(cust_id), avg(bigint_column1) from 
uniqdata_29 group by cust_id;

The datamap creation was successfull, but when I tried the following query:
select sum(decimal_column1), count(cust_id), avg(bigint_column1) from 
uniqdata_29 group by cust_id;

It throws the following exception:
Error: org.apache.spark.sql.AnalysisException: cannot resolve 
'(sum(uniqdata_29_uniqdata_agg_d.`uniqdata_29_bigint_column1_sum`) / 
sum(uniqdata_29_uniqdata_agg_d.`uniqdata_29_bigint_column1_count`))' due to 
data type mismatch: 
'(sum(uniqdata_29_uniqdata_agg_d.`uniqdata_29_bigint_column1_sum`) / 
sum(uniqdata_29_uniqdata_agg_d.`uniqdata_29_bigint_column1_count`))' requires 
(double or decimal) type, not bigint;;
'Aggregate [uniqdata_29_cust_id_count#244], 
[sum(uniqdata_29_decimal_column1_sum#243) AS sum(decimal_column1)#274, 
sum(cast(uniqdata_29_cust_id_count#244 as bigint)) AS count(cust_id)#276L, 
(sum(uniqdata_29_bigint_column1_sum#245L) / 
sum(uniqdata_29_bigint_column1_count#246L)) AS avg(bigint_column1)#279]
+- 
Relation[uniqdata_29_decimal_column1_sum#243,uniqdata_29_cust_id_count#244,uniqdata_29_bigint_column1_sum#245L,uniqdata_29_bigint_column1_count#246L]
 CarbonDatasourceHadoopRelation [ Database name :28dec, Table name 
:uniqdata_29_uniqdata_agg_d, Schema 
:Some(StructType(StructField(uniqdata_29_decimal_column1_sum,DecimalType(30,10),true),
 StructField(uniqdata_29_cust_id_count,IntegerType,true), 
StructField(uniqdata_29_bigint_column1_sum,LongType,true), 
StructField(uniqdata_29_bigint_column1_count,LongType,true))) ] (state=,code=0)

Steps for creation of maintable:
CREATE TABLE uniqdata_29(CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION 
string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 
bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 
decimal(36,10),Double_COLUMN1 do

[GitHub] carbondata issue #1748: [CARBONDATA-1967][PARTITION] Fix autocompaction and ...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1748
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2692/



---


[jira] [Created] (CARBONDATA-1981) Error occurs while building project in windows environment

2018-01-03 Thread xuchuanyin (JIRA)
xuchuanyin created CARBONDATA-1981:
--

 Summary: Error occurs while building project in windows environment
 Key: CARBONDATA-1981
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1981
 Project: CarbonData
  Issue Type: Bug
  Components: build, core
Affects Versions: 1.3.0
Reporter: xuchuanyin


Encounter error while building project in Windows environment, the error 
messages are shown as bellow:

```
[INFO] Apache CarbonData :: Parent  SUCCESS [ 3.915 s]
[INFO] Apache CarbonData :: Common  SUCCESS [ 12.007 s]
[INFO] Apache CarbonData :: Format  SUCCESS [ 30.059 s]
[INFO] Apache CarbonData :: Core .. FAILURE [ 3.604 s]

[INFO] 
[INFO] BUILD FAILURE
[INFO] 
[INFO] Total time: 50.099 s
[INFO] Finished at: 2018-01-04T11:19:28+08:00
[INFO] Final Memory: 62M/612M
[INFO] 
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-antrun-plugin:1.8:run (default) on project 
carbondata-core: An Ant BuildException has occured: Execute failed: 
java.io.IOException: Cannot run program "bash" (in directory 
"D:\01_workspace\carbondata\core"): CreateProcess error=2, 系统找不到指定的文件。(cannot 
find the specified file)
[ERROR] around Ant part .. @ 4:27 in 
D:\01_workspace\carbondata\core\target\antrun\build-main.xml
[ERROR] -> [Help 1]
```

It is because in the pom, it calls bash script, which is not incompatible in 
windows.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] carbondata issue #1729: [CARBONDATA-1936][PARTITION] Corrected bad record an...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1729
  
Build Failed with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1304/



---


[GitHub] carbondata issue #1760: [CARBONDATA-1979] ][IMPLICIT COLUMN] Modified implic...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1760
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2533/



---


[GitHub] carbondata pull request #1082: [CARBONDATA-1218] [GLOBAL SORT] In case of da...

2018-01-03 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1082#discussion_r159586543
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/util/CarbonDataProcessorUtil.java
 ---
@@ -145,6 +147,15 @@ public static void 
renameBadRecordsFromInProgressToNormal(
 }
   }
 
+  public static void renameBadRecord(CarbonDataLoadConfiguration 
configuration) {
+// rename the bad record in progress to normal
+CarbonTableIdentifier identifier =
+configuration.getTableIdentifier().getCarbonTableIdentifier();
+
CarbonDataProcessorUtil.renameBadRecordsFromInProgressToNormal(configuration,
--- End diff --

1. Avoid calling the method using classname. Method is present in the same 
class.
2. I think we can introduce a new class called BadRecordsUtil which will 
have all the functions related to bad records like 
renameBadRecord, hasBadRecord, renameBadRecordsFromInProgressToNormal (make 
it private in the new util class)


---


[GitHub] carbondata pull request #1082: [CARBONDATA-1218] [GLOBAL SORT] In case of da...

2018-01-03 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1082#discussion_r159584326
  
--- Diff: 
integration/spark-common-test/src/test/scala/org/apache/carbondata/spark/testsuite/badrecordloger/BadRecordLoggerTest.scala
 ---
@@ -247,6 +248,33 @@ class BadRecordLoggerTest extends QueryTest with 
BeforeAndAfterAll {
 }
   }
 
+  test(
+"test if first load failed due to bad record then second load should 
not failed if there is " +
+"no bad record") {
+sql("drop table IF EXISTS loadIssue")
+sql("""CREATE TABLE IF NOT EXISTS loadIssue(ID BigInt, item String) 
STORED BY 'carbondata'""")
+
CarbonProperties.getInstance().addProperty(CarbonCommonConstants.CARBON_BAD_RECORDS_ACTION,
+  "FAIL").addProperty(CarbonCommonConstants.LOAD_SORT_SCOPE, 
"GLOBAL_SORT")
+try {
+  sql("insert into loadIssue select 'x','Book'");
+} catch {
+  case ex: Exception =>
+assert(true)
+}
+try {
+  sql("insert into loadIssue select 1,'Book'");
+} catch {
+  case ex: Exception =>
+assert(false)
+} finally {
+  CarbonProperties.getInstance()
+.addProperty(CarbonCommonConstants.CARBON_BAD_RECORDS_ACTION, 
"FORCE")
+.addProperty(CarbonCommonConstants.LOAD_SORT_SCOPE, "LOCAL_SORT")
+}
+assert(true)
+sql("drop table IF EXISTS loadIssue")
+  }
+
--- End diff --

1. Use only one try and finally block inside test code..try should start 
from first line and end at last and finally should only reset the carbon 
property
2. Instead of try catch use intercept[Exception] only for the 1st case 
where exception is expected. For 2nd case try catch is not required...if 
exception comes anyways test case will fail


---


[GitHub] carbondata pull request #1732: [CARBONDATA-1946] Exception thrown after alte...

2018-01-03 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1732#discussion_r159585172
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/util/DataTypeUtil.java ---
@@ -775,4 +775,25 @@ public static DataType valueOf(String name) {
 }
   }
 
-}
\ No newline at end of file
+  /**
+   * Method to type case the data based on modified data type. This method 
will used for
+   * retrieving the data after change in data type restructure operation
+   *
+   * @param data
+   * @param restructuredDataType
+   * @param currentDataOffset
+   * @param length
+   * @return
+   */
+  public static long getDataBasedOnRestructuredDataType(byte[] data, 
DataType restructuredDataType,
--- End diff --

yes because alter datatype change is only for Int to BigInt and decimal 
from lower precision to higher precision. So float to double case is not 
required to be handled


---


[GitHub] carbondata pull request #1732: [CARBONDATA-1946] Exception thrown after alte...

2018-01-03 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/1732#discussion_r159585515
  
--- Diff: 
integration/spark2/src/test/scala/org/apache/spark/carbondata/restructure/vectorreader/ChangeDataTypeTestCases.scala
 ---
@@ -164,6 +164,21 @@ class ChangeDataTypeTestCases extends Spark2QueryTest 
with BeforeAndAfterAll {
 sql("drop table if exists PreAggMain_preagg1")
   }
 
+  test("test data type change for dictionary exclude INT type column") {
--- End diff --

Alter table datatype change does not support short to int, short to long 
and float to double. It only supports Int to BigInt and decimal from lower 
precision to higher precision. If any other data type change request comes 
validation is done and exception is thrown


---


[jira] [Resolved] (CARBONDATA-1975) Wrong input metrics displayed for carbon

2018-01-03 Thread Manish Gupta (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manish Gupta resolved CARBONDATA-1975.
--
   Resolution: Fixed
Fix Version/s: 1.3.0

> Wrong input metrics displayed for carbon
> 
>
> Key: CARBONDATA-1975
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1975
> Project: CarbonData
>  Issue Type: Bug
>Reporter: dhatchayani
>Assignee: dhatchayani
>Priority: Minor
> Fix For: 1.3.0
>
> Attachments: beeline.JPG
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> Input metrics is updated twice. Record count is updated twice and it is 
> wrongly displayed in Spark UI



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] carbondata issue #1729: [CARBONDATA-1936][PARTITION] Corrected bad record an...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1729
  
retest this please


---


[GitHub] carbondata pull request #1754: [CARBONDATA-1975] Wrong input metrics display...

2018-01-03 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/1754


---


[GitHub] carbondata issue #1754: [CARBONDATA-1975] Wrong input metrics displayed for ...

2018-01-03 Thread manishgupta88
Github user manishgupta88 commented on the issue:

https://github.com/apache/carbondata/pull/1754
  
LGTM...please attach the test results snapshots when the jira and github 
issue is resolved


---


[GitHub] carbondata issue #1733: [CARBONDATA-1944][PARTITION]Special character like c...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1733
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2691/



---


[jira] [Updated] (CARBONDATA-1975) Wrong input metrics displayed for carbon

2018-01-03 Thread dhatchayani (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

dhatchayani updated CARBONDATA-1975:

 Attachment: beeline.JPG
Description: 
Input metrics is updated twice. Record count is updated twice and it is wrongly 
displayed in Spark UI



  was:Input metrics is updated twice. Record count is updated twice and it is 
wrongly displayed in Spark UI


> Wrong input metrics displayed for carbon
> 
>
> Key: CARBONDATA-1975
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1975
> Project: CarbonData
>  Issue Type: Bug
>Reporter: dhatchayani
>Assignee: dhatchayani
>Priority: Minor
> Attachments: beeline.JPG
>
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> Input metrics is updated twice. Record count is updated twice and it is 
> wrongly displayed in Spark UI



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] carbondata issue #1718: [CARBONDATA-1929][Validation]carbon property configu...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1718
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2532/



---


[GitHub] carbondata issue #1733: [CARBONDATA-1944][PARTITION]Special character like c...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1733
  
Build Success with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1303/



---


[GitHub] carbondata issue #1729: [CARBONDATA-1936][PARTITION] Corrected bad record an...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1729
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2690/



---


[GitHub] carbondata pull request #1760: [CARBONDATA-1979] ][IMPLICIT COLUMN] Modified...

2018-01-03 Thread rahulforallp
GitHub user rahulforallp opened a pull request:

https://github.com/apache/carbondata/pull/1760

[CARBONDATA-1979] ][IMPLICIT COLUMN] Modified implicit column filtering 
logic to directly validate the blo…

…cklet ID as after driver pruning valid blocklets are identified and 
tasks are divided based on blocklets

Be sure to do all of the following checklist to help us incorporate 
your contribution quickly and easily:

 - [ ] Any interfaces changed?
 
 - [ ] Any backward compatibility impacted?
 
 - [ ] Document update required?

 - [ ] Testing done
Please provide details on 
- Whether new unit test cases have been added or why no new tests 
are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance 
test report.
- Any additional information to help reviewers in testing this 
change.
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/rahulforallp/incubator-carbondata 
implicit_column

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/1760.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #1760


commit 4df91bdf5608c280cd4c3f3320ddcb957aa496f2
Author: rahulforallp 
Date:   2018-01-03T09:51:37Z

Modified implicit column filtering logic to directly validate the blocklet 
ID as after driver pruning valid blocklets are identified and tasks are divided 
based on blocklets




---


[GitHub] carbondata issue #1718: [CARBONDATA-1929][Validation]carbon property configu...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1718
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2689/



---


[GitHub] carbondata issue #1753: [CARBONDATA-1970](Carbon1.3.0 - Spark 2.2) Use Spark...

2018-01-03 Thread zzcclp
Github user zzcclp commented on the issue:

https://github.com/apache/carbondata/pull/1753
  
The discussion about this on mailing list : 
[Should-we-use-Spark-2-2-1-as-default-version-for-Spark-2-2-supported](http://apache-carbondata-dev-mailing-list-archive.1130556.n5.nabble.com/Should-we-use-Spark-2-2-1-as-default-version-for-Spark-2-2-supported-td32713.html)


---


[GitHub] carbondata issue #1650: [CARBONDATA-1703] Refactored code for creation of fi...

2018-01-03 Thread sraghunandan
Github user sraghunandan commented on the issue:

https://github.com/apache/carbondata/pull/1650
  
retest sdv please


---


[GitHub] carbondata issue #1753: [CARBONDATA-1970](Carbon1.3.0 - Spark 2.2) Use Spark...

2018-01-03 Thread chenliang613
Github user chenliang613 commented on the issue:

https://github.com/apache/carbondata/pull/1753
  
please add mailing list discussion link to description.


---


[GitHub] carbondata issue #1718: [CARBONDATA-1929][Validation]carbon property configu...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1718
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2688/



---


[GitHub] carbondata issue #1678: [CARBONDATA-1903] Fix code issues in carbondata

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1678
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2531/



---


[GitHub] carbondata issue #1759: [CARBONDATA-1980][PARTITION]Fix adding partition inf...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1759
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2530/



---


[GitHub] carbondata issue #1748: [CARBONDATA-1967][PARTITION] Fix autocompaction and ...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1748
  
Build Success with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1302/



---


[GitHub] carbondata issue #1758: [CARBONDATA-1978] Handled preaggregate issues with h...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1758
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2687/



---


[GitHub] carbondata issue #1756: [HOTFIX] Remove unnecessary object creation in write...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1756
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2686/



---


[jira] [Resolved] (CARBONDATA-1903) Fix some code issues in carbondata

2018-01-03 Thread QiangCai (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-1903?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

QiangCai resolved CARBONDATA-1903.
--
   Resolution: Fixed
Fix Version/s: 1.3.0

> Fix some code issues in carbondata
> --
>
> Key: CARBONDATA-1903
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1903
> Project: CarbonData
>  Issue Type: Improvement
>Reporter: xuchuanyin
>Assignee: xuchuanyin
>Priority: Minor
> Fix For: 1.3.0
>
>  Time Spent: 12h 10m
>  Remaining Estimate: 0h
>
> Remove useless code in carbondata



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] carbondata pull request #1678: [CARBONDATA-1903] Fix code issues in carbonda...

2018-01-03 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/1678


---


[GitHub] carbondata issue #1718: [CARBONDATA-1929][Validation]carbon property configu...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1718
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2529/



---


[GitHub] carbondata issue #1752: [CARBONDATA-1972][PARTITION] Fix compaction after up...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1752
  
Build Success with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1301/



---


[GitHub] carbondata issue #1678: [CARBONDATA-1903] Fix code issues in carbondata

2018-01-03 Thread QiangCai
Github user QiangCai commented on the issue:

https://github.com/apache/carbondata/pull/1678
  
LGTM


---


[GitHub] carbondata issue #1757: [CARBONDATA-1977][PARTITION] Fix aggregation table l...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1757
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2685/



---


[GitHub] carbondata issue #1678: [CARBONDATA-1903] Fix code issues in carbondata

2018-01-03 Thread xuchuanyin
Github user xuchuanyin commented on the issue:

https://github.com/apache/carbondata/pull/1678
  
retest this please


---


[GitHub] carbondata issue #1729: [CARBONDATA-1936][PARTITION] Corrected bad record an...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1729
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2528/



---


[GitHub] carbondata issue #1754: [CARBONDATA-1975] Wrong input metrics displayed for ...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1754
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2684/



---


[GitHub] carbondata issue #1759: [CARBONDATA-1980][PARTITION]Fix adding partition inf...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1759
  
Build Failed with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1300/



---


[GitHub] carbondata issue #1733: [CARBONDATA-1944][PARTITION]Special character like c...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1733
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2527/



---


[GitHub] carbondata issue #1755: [CARBONDATA-1976][PARTITION] Support combination of ...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1755
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2683/



---


[GitHub] carbondata issue #1678: [CARBONDATA-1903] Fix code issues in carbondata

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1678
  
Build Success with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1299/



---


[GitHub] carbondata issue #1748: [CARBONDATA-1967][PARTITION] Fix autocompaction and ...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1748
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2526/



---


[GitHub] carbondata issue #1713: [WIP] [CARBONDATA-1899] Optimize CarbonData concurre...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1713
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2682/



---


[GitHub] carbondata issue #1713: [WIP] [CARBONDATA-1899] Optimize CarbonData concurre...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1713
  
Build Success with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1298/



---


[GitHub] carbondata issue #1752: [CARBONDATA-1972][PARTITION] Fix compaction after up...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1752
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2525/



---


[GitHub] carbondata issue #1743: [CARBONDATA-1954] [Pre-Aggregate] CarbonHiveMetastor...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1743
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2681/



---


[GitHub] carbondata issue #1718: [CARBONDATA-1929][Validation]carbon property configu...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1718
  
Build Success with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1297/



---


[GitHub] carbondata issue #1754: [CARBONDATA-1975] Wrong input metrics displayed for ...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1754
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2524/



---


[GitHub] carbondata issue #1747: [Compatibility] Added changes for backward compatibi...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1747
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2680/



---


[GitHub] carbondata issue #1743: [CARBONDATA-1954] [Pre-Aggregate] CarbonHiveMetastor...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1743
  
Build Success with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1296/



---


[GitHub] carbondata issue #1755: [CARBONDATA-1976][PARTITION] Support combination of ...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1755
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2523/



---


[GitHub] carbondata issue #1697: [CARBONDATA-1719][Pre-Aggregate][Bug] Fixed bug to h...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1697
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2679/



---


[GitHub] carbondata issue #1747: [Compatibility] Added changes for backward compatibi...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1747
  
Build Success with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1295/



---


[GitHub] carbondata issue #1756: [HOTFIX] Remove unnecessary object creation in write...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1756
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2522/



---


[GitHub] carbondata issue #1702: [WIP][CARBONDATA-1896] Clean files operation improve...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1702
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2678/



---


[GitHub] carbondata issue #1754: [CARBONDATA-1975] Wrong input metrics displayed for ...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1754
  
Build Success with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1294/



---


[GitHub] carbondata issue #1757: [CARBONDATA-1977][PARTITION] Fix aggregation table l...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1757
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2521/



---


[GitHub] carbondata issue #1758: [CARBONDATA-1978] Handled preaggregate issues with h...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1758
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2520/



---


[GitHub] carbondata issue #1718: [CARBONDATA-1929][Validation]carbon property configu...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1718
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2677/



---


[GitHub] carbondata issue #1755: [CARBONDATA-1976][PARTITION] Support combination of ...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1755
  
Build Failed with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1293/



---


[GitHub] carbondata issue #1116: [CARBONDATA-1249] Wrong order of columns in redirect...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1116
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2519/



---


[GitHub] carbondata pull request #1759: [CARBONDATA-1980][PARTITION]Fix adding partit...

2018-01-03 Thread ravipesala
GitHub user ravipesala opened a pull request:

https://github.com/apache/carbondata/pull/1759

[CARBONDATA-1980][PARTITION]Fix adding partition information while doing 
refresh(restore) table.And fix the case sensitivity issue of partition column

Problem 
While restoring the table the partitions are not added to the hive.  And 
also if any uppercase letters are present in partition column then filter 
queries on that column gives error.

Solution 
Read all partitionmap files of all segments to add back the partition 
information to the hive metatstore. 
And always update the partition filter columns to lower case before 
querying partition information.

Be sure to do all of the following checklist to help us incorporate 
your contribution quickly and easily:

 - [X] Any interfaces changed? NO
 
 - [X] Any backward compatibility impacted? NO
 
 - [X] Document update required? NO

 - [X] Testing done
   Tests added
   
 - [X] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/ravipesala/incubator-carbondata 
restore-partition

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/1759.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #1759


commit 20e411487a11c2219e08e966ff83c6cc0e718678
Author: ravipesala 
Date:   2018-01-03T19:04:33Z

Fix adding partition information while doing refresh(restore) table. And 
fix the case sensitivity issue of partition column




---


[GitHub] carbondata issue #1756: [HOTFIX] Remove unnecessary object creation in write...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1756
  
Build Success with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1292/



---


[jira] [Created] (CARBONDATA-1980) Partition information is added while restore or refresh the table. And also query is not working if there is nay upper case letter in partition column.

2018-01-03 Thread Ravindra Pesala (JIRA)
Ravindra Pesala created CARBONDATA-1980:
---

 Summary: Partition information is added while restore or refresh 
the table. And also query is not working if there is nay upper case letter in 
partition column.
 Key: CARBONDATA-1980
 URL: https://issues.apache.org/jira/browse/CARBONDATA-1980
 Project: CarbonData
  Issue Type: Bug
Reporter: Ravindra Pesala


Partition information is added while restore or refresh the table. And also 
query is not working if there is nay upper case letter in partition column.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] carbondata issue #1678: [CARBONDATA-1903] Fix code issues in carbondata

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1678
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2518/



---


[GitHub] carbondata issue #1697: [CARBONDATA-1719][Pre-Aggregate][Bug] Fixed bug to h...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1697
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2676/



---


[GitHub] carbondata issue #1757: [CARBONDATA-1977][PARTITION] Fix aggregation table l...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1757
  
Build Failed with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1291/



---


[GitHub] carbondata issue #1758: [CARBONDATA-1978] Handled preaggregate issues with h...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1758
  
Build Success with Spark 2.2.0, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1290/



---


[GitHub] carbondata issue #1697: [CARBONDATA-1719][Pre-Aggregate][Bug] Fixed bug to h...

2018-01-03 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/1697
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2517/



---


[GitHub] carbondata issue #1697: [CARBONDATA-1719][Pre-Aggregate][Bug] Fixed bug to h...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1697
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2675/



---


[GitHub] carbondata issue #1728: [CARBONDATA-1926][CARBONDATA-1927][Pre-Aggregate] Ex...

2018-01-03 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1728
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/2674/



---


  1   2   3   >