[GitHub] carbondata issue #2453: [CARBONDATA-2528][MV] Fixed order by in mv and aggre...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2453
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5875/



---


[GitHub] carbondata issue #2423: [CARBONDATA-2530][MV] Fix wrong data displayed when ...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2423
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7104/



---


[GitHub] carbondata issue #2431: [MINOR] Adding a testcase for stream-table join in S...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2431
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7103/



---


[GitHub] carbondata issue #2489: [CARBONDATA-2606][Complex DataType Enhancements]Fix ...

2018-07-13 Thread kunal642
Github user kunal642 commented on the issue:

https://github.com/apache/carbondata/pull/2489
  
LGTM


---


[GitHub] carbondata issue #2455: [CARBONDATA-2512][32k] Support writing longstring th...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2455
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5874/



---


[jira] [Created] (CARBONDATA-2743) [MV] After MV creation limit queries Throws Exceptions for table which does not have mv data map

2018-07-13 Thread Babulal (JIRA)
Babulal created CARBONDATA-2743:
---

 Summary: [MV] After MV creation limit queries Throws Exceptions 
for table which does not have mv data map
 Key: CARBONDATA-2743
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2743
 Project: CarbonData
  Issue Type: Bug
Reporter: Babulal


0: jdbc:hive2://10.18.16.173:23040/default> create table mytest_50_s13 (name 
string,rownumber string, m1 float) stored by 'carbondata' 
TBLPROPERTIES('sort_scope'='global_sort');
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (12.209 seconds)
0: jdbc:hive2://10.18.16.173:23040/default> load data inpath 
'hdfs://hacluster/tmp/data/cbo_1.csv' into table mytest_50_s13 
options('FILEHEADER'='name,rownumber,m1');
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (79.901 seconds)
0: jdbc:hive2://10.18.16.173:23040/default> create datamap map10 using 'mv' as 
select sum(m1),rownumber from mytest_50_s13 group by rownumber;
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (17.05 seconds)

0: jdbc:hive2://10.18.16.173:23040/default> show datamap on table mytest_50_s13;
+--++---+-+--+
| DataMapName  | ClassName  | Associated Table  | DataMap Properties  |
+--++---+-+--+
| map10    | mv | babu.map10_table  | |
+--++---+-+--+
1 row selected (0.815 seconds)



Now create one more table without mv

: jdbc:hive2://10.18.16.173:23040/default> create table mytest_50_s14 (name 
string,rownumber string, m1 float) stored by 'carbondata' 
TBLPROPERTIES('sort_scope'='global_sort');
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (12.209 seconds)
0: jdbc:hive2://10.18.16.173:23040/default> load data inpath 
'hdfs://hacluster/tmp/data/cbo_1.csv'' into table mytest_50_s14 
options('FILEHEADER'='name,rownumber,m1');
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (79.901 seconds)
0: jdbc:hive2://10.18.16.173:23040/default> select * from mytest_50_s14 limit 
10;
Error: java.lang.UnsupportedOperationException: unsupported operation: Modular 
plan not supported (e.g. has subquery expression) for
GlobalLimit 10
+- LocalLimit 10
   +- Relation[name#1026,rownumber#1027,m1#1028] CarbonDatasourceHadoopRelation 
[ Database name :babu, Table name :mytest_50_s14, Schema 
:Some(StructType(StructField(name,StringType,true), 
StructField(rownumber,StringType,true), StructField(m1,DoubleType,true))) ] 
(state=,code=0)





2018-07-13 00:42:51,540 | INFO  | [pool-25-thread-32] | 
OperationId=b5c2c8b2-1ef4-4894-a709-2a738bd81f76    Result=FAIL | 
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:280)
2018-07-13 00:42:51,540 | ERROR | [pool-25-thread-32] | Error executing query, 
currentState RUNNING,  | 
org.apache.spark.internal.Logging$class.logError(Logging.scala:91)
java.lang.UnsupportedOperationException: unsupported operation: Modular plan 
not supported (e.g. has subquery expression) for
GlobalLimit 10
+- LocalLimit 10
   +- Relation[name#1026,rownumber#1027,m1#1028] CarbonDatasourceHadoopRelation 
[ Database name :babu, Table name :mytest_50_s14, Schema 
:Some(StructType(StructField(name,StringType,true), 
StructField(rownumber,StringType,true), StructField(m1,DoubleType,true))) ]

    at org.apache.carbondata.mv.plans.package$.supports(package.scala:52)
    at 
org.apache.carbondata.mv.plans.modular.Modularizer.org$apache$carbondata$mv$plans$modular$Modularizer$$modularizeCore(Modularizer.scala:102)
    at 
org.apache.carbondata.mv.plans.modular.Modularizer.modularize(Modularizer.scala:65)
    at 
org.apache.carbondata.mv.rewrite.QueryRewrite.modularPlan$lzycompute(QueryRewrite.scala:50)
    at 
org.apache.carbondata.mv.rewrite.QueryRewrite.modularPlan(QueryRewrite.scala:49)
    at 
org.apache.carbondata.mv.rewrite.QueryRewrite.withSummaryData$lzycompute(QueryRewrite.scala:53)
    at 
org.apache.carbondata.mv.rewrite.QueryRewrite.withSummaryData(QueryRewrite.scala:52)
    at 
org.apache.carbondata.mv.rewrite.QueryRewrite.withMVTable$lzycompute(QueryRewrite.scala:55)
    at 
org.apache.carbondata.mv.rewrite.QueryRewrite.withMVTable(QueryRewrite.scala:55)
    at 
org.apache.carbondata.mv.datamap.MVAnalyzerRule.apply(MVAnalyzerRule.scala:68)
    at 
org.apache.carbondata.mv.datamap.MVAnalyzerRule.apply(MVAnalyzerRule.scala:38)
    at 
org.apache.spark.sql.hive.CarbonAnalyzer.execute(CarbonAnalyzer.scala:46)
    at 
org.apache.spark.sql.hive.CarbonAnalyzer.execute(CarbonAnalyzer.scala:27)
    at 
org.apache.spark.sql.execution.Que

[jira] [Created] (CARBONDATA-2742) [MV] Wrong data displayed after MV creation.

2018-07-13 Thread Babulal (JIRA)
Babulal created CARBONDATA-2742:
---

 Summary: [MV] Wrong data displayed  after MV creation. 
 Key: CARBONDATA-2742
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2742
 Project: CarbonData
  Issue Type: Bug
Reporter: Babulal


0: jdbc:hive2://10.18.16.173:23040/default> create table mytest_48 (rownumber 
int,name string, m1 int) stored by 'carbondata';
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (1.267 seconds)
0: jdbc:hive2://10.18.16.173:23040/default> load data inpath 
'hdfs://hacluster/tmp/babu/testdata_1.csv' into table mytest_48 ;
+-+--+
| Result  |
+-+--+
+-+–+

 

0: jdbc:hive2://10.18.16.173:23040/default> show datamap  on table mytest_48;
+--++---+-+--+
| DataMapName  | ClassName  | Associated Table  | DataMap Properties  |
+--++---+-+--+
+--++---+-+--+
No rows selected (0.162 seconds)
0: jdbc:hive2://10.18.16.173:23040/default>
0: jdbc:hive2://10.18.16.173:23040/default> select * from mytest_48;
++---+--+--+
| rownumber  | name  |    m1    |
++---+--+--+
| 1  | aaa   | 1000 |
| 2  | aaa   | 65000    |
| 3  | aaa   | 100  |
| 1  | ddd   | 1000 |
| 2  | ddd   | 65000    |
| 3  | ddd   | 100  |
++---+--+--+
6 rows selected (1.266 seconds)
0: jdbc:hive2://10.18.16.173:23040/default> create datamap map9 using 'mv' as 
select sum(m1),name from mytest_48 group by name;
+-+--+
| Result  |
+-+--+
+-+--+
No rows selected (0.82 seconds)
0: jdbc:hive2://10.18.16.173:23040/default> select sum(m1),name from mytest_48 
group by name;
+--+---+--+
| sum(m1)  | name  |
+--+---+--+
+--+---+--+
No rows selected (2.615 seconds)


0: jdbc:hive2://10.18.16.173:23040/default> explain select sum(m1),name from 
mytest_48 group by name;
+-+--+
|   


   plan 


  |
+-+--+
| == CarbonData Profiler ==





  |
| == Physical Plan ==
*HashAggregate(keys=[mytest_48_name#297], functions=[sum(sum_m1#296L)])
+- Exchange hashpartitioning(mytest_48_name#297, 200)
   +- *HashAggregate(keys=[mytest_48_name#297], 
functions=[partial_sum(sum_m1#296L)])
  +- *BatchedScan CarbonDatasourceHadoopRelation [ Database name :babu, 
Table name :map9_table, Schema 
:Some(StructType(StructField(sum_m1,LongType,true), 
StructField(mytest_48_name,StringType,true))) ] 
babu.map9_table[sum_m1#296L,mytest_48_name#297]  |
+

[GitHub] carbondata issue #2504: [CARBONDATA-2740]flat folder handling for implicit c...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2504
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/5826/



---


[jira] [Created] (CARBONDATA-2741) Exception occurs after alter add few columns and selecting in random order

2018-07-13 Thread Jatin (JIRA)
Jatin created CARBONDATA-2741:
-

 Summary: Exception occurs after alter add few columns and 
selecting in random order
 Key: CARBONDATA-2741
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2741
 Project: CarbonData
  Issue Type: Bug
  Components: spark-integration
Affects Versions: 1.5.0
 Environment: 3 node cluster with spark2.2
Reporter: Jatin
Assignee: Jatin
 Fix For: 1.5.0


create table tb1 (imei string,AMSize string,channelsId string,ActiveCountry 
string, Activecity string,gamePointId double,deviceInformationId 
double,productionDate Timestamp,deliveryDate timestamp,deliverycharge double) 
STORED BY 'org.apache.carbondata.format' 
TBLPROPERTIES('table_blocksize'='1','COLUMN_META_CACHE'='AMSize');

 LOAD DATA INPATH 'hdfs://hacluster/csv/vardhandaterestruct.csv' INTO TABLE tb1 
OPTIONS('DELIMITER'=',', 'QUOTECHAR'= 
'"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'= 
'imei,deviceInformationId,AMSize,channelsId,ActiveCountry,Activecity,gamePointId,productionDate,deliveryDate,deliverycharge');

alter table tb1 add columns(age int, name string);

select * from tb1 where name is NULL or channelsId =4;

Exception occurs :

*Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 
0 in stage 6508.0 failed 4 times, most recent failure: Lost task 0.3 in stage 
6508.0 (TID 140476, linux-49, executor 3): java.lang.RuntimeException: internal 
error:* 
org.apache.carbondata.core.datastore.page.encoding.adaptive.AdaptiveFloatingCodec[src
 type: DOUBLE, target type: INT, stats(min: 1.0, max: 100.0, decimal: 1 )]
 at 
org.apache.carbondata.core.datastore.page.encoding.adaptive.AdaptiveFloatingCodec$3.decodeLong(AdaptiveFloatingCodec.java:185)
 at 
org.apache.carbondata.core.datastore.page.LazyColumnPage.getLong(LazyColumnPage.java:64)
 at 
org.apache.carbondata.core.scan.result.vector.MeasureDataVectorProcessor$IntegralMeasureVectorFiller.fillMeasureVector(MeasureDataVectorProcessor.java:73)
 at 
org.apache.carbondata.core.scan.result.impl.FilterQueryScannedResult.fillColumnarMeasureBatch(FilterQueryScannedResult.java:129)
 at 
org.apache.carbondata.core.scan.collector.impl.DictionaryBasedVectorResultCollector.fillResultToColumnarBatch(DictionaryBasedVectorResultCollector.java:167)
 at 
org.apache.carbondata.core.scan.collector.impl.RestructureBasedVectorResultCollector.collectResultInColumnarBatch(RestructureBasedVectorResultCollector.java:127)
 at 
org.apache.carbondata.core.scan.processor.DataBlockIterator.processNextBatch(DataBlockIterator.java:245)
 at 
org.apache.carbondata.core.scan.result.iterator.VectorDetailQueryResultIterator.processNextBatch(VectorDetailQueryResultIterator.java:48)
 at 
org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextBatch(VectorizedCarbonRecordReader.java:290)
 at 
org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.nextKeyValue(VectorizedCarbonRecordReader.java:180)
 at 
org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.hasNext(CarbonScanRDD.scala:497)
 at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.scan_nextBatch$(Unknown
 Source)
 at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
 Source)
 at 
org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
 at 
org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:381)
 at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown
 Source)
 at 
org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
 at 
org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:381)
 at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:231)
 at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:225)
 at 
org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:828)
 at 
org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:828)
 at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
 at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:324)
 at org.apache.spark.rdd.RDD.iterator(RDD.scala:288)
 at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
 at org.apache.spark.scheduler.Task.run(Task.scala:99)
 at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:325)
 at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
 at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
 at java.lang.Thread.run(Thread.java:748)

Driver stacktrace: (state=,code=0)

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2441: [CARBONDATA-2625] optimize CarbonReader performance

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2441
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7102/



---


[GitHub] carbondata issue #2448: [HotFix] Getting carbon table identifier to datamap ...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2448
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7101/



---


[jira] [Resolved] (CARBONDATA-2721) [SDK] [JsonWriter] NPE when schema and data are not of same length or Data is null.

2018-07-13 Thread Ravindra Pesala (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2721?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravindra Pesala resolved CARBONDATA-2721.
-
   Resolution: Fixed
Fix Version/s: 1.4.1

> [SDK] [JsonWriter] NPE when schema and data are not of same length or Data is 
> null.
> ---
>
> Key: CARBONDATA-2721
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2721
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Ajantha Bhat
>Assignee: Ajantha Bhat
>Priority: Major
> Fix For: 1.4.1
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (CARBONDATA-2722) [SDK] [JsonWriter] Json writer is writing only first element of an array and discarding the rest of the elements

2018-07-13 Thread Ravindra Pesala (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2722?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravindra Pesala resolved CARBONDATA-2722.
-
   Resolution: Fixed
Fix Version/s: 1.4.1

> [SDK] [JsonWriter] Json writer is writing only first element of an array and 
> discarding the rest of the elements
> 
>
> Key: CARBONDATA-2722
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2722
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Ajantha Bhat
>Assignee: Ajantha Bhat
>Priority: Major
> Fix For: 1.4.1
>
>  Time Spent: 2h 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2456: [CARBONDATA-2694][32k] Show longstring table propert...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2456
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5873/



---


[GitHub] carbondata pull request #2485: [CARBONDATA-2722] [CARBONDATA-2721] JsonWrite...

2018-07-13 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2485


---


[GitHub] carbondata issue #2485: [CARBONDATA-2722] [CARBONDATA-2721] JsonWriter issue...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2485
  
LGTM


---


[GitHub] carbondata issue #2504: [CARBONDATA-2740]flat folder handling for implicit c...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2504
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/5825/



---


[GitHub] carbondata issue #2501: [CARBONDATA-2738]Block Preaggregate, Dictionary Excl...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2501
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/5824/



---


[GitHub] carbondata issue #2462: [CARBONDATA-2704] Index file size in describe format...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2462
  
Build Failed with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5872/



---


[GitHub] carbondata issue #2452: [CARBONDATA-2693][BloomDataMap]Fix bug for alter ren...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2452
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7100/



---


[GitHub] carbondata issue #2506: [CARBONDATA-2682][32K] fix create table with long_st...

2018-07-13 Thread xuchuanyin
Github user xuchuanyin commented on the issue:

https://github.com/apache/carbondata/pull/2506
  
retest this please


---


[GitHub] carbondata issue #2505: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA-2732][...

2018-07-13 Thread xuchuanyin
Github user xuchuanyin commented on the issue:

https://github.com/apache/carbondata/pull/2505
  
retest this please


---


[GitHub] carbondata issue #2453: [CARBONDATA-2528][MV] Fixed order by in mv and aggre...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2453
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7099/



---


[GitHub] carbondata issue #2463: [CARBONDATA-2708][BloomDataMap] clear index file in ...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2463
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5871/



---


[GitHub] carbondata issue #2482: [CARBONDATA-2714] Support merge index files for the ...

2018-07-13 Thread dhatchayani
Github user dhatchayani commented on the issue:

https://github.com/apache/carbondata/pull/2482
  
retest this please


---


[GitHub] carbondata issue #2485: [CARBONDATA-2722] [CARBONDATA-2721] JsonWriter issue...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2485
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/5823/



---


[GitHub] carbondata issue #2455: [CARBONDATA-2512][32k] Support writing longstring th...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2455
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7098/



---


[GitHub] carbondata issue #2464: [CARBONDATA-2618][32K] Split to multiple pages if va...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2464
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5870/



---


[GitHub] carbondata issue #2456: [CARBONDATA-2694][32k] Show longstring table propert...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2456
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7097/



---


[GitHub] carbondata issue #2465: [WIP] Refactored CarbonFile interface

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2465
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5869/



---


[GitHub] carbondata issue #2503: [CARBONDATA-2734] Update is not working on the table...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2503
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/5822/



---


[GitHub] carbondata issue #2462: [CARBONDATA-2704] Index file size in describe format...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2462
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7096/



---


[GitHub] carbondata pull request #2504: [CARBONDATA-2740]flat folder handling for imp...

2018-07-13 Thread KanakaKumar
Github user KanakaKumar commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2504#discussion_r202329272
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/hive/CarbonRelation.scala
 ---
@@ -173,15 +172,38 @@ case class CarbonRelation(
 .getValidAndInvalidSegments.getValidSegments.asScala
   var size = 0L
   // for each segment calculate the size
-  segments.foreach {validSeg =>
-// for older store
-if (null != validSeg.getLoadMetadataDetails.getDataSize &&
-null != validSeg.getLoadMetadataDetails.getIndexSize) {
-  size = size + 
validSeg.getLoadMetadataDetails.getDataSize.toLong +
- validSeg.getLoadMetadataDetails.getIndexSize.toLong
-} else {
-  size = size + FileFactory.getDirectorySize(
-CarbonTablePath.getSegmentPath(tablePath, 
validSeg.getSegmentNo))
+  if 
(carbonTable.getTableInfo.getFactTable.getTableProperties.asScala
+.get(CarbonCommonConstants.FLAT_FOLDER).isDefined &&
+  
carbonTable.getTableInfo.getFactTable.getTableProperties.asScala
+  (CarbonCommonConstants.FLAT_FOLDER).toBoolean) {
+val tableDirectorySize = 
FileFactory.getDirectorySize(carbonTable.getTablePath)
+val metaDirectorySize = 
FileFactory.getDirectorySize(carbonTable.getMetadataPath)
+val factDirectorySize = FileFactory
+  
.getDirectorySize(CarbonTablePath.getFactDir(carbonTable.getTablePath))
+val lockDirSize = FileFactory
+  
.getDirectorySize(CarbonTablePath.getLockFilesDirPath(carbonTable.getTablePath))
+val datamaps = 
DataMapStoreManager.getInstance().getAllDataMap(carbonTable)
+var datamapsDirectorySize = 0L
+if (datamaps.size() > 0) {
+  datamaps.asScala.foreach { datamap =>
+datamapsDirectorySize = datamapsDirectorySize + FileFactory
+  .getDirectorySize(
+carbonTable.getTablePath + 
datamap.getDataMapSchema.getDataMapName)
--- End diff --

Please use a utility method to make to dm path like 
org.apache.carbondata.core.util.path.CarbonTablePath#getDataMapStorePath.


---


[GitHub] carbondata issue #2452: [CARBONDATA-2693][BloomDataMap]Fix bug for alter ren...

2018-07-13 Thread ndwangsen
Github user ndwangsen commented on the issue:

https://github.com/apache/carbondata/pull/2452
  
retest this please


---


[GitHub] carbondata issue #2463: [CARBONDATA-2708][BloomDataMap] clear index file in ...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2463
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7095/



---


[GitHub] carbondata issue #2472: [CARBONDATA-2717] fixed table id empty problem while...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2472
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5867/



---


[GitHub] carbondata issue #2466: [CARBONDATA-2710][Spark Integration] Refactor Carbon...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2466
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5868/



---


[jira] [Updated] (CARBONDATA-2717) Table id is empty when taking drop lock which causes failure

2018-07-13 Thread Venkata Ramana G (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2717?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Venkata Ramana G updated CARBONDATA-2717:
-
Issue Type: Bug  (was: Improvement)

> Table id is empty when taking drop lock which causes failure
> 
>
> Key: CARBONDATA-2717
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2717
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Kunal Kapoor
>Assignee: Kunal Kapoor
>Priority: Minor
> Fix For: 1.4.1
>
>  Time Spent: 3h 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2488: [CARBONDATA-2724][DataMap]Unsupported create datamap...

2018-07-13 Thread ndwangsen
Github user ndwangsen commented on the issue:

https://github.com/apache/carbondata/pull/2488
  
retest sdv please



---


[jira] [Resolved] (CARBONDATA-2717) Table id is empty when taking drop lock which causes failure

2018-07-13 Thread Venkata Ramana G (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2717?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Venkata Ramana G resolved CARBONDATA-2717.
--
   Resolution: Fixed
Fix Version/s: 1.4.1

> Table id is empty when taking drop lock which causes failure
> 
>
> Key: CARBONDATA-2717
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2717
> Project: CarbonData
>  Issue Type: Bug
>Reporter: Kunal Kapoor
>Assignee: Kunal Kapoor
>Priority: Minor
> Fix For: 1.4.1
>
>  Time Spent: 3h 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2472: [CARBONDATA-2717] fixed table id empty proble...

2018-07-13 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2472


---


[GitHub] carbondata issue #2489: [CARBONDATA-2606][Complex DataType Enhancements]Fix ...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2489
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/5821/



---


[GitHub] carbondata issue #2472: [CARBONDATA-2717] fixed table id empty problem while...

2018-07-13 Thread gvramana
Github user gvramana commented on the issue:

https://github.com/apache/carbondata/pull/2472
  
LGTM


---


[GitHub] carbondata pull request #2487: [CARBONDATA-2735]Fixed Performance issue for ...

2018-07-13 Thread gvramana
Github user gvramana commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2487#discussion_r202321022
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/datastore/page/VarLengthColumnPageBase.java
 ---
@@ -201,31 +209,32 @@ private static ColumnPage 
getComplexLVBytesColumnPage(TableSpec.ColumnSpec colum
   throws MemoryException {
 // extract length and data, set them to rowOffset and unsafe memory 
correspondingly
 int rowId = 0;
-List rowOffset = new ArrayList<>();
-List rowLength = new ArrayList<>();
+TableSpec.ColumnSpec spec = TableSpec.ColumnSpec
+.newInstance(columnSpec.getFieldName(), DataTypes.INT, 
ColumnType.MEASURE);
+ColumnPage rowOffset = ColumnPage.newPage(spec, DataTypes.INT, 1024);
--- End diff --

This default can be equal NUMBER_OF_ROWS_PER_BLOCKLET_COLUMN_PAGE_DEFAULT


---


[GitHub] carbondata pull request #2487: [CARBONDATA-2735]Fixed Performance issue for ...

2018-07-13 Thread gvramana
Github user gvramana commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2487#discussion_r202320940
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/datastore/page/VarLengthColumnPageBase.java
 ---
@@ -175,24 +182,25 @@ private static ColumnPage 
getLVBytesColumnPage(TableSpec.ColumnSpec columnSpec,
   throws MemoryException {
 // extract length and data, set them to rowOffset and unsafe memory 
correspondingly
 int rowId = 0;
-List rowOffset = new ArrayList<>();
-List rowLength = new ArrayList<>();
+TableSpec.ColumnSpec spec = TableSpec.ColumnSpec
+.newInstance(columnSpec.getFieldName(), DataTypes.INT, 
ColumnType.MEASURE);
+ColumnPage rowOffset = ColumnPage.newPage(spec, DataTypes.INT, 1024);
--- End diff --

This default can be equal NUMBER_OF_ROWS_PER_BLOCKLET_COLUMN_PAGE_DEFAULT


---


[GitHub] carbondata pull request #2487: [CARBONDATA-2735]Fixed Performance issue for ...

2018-07-13 Thread gvramana
Github user gvramana commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2487#discussion_r202319613
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/datastore/page/SafeVarLengthColumnPage.java
 ---
@@ -40,6 +40,7 @@
   @Override
   public void freeMemory() {
 byteArrayData = null;
+rowOffset.freeMemory();
--- End diff --

rowOffset.freeMemory also required in SafeDecimalColumnPage.
It will we good practice to call rowOffset.freeMemory  in same class it is 
allocated, call abstract freememory class from derived to ensure any clean up 
from abstract class


---


[GitHub] carbondata issue #2464: [CARBONDATA-2618][32K] Split to multiple pages if va...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2464
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7094/



---


[GitHub] carbondata issue #2465: [WIP] Refactored CarbonFile interface

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2465
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7093/



---


[GitHub] carbondata issue #2307: [CARBONDATA-2482] Pass uuid while writing segment fi...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2307
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/5820/



---


[GitHub] carbondata issue #2482: [CARBONDATA-2714] Support merge index files for the ...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2482
  
Build Failed with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5866/



---


[GitHub] carbondata pull request #2307: [CARBONDATA-2482] Pass uuid while writing seg...

2018-07-13 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2307#discussion_r202310789
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/writer/CarbonIndexFileMergeWriter.java
 ---
@@ -145,9 +145,10 @@ private String writeMergeIndexFileBasedOnSegmentFile(
   }
 }
 
-String uniqueId = String.valueOf(System.currentTimeMillis());
-String newSegmentFileName =
-SegmentFileStore.genSegmentFileName(segmentId, 
String.valueOf(uniqueId))
+String uniqueId = (uuid.endsWith(CarbonTablePath.SEGMENT_EXT)) ?
--- End diff --

Don't pass segment file name, just pass only uuid . 


---


[GitHub] carbondata issue #2484: [WIP] added hadoop conf to thread local

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2484
  
Build Failed with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5865/



---


[jira] [Resolved] (CARBONDATA-2723) Failed to recreate the table which has bloomfilter on it with same table name but different bloom index

2018-07-13 Thread Ravindra Pesala (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2723?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravindra Pesala resolved CARBONDATA-2723.
-
   Resolution: Fixed
Fix Version/s: 1.4.1

> Failed to recreate the table which has bloomfilter on it with same table name 
> but different bloom index
> ---
>
> Key: CARBONDATA-2723
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2723
> Project: CarbonData
>  Issue Type: Sub-task
>Reporter: xuchuanyin
>Assignee: xuchuanyin
>Priority: Major
> Fix For: 1.4.1
>
>  Time Spent: 3h 20m
>  Remaining Estimate: 0h
>
> Steps to reproduce:
> ```
> create table xu_t11(id int,name string,sal float) stored by 'carbondata' 
> tblproperties('sort_columns'='id')
> create datamap xu_dm_t1_new1 on table xu_t11 using 'bloomfilter' 
> dmproperties('index_columns'='id')
> insert into xu_t11 select 1,'m',122.33
> // id.bloomindex is generated
> drop table if exists xu_t11
> // recreate table with same datamap but with different index column
> create table xu_t11(id int,name string,sal float) stored by 'carbondata' 
> tblproperties('sort_columns'='id')
> create datamap xu_dm_t1_new1 on table xu_t11 using 'bloomfilter' 
> dmproperties('index_columns'='sal')
> insert into xu_t11 select 1,'m',122.33
> // Expecte sal.bloomindex to be generated, but actually id.bloomindex is 
> generated.
> ```
> This will not be reproduced in testcase, can only be generated in cluster 
> (with multiple executors).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2476: [CARBONDATA-2534][MV] Fix substring expression not w...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2476
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7092/



---


[GitHub] carbondata pull request #2486: [CARBONDATA-2723][DataMap] Fix bugs in recrea...

2018-07-13 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2486


---


[GitHub] carbondata issue #2486: [CARBONDATA-2723][DataMap] Fix bugs in recreate data...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2486
  
LGTM


---


[GitHub] carbondata issue #2486: [CARBONDATA-2723][DataMap] Fix bugs in recreate data...

2018-07-13 Thread xuchuanyin
Github user xuchuanyin commented on the issue:

https://github.com/apache/carbondata/pull/2486
  
The error in SDV is not related to this PR:
```

org.apache.carbondata.cluster.sdv.generated.QueriesIncludeDictionaryTestCase.TABLE_DICTIONARY_INCLUDE_PushUP_044
Results do not match for query:
== Results ==
!== Correct Answer - 1 ==   == Spark Answer - 1 ==
![7.452597748961225E39] [7.452597748961223E39]
  
  at 
org.scalatest.Assertions$class.newAssertionFailedException(Assertions.scala:495)
  at 
org.scalatest.FunSuite.newAssertionFailedException(FunSuite.scala:1555)
  at org.scalatest.Assertions$class.fail(Assertions.scala:1328)
  at org.scalatest.FunSuite.fail(FunSuite.scala:1555)
  at 
org.apache.spark.sql.common.util.QueryTest.checkAnswer(QueryTest.scala:75)
  at 
org.apache.spark.sql.common.util.QueryTest.writeAndCheckAnswer(QueryTest.scala:113)
```


---


[GitHub] carbondata pull request #2482: [CARBONDATA-2714] Support merge index files f...

2018-07-13 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2482#discussion_r202301535
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/metadata/SegmentFileStore.java ---
@@ -488,6 +485,49 @@ private void readIndexFiles(SegmentStatus status, 
boolean ignoreStatus) throws I
 }
   }
 
+  /**
+   * Reads all merge index / index files as per the status of the file.
+   * In case of @ignoreStatus is true it just reads all merge index / 
index files
+   *
+   * @param status
+   * @param ignoreStatus
+   * @return
+   * @throws IOException
+   */
+  private List readIndexOrMergeFiles(SegmentStatus status, boolean 
ignoreStatus)
--- End diff --

why do you need this method, Already merge files are available in 
`SegmentIndexFileStore.getCarbonMergeFileToIndexFilesMap`


---


[GitHub] carbondata issue #2479: [CARBONDATA-2542][MV] Fix the mv query from table wi...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2479
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7091/



---


[GitHub] carbondata issue #2490: [CARBONDATA-2727][BloomDataMap] Support create bloom...

2018-07-13 Thread kevinjmh
Github user kevinjmh commented on the issue:

https://github.com/apache/carbondata/pull/2490
  
retest this please


---


[GitHub] carbondata issue #2486: [CARBONDATA-2723][DataMap] Fix bugs in recreate data...

2018-07-13 Thread xuchuanyin
Github user xuchuanyin commented on the issue:

https://github.com/apache/carbondata/pull/2486
  
retest sdv please


---


[GitHub] carbondata issue #2486: [CARBONDATA-2723][DataMap] Fix bugs in recreate data...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2486
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/5819/



---


[GitHub] carbondata issue #2307: [CARBONDATA-2482] Pass uuid while writing segment fi...

2018-07-13 Thread manishgupta88
Github user manishgupta88 commented on the issue:

https://github.com/apache/carbondata/pull/2307
  
LGTM...can be merged once build passes


---


[GitHub] carbondata pull request #2482: [CARBONDATA-2714] Support merge index files f...

2018-07-13 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2482#discussion_r202298843
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/events/AlterTableMergeIndexEventListener.scala
 ---
@@ -0,0 +1,95 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.spark.sql.events
+
+import java.util
+
+import scala.collection.JavaConverters._
+import scala.collection.mutable
+
+import org.apache.spark.internal.Logging
+import org.apache.spark.sql.util.CarbonException
+
+import org.apache.carbondata.common.logging.{LogService, LogServiceFactory}
+import org.apache.carbondata.core.datamap.Segment
+import org.apache.carbondata.core.locks.{CarbonLockFactory, LockUsage}
+import org.apache.carbondata.core.statusmanager.SegmentStatusManager
+import org.apache.carbondata.events.{AlterTableMergeIndexEvent, Event, 
OperationContext, OperationEventListener}
+import org.apache.carbondata.processing.merger.CarbonDataMergerUtil
+import org.apache.carbondata.spark.util.CommonUtil
+
+class AlterTableMergeIndexEventListener extends OperationEventListener 
with Logging {
--- End diff --

This listener also merge with `MergeIndexEventListener` as another `case `


---


[GitHub] carbondata issue #2505: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA-2732][...

2018-07-13 Thread Sssan520
Github user Sssan520 commented on the issue:

https://github.com/apache/carbondata/pull/2505
  
retest this please.


---


[GitHub] carbondata issue #2506: [CARBONDATA-2682][32K] fix create table with long_st...

2018-07-13 Thread Sssan520
Github user Sssan520 commented on the issue:

https://github.com/apache/carbondata/pull/2506
  
retest this please.


---


[GitHub] carbondata issue #2506: [CARBONDATA-2682][32K] fix create table with long_st...

2018-07-13 Thread Sssan520
Github user Sssan520 commented on the issue:

https://github.com/apache/carbondata/pull/2506
  
There are some problems in the CI env, so I closed #2436 and opened this PR.


---


[GitHub] carbondata issue #2505: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA-2732][...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2505
  
Can one of the admins verify this patch?


---


[GitHub] carbondata issue #2506: [CARBONDATA-2682][32K] fix create table with long_st...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2506
  
Can one of the admins verify this patch?


---


[GitHub] carbondata pull request #2506: [CARBONDATA-2682][32K] fix create table with ...

2018-07-13 Thread Sssan520
GitHub user Sssan520 opened a pull request:

https://github.com/apache/carbondata/pull/2506

[CARBONDATA-2682][32K] fix create table with long_string_columns properties 
bugs

Changes
This PR fixes create table with long_string_columns bugs which are:
1.create table with columns both in long_string_columns and partition or 
no_inverted_index property should be blocked.
2.create table with duplicate columns in long_string_column property should 
be blocked.
And already add new test cases to take care of testing this.

Be sure to do all of the following checklist to help us incorporate 
your contribution quickly and easily:

 - [ ] Any interfaces changed?
 
 - [ ] Any backward compatibility impacted?
 
 - [ ] Document update required?

 - [ ] Testing done
Please provide details on 
- Whether new unit test cases have been added or why no new tests 
are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance 
test report.
- Any additional information to help reviewers in testing this 
change.
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/Sssan520/carbondata dts_longstringcols

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2506.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2506


commit 8f5d6a0ded3080f6c6d7021bf6618e4f823d34cd
Author: Sssan520 
Date:   2018-07-12T11:58:23Z

fix create table with long_string_columns properties bugs




---


[GitHub] carbondata issue #2505: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA-2732][...

2018-07-13 Thread Sssan520
Github user Sssan520 commented on the issue:

https://github.com/apache/carbondata/pull/2505
  
There are some problems in the CI env, so I closed #2491 and opened this PR


---


[GitHub] carbondata pull request #2505: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA...

2018-07-13 Thread Sssan520
GitHub user Sssan520 opened a pull request:

https://github.com/apache/carbondata/pull/2505

[CARBONDATA-2698][CARBONDATA-2700][CARBONDATA-2732][BloomDataMap] block 
some oprerations of bloomfilter datamap

1.Block create bloomfilter datamap index on column which its datatype is 
complex type;
2.Block change datatype for bloomfilter index datamap;
3.Block dropping index columns for bloomfilter index datamap

Be sure to do all of the following checklist to help us incorporate 
your contribution quickly and easily:

 - [ ] Any interfaces changed? 
 
 - [ ] Any backward compatibility impacted?
 
 - [ ] Document update required?

 - [ ] Testing done
Please provide details on 
- Whether new unit test cases have been added or why no new tests 
are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance 
test report.
- Any additional information to help reviewers in testing this 
change.
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/Sssan520/carbondata bloomfilterdatamap

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2505.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2505


commit 8e40f663070022b4b5d919fcc8dc5503fd37fa21
Author: Sssan520 
Date:   2018-07-11T12:17:09Z

Block some operations of bloomfilter datamap




---


[GitHub] carbondata pull request #2436: [CARBONDATA-2682][32K] fix create table with ...

2018-07-13 Thread Sssan520
Github user Sssan520 closed the pull request at:

https://github.com/apache/carbondata/pull/2436


---


[GitHub] carbondata pull request #2504: [CARBONDATA-2740]flat folder handling for imp...

2018-07-13 Thread akashrn5
GitHub user akashrn5 opened a pull request:

https://github.com/apache/carbondata/pull/2504

[CARBONDATA-2740]flat folder handling for implicit column and other bug fix

### Problem
1) When flat folder is enabled for table, pruning is not happening for 
implicit column.
2) error message is wrong when complex column which does not have any 
string datatype column is given in local dictionary and exclude column.
3) when data load is failed, corresponding segment file is not getting 
deleted.

### Solution:
1) when flat folder is enabled, based on the segment number present in the 
carbondata file name,  pruning is taken care
2) error message is corrected
3) when data load is failed, segment file will be deleted




You can merge this pull request into a Git repository by running:

$ git pull https://github.com/akashrn5/incubator-carbondata implicit

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2504.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2504


commit 037769a4ccedd9a5a76cf1c9992a36b99cb4926d
Author: akashrn5 
Date:   2018-07-13T09:04:42Z

flat folder handling for implicit column and other bug fix




---


[GitHub] carbondata issue #2480: [CARBONDATA-2550][CARBONDATA-2576][MV] Fix limit and...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2480
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7090/



---


[GitHub] carbondata issue #2453: [CARBONDATA-2528][MV] Fixed order by in mv and aggre...

2018-07-13 Thread brijoobopanna
Github user brijoobopanna commented on the issue:

https://github.com/apache/carbondata/pull/2453
  
retest this please


---


[GitHub] carbondata issue #2485: [CARBONDATA-2722] [CARBONDATA-2721] JsonWriter issue...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2485
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5864/



---


[jira] [Created] (CARBONDATA-2740) flat folder structure is not handled for implicit column and segment file is not getting deleted after load is failed

2018-07-13 Thread Akash R Nilugal (JIRA)
Akash R Nilugal created CARBONDATA-2740:
---

 Summary: flat folder structure is not handled for implicit column 
and segment file is not getting deleted after load is failed
 Key: CARBONDATA-2740
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2740
 Project: CarbonData
  Issue Type: Bug
Reporter: Akash R Nilugal
Assignee: Akash R Nilugal


flat folder structure is not handled for implicit column and segment file is 
not getting deleted after load is failed



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2491: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA...

2018-07-13 Thread Sssan520
Github user Sssan520 closed the pull request at:

https://github.com/apache/carbondata/pull/2491


---


[GitHub] carbondata pull request #2501: [CARBONDATA-2738]Block Preaggregate, Dictiona...

2018-07-13 Thread akashrn5
Github user akashrn5 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2501#discussion_r202285088
  
--- Diff: 
integration/spark-common/src/main/scala/org/apache/spark/sql/catalyst/CarbonDDLSqlParser.scala
 ---
@@ -720,18 +720,18 @@ abstract class CarbonDDLSqlParser extends 
AbstractCarbonSparkSQLParser {
 
tableProperties.get(CarbonCommonConstants.DICTIONARY_EXCLUDE).get.split(',').map(_.trim)
   dictExcludeCols
 .foreach { dictExcludeCol =>
+  if (dictExcludeCol.contains(".") || 
dictExcludeCol.contains("[")) {
--- End diff --

do not check, for contains ".", get from field whether it is child column, 
then you can check whether that is given in include or exclude and throw 
exception


---


[GitHub] carbondata pull request #2501: [CARBONDATA-2738]Block Preaggregate, Dictiona...

2018-07-13 Thread akashrn5
Github user akashrn5 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2501#discussion_r202285188
  
--- Diff: 
integration/spark-common/src/main/scala/org/apache/spark/sql/catalyst/CarbonDDLSqlParser.scala
 ---
@@ -748,6 +748,10 @@ abstract class CarbonDDLSqlParser extends 
AbstractCarbonSparkSQLParser {
   dictIncludeCols =
 
tableProperties(CarbonCommonConstants.DICTIONARY_INCLUDE).split(",").map(_.trim)
   dictIncludeCols.foreach { distIncludeCol =>
+if (distIncludeCol.contains(".") || distIncludeCol.contains("[")) {
--- End diff --

same as above


---


[GitHub] carbondata pull request #2501: [CARBONDATA-2738]Block Preaggregate, Dictiona...

2018-07-13 Thread akashrn5
Github user akashrn5 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2501#discussion_r202286122
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/mutation/CarbonProjectForUpdateCommand.scala
 ---
@@ -59,6 +60,13 @@ private[sql] case class CarbonProjectForUpdateCommand(
   return Seq.empty
 }
 val carbonTable = CarbonEnv.getCarbonTable(databaseNameOp, 
tableName)(sparkSession)
+columns.foreach { col =>
+  var dataType = carbonTable.getColumnByName(tableName, 
col).getColumnSchema.getDataType
+  if(dataType.getName.equals("STRUCT") || 
dataType.getName.equals("ARRAY")) {
--- End diff --

do not check for hard coded value, get the datatype and dicrectly check 
`isComplexType` and throw error


---


[GitHub] carbondata issue #2486: [CARBONDATA-2723][DataMap] Fix bugs in recreate data...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2486
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5863/



---


[GitHub] carbondata issue #2489: [CARBONDATA-2606][Complex DataType Enhancements]Fix ...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2489
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/5818/



---


[GitHub] carbondata issue #2482: [CARBONDATA-2714] Support merge index files for the ...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2482
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7089/



---


[GitHub] carbondata pull request #2491: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA...

2018-07-13 Thread Sssan520
Github user Sssan520 closed the pull request at:

https://github.com/apache/carbondata/pull/2491


---


[GitHub] carbondata pull request #2491: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA...

2018-07-13 Thread Sssan520
GitHub user Sssan520 reopened a pull request:

https://github.com/apache/carbondata/pull/2491

[CARBONDATA-2698][CARBONDATA-2700][CARBONDATA-2732][BloomDataMap] block 
some oprerations of bloomfilter datamap

1.Block create bloomfilter datamap index on column which its datatype is 
complex type;
2.Block change datatype for bloomfilter index datamap;
3.Block dropping index columns for bloomfilter index datamap

Be sure to do all of the following checklist to help us incorporate 
your contribution quickly and easily:

 - [ ] Any interfaces changed? 
 add a parameter "targets" for method "canAllow" of "CarbonTable" class

 - [ ] Any backward compatibility impacted?  No
 
 - [ ] Document update required? No

 - [ ] Testing done NA
Please provide details on 
- Whether new unit test cases have been added or why no new tests 
are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance 
test report.
- Any additional information to help reviewers in testing this 
change.
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/Sssan520/carbondata bloomfilter

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2491.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2491


commit 815c497380fa52b98aa270d35ce67f2cb9c67191
Author: lianganping 00251374 
Date:   2018-07-11T12:17:09Z

1.Block create bloomfilter datamap index on column which its datatype is 
complex;2.Block create bloomfilter datamap index on local_dictionary 
column;3.Block change datatype for bloomfilter index datamap;4.Block dropping 
index columns for bloomfilter index datamap




---


[GitHub] carbondata pull request #2491: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA...

2018-07-13 Thread Sssan520
Github user Sssan520 closed the pull request at:

https://github.com/apache/carbondata/pull/2491


---


[GitHub] carbondata pull request #2491: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA...

2018-07-13 Thread Sssan520
GitHub user Sssan520 reopened a pull request:

https://github.com/apache/carbondata/pull/2491


[CARBONDATA-2698][CARBONDATA-2700][CARBONDATA-2730][CARBONDATA-2732][BloomDataMap]
 block some oprerations of bloomfilter datamap

1.Block create bloomfilter datamap index on column which its datatype is 
complex type;
2.Block change datatype for bloomfilter index datamap;
3.Block dropping index columns for bloomfilter index datamap

Be sure to do all of the following checklist to help us incorporate 
your contribution quickly and easily:

 - [ ] Any interfaces changed? 
 add a parameter "targets" for method "canAllow" of "CarbonTable" class

 - [ ] Any backward compatibility impacted?  No
 
 - [ ] Document update required? No

 - [ ] Testing done NA
Please provide details on 
- Whether new unit test cases have been added or why no new tests 
are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance 
test report.
- Any additional information to help reviewers in testing this 
change.
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/Sssan520/carbondata bloomfilter

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2491.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2491


commit 815c497380fa52b98aa270d35ce67f2cb9c67191
Author: lianganping 00251374 
Date:   2018-07-11T12:17:09Z

1.Block create bloomfilter datamap index on column which its datatype is 
complex;2.Block create bloomfilter datamap index on local_dictionary 
column;3.Block change datatype for bloomfilter index datamap;4.Block dropping 
index columns for bloomfilter index datamap




---


[jira] [Closed] (CARBONDATA-2730) Block create bloomfilter datamap index on local_dictionary column

2018-07-13 Thread lianganping (JIRA)


 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2730?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lianganping closed CARBONDATA-2730.
---
Resolution: Not A Problem

> Block create bloomfilter datamap index on local_dictionary column
> -
>
> Key: CARBONDATA-2730
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2730
> Project: CarbonData
>  Issue Type: Sub-task
>Reporter: lianganping
>Assignee: lianganping
>Priority: Major
>
> Block create bloomfilter datamap index on local_dictionary column;



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2491: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA-2730][...

2018-07-13 Thread Sssan520
Github user Sssan520 commented on the issue:

https://github.com/apache/carbondata/pull/2491
  
retest this please


---


[GitHub] carbondata issue #2491: [CARBONDATA-2698][CARBONDATA-2700][CARBONDATA-2730][...

2018-07-13 Thread Sssan520
Github user Sssan520 commented on the issue:

https://github.com/apache/carbondata/pull/2491
  
all code review comments has been handled.


---


[GitHub] carbondata issue #2436: [CARBONDATA-2682][32K] fix create table with long_st...

2018-07-13 Thread Sssan520
Github user Sssan520 commented on the issue:

https://github.com/apache/carbondata/pull/2436
  
retest this please


---


[GitHub] carbondata pull request #2499: [CARBONDATA-2648] Fixed NPE issue with legacy...

2018-07-13 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2499


---


[GitHub] carbondata issue #2499: [CARBONDATA-2648] Fixed NPE issue with legacy store ...

2018-07-13 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2499
  
LGTM


---


[GitHub] carbondata pull request #2499: [CARBONDATA-2648] Fixed NPE issue with legacy...

2018-07-13 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2499#discussion_r202278145
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/indexstore/TableBlockIndexUniqueIdentifierWrapper.java
 ---
@@ -35,18 +35,34 @@
 
   // holds the reference to CarbonTable
   private CarbonTable carbonTable;
+  /**
+   * flag to specify whether to load table block metadata in unsafe or 
safe. Default value is true
+   */
+  private boolean addTableBlockToUnsafe = true;
 
   public TableBlockIndexUniqueIdentifierWrapper(
   TableBlockIndexUniqueIdentifier tableBlockIndexUniqueIdentifier, 
CarbonTable carbonTable) {
 this.tableBlockIndexUniqueIdentifier = tableBlockIndexUniqueIdentifier;
 this.carbonTable = carbonTable;
   }
 
+  public TableBlockIndexUniqueIdentifierWrapper(
--- End diff --

As discussed I have added a Note


---


[GitHub] carbondata pull request #2485: [CARBONDATA-2722] [CARBONDATA-2721] JsonWrite...

2018-07-13 Thread ajantha-bhat
Github user ajantha-bhat commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2485#discussion_r202277326
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/loading/parser/impl/JsonRowParser.java
 ---
@@ -56,6 +56,9 @@ public JsonRowParser(DataField[] dataFields) {
   Map jsonNodeMap =
   objectMapper.readValue(jsonString, new TypeReference>() {
   });
+  if (jsonNodeMap == null) {
--- End diff --

yeah, not required. removed it.


---


[GitHub] carbondata issue #2485: [CARBONDATA-2722] [CARBONDATA-2721] JsonWriter issue...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2485
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7087/



---


[GitHub] carbondata issue #2487: [CARBONDATA-2735]Fixed Performance issue for complex...

2018-07-13 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2487
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/5862/



---


[GitHub] carbondata pull request #2503: [CARBONDATA-2734] Update is not working on th...

2018-07-13 Thread ravipesala
GitHub user ravipesala opened a pull request:

https://github.com/apache/carbondata/pull/2503

[CARBONDATA-2734] Update is not working on the table which has segmentfile 
present

It reverts to the PR https://github.com/apache/carbondata/pull/2385 
It fixes the IUD on the flat folder.

Be sure to do all of the following checklist to help us incorporate 
your contribution quickly and easily:

 - [ ] Any interfaces changed?
 
 - [ ] Any backward compatibility impacted?
 
 - [ ] Document update required?

 - [ ] Testing done
Please provide details on 
- Whether new unit test cases have been added or why no new tests 
are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance 
test report.
- Any additional information to help reviewers in testing this 
change.
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/ravipesala/incubator-carbondata 
flat-folder-update-issue

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2503.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2503


commit caf2304736741748e6ba896d5ba29333ee95defc
Author: ravipesala 
Date:   2018-07-13T07:45:15Z

Update is not working on the table which has segmentfile present




---


<    1   2   3   >