[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184919901
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/store/TablePage.java 
---
@@ -222,8 +222,9 @@ private void addComplexColumn(int index, int rowId, 
byte[] complexColumns) {
   ByteBuffer byteArrayInput = ByteBuffer.wrap(complexColumns);
   ByteArrayOutputStream byteArrayOutput = new ByteArrayOutputStream();
   DataOutputStream dataOutputStream = new 
DataOutputStream(byteArrayOutput);
-  complexDataType.parseAndBitPack(byteArrayInput, dataOutputStream,
-  model.getComplexDimensionKeyGenerator());
+  int startOffset = 0;
--- End diff --

What is the use of `startOffset` when not used?


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184919620
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/loading/CarbonDataLoadConfiguration.java
 ---
@@ -348,6 +356,33 @@ public void 
setCardinalityFinder(DictionaryCardinalityFinder cardinalityFinder)
 return complexKeyGenerators;
   }
 
+  public Boolean[][] createComplexDictionaryFieldIdentification() {
--- End diff --

If not required, please remove this method


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184919569
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java
 ---
@@ -245,6 +312,28 @@ public void parseAndBitPack(ByteBuffer byteArrayInput, 
DataOutputStream dataOutp
 dataOutputStream.write(v);
   }
 
+  @Override
+  public int parseComplexValue(ByteBuffer byteArrayInput, DataOutputStream 
dataOutputStream,
+  KeyGenerator[] generator, Boolean[][] 
complexDictionaryIndentification, int startOffset)
--- End diff --

What is the use of `complexDictionaryIndentification` if not used


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread kumarvishal09
Github user kumarvishal09 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184919435
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/util/CarbonDataProcessorUtil.java
 ---
@@ -267,14 +267,24 @@ private static String 
getComplexTypeString(DataField[] dataFields) {
 return dimString.toString();
   }
 
+  private static String isDictionaryType(CarbonDimension dimension) {
--- End diff --

why this method return type is string??? Change this to string


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184919308
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java
 ---
@@ -245,6 +312,28 @@ public void parseAndBitPack(ByteBuffer byteArrayInput, 
DataOutputStream dataOutp
 dataOutputStream.write(v);
   }
 
+  @Override
+  public int parseComplexValue(ByteBuffer byteArrayInput, DataOutputStream 
dataOutputStream,
+  KeyGenerator[] generator, Boolean[][] 
complexDictionaryIndentification, int startOffset)
+  throws IOException, KeyGenException {
+if (!this.isDictionary) {
+  int sizeOfData = byteArrayInput.getInt();
+  startOffset += Integer.SIZE / Byte.SIZE;
+  dataOutputStream.writeInt(sizeOfData);
+  byte[] bb = new byte[sizeOfData];
+  byteArrayInput.get(bb, 0, sizeOfData);
+  dataOutputStream.write(bb);
+  startOffset += sizeOfData;
+} else {
+  int data = byteArrayInput.getInt();
+  startOffset += Integer.SIZE / Byte.SIZE;
--- End diff --

I think this is not required, if it is only writeoffset


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread kumarvishal09
Github user kumarvishal09 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184919245
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/store/CarbonFactDataHandlerModel.java
 ---
@@ -609,5 +616,13 @@ public short getWritingCoresCount() {
   public DataMapWriterListener getDataMapWriterlistener() {
 return dataMapWriterlistener;
   }
+
+  public Boolean[][] getComplexDictionaryFields() {
--- End diff --

change this to boolean[][] getComplexDictionaryFields()


---


[GitHub] carbondata issue #2224: [CARBONDATA-2393]TaskNo is not working for SDK

2018-04-29 Thread ajantha-bhat
Github user ajantha-bhat commented on the issue:

https://github.com/apache/carbondata/pull/2224
  
retest this please


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184919017
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java
 ---
@@ -245,6 +312,28 @@ public void parseAndBitPack(ByteBuffer byteArrayInput, 
DataOutputStream dataOutp
 dataOutputStream.write(v);
   }
 
+  @Override
+  public int parseComplexValue(ByteBuffer byteArrayInput, DataOutputStream 
dataOutputStream,
+  KeyGenerator[] generator, Boolean[][] 
complexDictionaryIndentification, int startOffset)
+  throws IOException, KeyGenException {
+if (!this.isDictionary) {
+  int sizeOfData = byteArrayInput.getInt();
+  startOffset += Integer.SIZE / Byte.SIZE;
--- End diff --

No need to calculate every time , just use 4 and add comment , Do it in all 
places


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread kumarvishal09
Github user kumarvishal09 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184918803
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java
 ---
@@ -245,6 +312,28 @@ public void parseAndBitPack(ByteBuffer byteArrayInput, 
DataOutputStream dataOutp
 dataOutputStream.write(v);
   }
 
+  @Override
+  public int parseComplexValue(ByteBuffer byteArrayInput, DataOutputStream 
dataOutputStream,
+  KeyGenerator[] generator, Boolean[][] 
complexDictionaryIndentification, int startOffset)
--- End diff --

Boolean[][] complexDictionaryIndentification is not required as each 
Primitive type field knows whether it is of dictionary/ no dictionary type.


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184918720
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/ArrayDataType.java
 ---
@@ -183,7 +187,27 @@ public void parseAndBitPack(ByteBuffer byteArrayInput, 
DataOutputStream dataOutp
 for (int i = 0; i < dataLength; i++) {
   children.parseAndBitPack(byteArrayInput, dataOutputStream, 
generator);
 }
+  }
+
+  @Override
+  public int parseComplexValue(ByteBuffer byteArrayInput, DataOutputStream 
dataOutputStream,
+  KeyGenerator[] generator, Boolean[][] 
complexDictionaryIndentification, int startOffset)
+  throws IOException, KeyGenException {
+int dataLength = byteArrayInput.getInt();
+startOffset += Integer.SIZE / Byte.SIZE;
--- End diff --

No need to calculate every time , just use `4` and add comment


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184918753
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/ArrayDataType.java
 ---
@@ -183,7 +187,27 @@ public void parseAndBitPack(ByteBuffer byteArrayInput, 
DataOutputStream dataOutp
 for (int i = 0; i < dataLength; i++) {
   children.parseAndBitPack(byteArrayInput, dataOutputStream, 
generator);
 }
+  }
+
+  @Override
+  public int parseComplexValue(ByteBuffer byteArrayInput, DataOutputStream 
dataOutputStream,
+  KeyGenerator[] generator, Boolean[][] 
complexDictionaryIndentification, int startOffset)
+  throws IOException, KeyGenException {
+int dataLength = byteArrayInput.getInt();
+startOffset += Integer.SIZE / Byte.SIZE;
 
+dataOutputStream.writeInt(dataLength);
+if (children instanceof PrimitiveDataType) {
+  if (children.getIsColumnDictionary()) {
+
dataOutputStream.writeInt(generator[children.getSurrogateIndex()].getKeySizeInBytes());
+startOffset += Integer.SIZE / Byte.SIZE;
--- End diff --

No need to calculate every time , just use 4 and add comment


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread kumarvishal09
Github user kumarvishal09 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184918635
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java
 ---
@@ -211,29 +241,66 @@ public int getSurrogateIndex() {
   /*
* set surrogate index
*/
-  @Override
-  public void setSurrogateIndex(int surrIndex) {
-index = surrIndex;
+  @Override public void setSurrogateIndex(int surrIndex) {
+if (this.carbonDimension != null && 
!this.carbonDimension.hasEncoding(Encoding.DICTIONARY)) {
+  index = 0;
+} else if (this.carbonDimension == null && isDictionary == false) {
+  index = 0;
+} else {
+  index = surrIndex;
+}
+  }
+
+  @Override public Boolean getIsColumnDictionary() {
+return isDictionary;
   }
 
   @Override public void writeByteArray(Object input, DataOutputStream 
dataOutputStream)
   throws IOException, DictionaryGenerationException {
 String parsedValue =
 input == null ? null : DataTypeUtil.parseValue(input.toString(), 
carbonDimension);
-Integer surrogateKey;
-if (null == parsedValue) {
-  surrogateKey = 
CarbonCommonConstants.MEMBER_DEFAULT_VAL_SURROGATE_KEY;
-} else {
-  surrogateKey = dictionaryGenerator.getOrGenerateKey(parsedValue);
-  if (surrogateKey == CarbonCommonConstants.INVALID_SURROGATE_KEY) {
+if (this.isDictionary) {
+  Integer surrogateKey;
+  if (null == parsedValue) {
 surrogateKey = 
CarbonCommonConstants.MEMBER_DEFAULT_VAL_SURROGATE_KEY;
+  } else {
+surrogateKey = dictionaryGenerator.getOrGenerateKey(parsedValue);
+if (surrogateKey == CarbonCommonConstants.INVALID_SURROGATE_KEY) {
+  surrogateKey = 
CarbonCommonConstants.MEMBER_DEFAULT_VAL_SURROGATE_KEY;
+}
+  }
+  dataOutputStream.writeInt(surrogateKey);
+} else {
+  // Transform into ByteArray for No Dictionary.
+  // TODO have to refactor and place all the cases present in 
NonDictionaryFieldConverterImpl
+  if (null == parsedValue && this.carbonDimension.getDataType() != 
DataTypes.STRING) {
+updateNullValue(dataOutputStream);
+  } else {
+byte[] value = 
DataTypeUtil.getBytesBasedOnDataTypeForNoDictionaryColumn(parsedValue,
--- End diff --

In case of no dictionary primitive type column NoDictionary field converter 
will convert and give the parsed value so no need to call  
DataTypeUtil.getBytesBasedOnDataTypeForNoDictionaryColumn again, dictionary 
write this value in LV format to outputstream


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184918548
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/ArrayDataType.java
 ---
@@ -171,9 +175,9 @@ public void fillCardinality(List 
dimCardWithComplex) {
   /**
* parse byte array and bit pack
*/
-  @Override
-  public void parseAndBitPack(ByteBuffer byteArrayInput, DataOutputStream 
dataOutputStream,
-  KeyGenerator[] generator) throws IOException, KeyGenException {
+  @Override public void parseAndBitPack(ByteBuffer byteArrayInput,
--- End diff --

Now this method is not used? if not delete it


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184918197
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/util/SparkTypeConverter.scala
 ---
@@ -123,13 +137,29 @@ private[spark] object SparkTypeConverter {
   private def recursiveMethod(
   table: CarbonTable, dimName: String, childDim: CarbonDimension) = {
 childDim.getDataType.getName.toLowerCase match {
-  case "array" => s"${
-childDim.getColName.substring(dimName.length + 1)
-  }:array<${ getArrayChildren(table, childDim.getColName) }>"
-  case "struct" => s"${
-childDim.getColName.substring(dimName.length + 1)
-  }:struct<${ getStructChildren(table, childDim.getColName) }>"
-  case dType => s"${ childDim.getColName.substring(dimName.length + 1) 
}:${ dType }"
+  case "array" => if (table.isTransactionalTable) {
--- End diff --

Better write another method rather than keeping so many if checks


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184918157
  
--- Diff: 
integration/spark2/src/main/scala/org/apache/spark/util/SparkTypeConverter.scala
 ---
@@ -97,16 +97,30 @@ private[spark] object SparkTypeConverter {
   def getStructChildren(table: CarbonTable, dimName: String): String = {
 table.getChildren(dimName).asScala.map(childDim => {
   childDim.getDataType.getName.toLowerCase match {
-case "array" => s"${
+case "array" => if (table.isTransactionalTable) {s"${
--- End diff --

Could explain in comments why this many checks are required for non 
transactional table? 
Better write another method rather than keeping so many if checks


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread kumarvishal09
Github user kumarvishal09 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184918097
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java
 ---
@@ -211,29 +241,66 @@ public int getSurrogateIndex() {
   /*
* set surrogate index
*/
-  @Override
-  public void setSurrogateIndex(int surrIndex) {
-index = surrIndex;
+  @Override public void setSurrogateIndex(int surrIndex) {
+if (this.carbonDimension != null && 
!this.carbonDimension.hasEncoding(Encoding.DICTIONARY)) {
+  index = 0;
+} else if (this.carbonDimension == null && isDictionary == false) {
+  index = 0;
+} else {
+  index = surrIndex;
+}
+  }
+
+  @Override public Boolean getIsColumnDictionary() {
+return isDictionary;
   }
 
   @Override public void writeByteArray(Object input, DataOutputStream 
dataOutputStream)
   throws IOException, DictionaryGenerationException {
 String parsedValue =
 input == null ? null : DataTypeUtil.parseValue(input.toString(), 
carbonDimension);
-Integer surrogateKey;
-if (null == parsedValue) {
-  surrogateKey = 
CarbonCommonConstants.MEMBER_DEFAULT_VAL_SURROGATE_KEY;
-} else {
-  surrogateKey = dictionaryGenerator.getOrGenerateKey(parsedValue);
-  if (surrogateKey == CarbonCommonConstants.INVALID_SURROGATE_KEY) {
+if (this.isDictionary) {
+  Integer surrogateKey;
+  if (null == parsedValue) {
 surrogateKey = 
CarbonCommonConstants.MEMBER_DEFAULT_VAL_SURROGATE_KEY;
+  } else {
+surrogateKey = dictionaryGenerator.getOrGenerateKey(parsedValue);
+if (surrogateKey == CarbonCommonConstants.INVALID_SURROGATE_KEY) {
+  surrogateKey = 
CarbonCommonConstants.MEMBER_DEFAULT_VAL_SURROGATE_KEY;
+}
+  }
+  dataOutputStream.writeInt(surrogateKey);
+} else {
+  // Transform into ByteArray for No Dictionary.
+  // TODO have to refactor and place all the cases present in 
NonDictionaryFieldConverterImpl
+  if (null == parsedValue && this.carbonDimension.getDataType() != 
DataTypes.STRING) {
--- End diff --

null  value updation is also required for String data type second condition 
in If check is not correct


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184917833
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/scan/complextypes/PrimitiveQueryType.java
 ---
@@ -108,4 +128,14 @@ public PrimitiveQueryType(String name, String 
parentname, int blockIndex,
 }
 return actualData;
   }
+
+  @Override public Object 
getDataBasedOnDataTypeFromNoDictionary(ByteBuffer data) {
--- End diff --

I don't think it is required to have one more method, just rename and 
handle in same method `getDataBasedOnDataTypeFromSurrogates` , it can avoid 
duplicate code and extra handling


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread kumarvishal09
Github user kumarvishal09 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184917713
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java
 ---
@@ -94,6 +99,11 @@
 
   private CarbonDimension carbonDimension;
 
+  private Boolean isDictionary;
+
+  private FieldConverter fieldConverterForNoDictionary;
--- End diff --

FieldConverter is not required in this class, as caller will create for 
primitive type like existing flow..same way we can handle for no dictionary


---


[GitHub] carbondata issue #1812: [CARBONDATA-2033]Support user specified segments in ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/1812
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4626/



---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184917500
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/scan/complextypes/PrimitiveQueryType.java
 ---
@@ -84,6 +101,9 @@ public PrimitiveQueryType(String name, String 
parentname, int blockIndex,
   DimensionRawColumnChunk[] rawColumnChunks, int rowNumber,
   int pageNumber, DataOutputStream dataOutputStream) throws 
IOException {
 byte[] currentVal = copyBlockDataChunk(rawColumnChunks, rowNumber, 
pageNumber);
+if (!this.isDictionary) {
+  dataOutputStream.writeInt(currentVal.length);
--- End diff --

Just writing length? where are you writing data?


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184917127
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/scan/complextypes/PrimitiveQueryType.java
 ---
@@ -53,6 +55,21 @@ public PrimitiveQueryType(String name, String 
parentname, int blockIndex,
 this.name = name;
 this.parentname = parentname;
 this.isDirectDictionary = isDirectDictionary;
+this.isDictionary = true;
+  }
+
+
+  public PrimitiveQueryType(String name, String parentname, int blockIndex,
+  org.apache.carbondata.core.metadata.datatype.DataType dataType, int 
keySize,
+  Dictionary dictionary, boolean isDirectDictionary, boolean 
isDictionary) {
--- End diff --

I don't think it is required to have a separate constructor, if (dictionary 
== null && !isDirectDictionary) then it becomes nodictionary


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread kumarvishal09
Github user kumarvishal09 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184917041
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java
 ---
@@ -94,6 +99,11 @@
 
   private CarbonDimension carbonDimension;
 
+  private Boolean isDictionary;
--- End diff --

change this to  boolean isDictionary


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread kumarvishal09
Github user kumarvishal09 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184916986
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/GenericDataType.java
 ---
@@ -71,6 +71,8 @@ void writeByteArray(T input, DataOutputStream 
dataOutputStream)
*/
   void setSurrogateIndex(int surrIndex);
 
+  Boolean getIsColumnDictionary();
--- End diff --

change this to boolean isDictionaryEncoded()


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread kumarvishal09
Github user kumarvishal09 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184916825
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/ArrayDataType.java
 ---
@@ -183,7 +187,27 @@ public void parseAndBitPack(ByteBuffer byteArrayInput, 
DataOutputStream dataOutp
 for (int i = 0; i < dataLength; i++) {
   children.parseAndBitPack(byteArrayInput, dataOutputStream, 
generator);
 }
+  }
+
+  @Override
+  public int parseComplexValue(ByteBuffer byteArrayInput, DataOutputStream 
dataOutputStream,
+  KeyGenerator[] generator, Boolean[][] 
complexDictionaryIndentification, int startOffset)
--- End diff --

complexDictionaryIndentification is not required as each primitive children 
knows it is not of dictionary type or no dictionary type


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184916731
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/scan/collector/impl/DictionaryBasedResultCollector.java
 ---
@@ -134,7 +134,13 @@ void fillDimensionData(BlockletScannedResult 
scannedResult, int[] surrogateResul
   row[order[i]] =
   
DataTypeUtil.getDataBasedOnDataType(scannedResult.getBlockletId(), 
DataTypes.STRING);
 }
-  } else {
+  } else if (complexDataTypeArray[i]) {
+// Complex Type With No Dictionary Encoding.
+row[order[i]] = 
comlexDimensionInfoMap.get(queryDimensions[i].getDimension().getOrdinal())
+.getDataBasedOnDataTypeFromNoDictionary(
+
ByteBuffer.wrap(complexTypeKeyArray[complexTypeColumnIndex++]));
+  } else
+  {
--- End diff --

Move the braces up


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread kumarvishal09
Github user kumarvishal09 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184916594
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/ArrayDataType.java
 ---
@@ -171,9 +175,9 @@ public void fillCardinality(List 
dimCardWithComplex) {
   /**
* parse byte array and bit pack
*/
-  @Override
-  public void parseAndBitPack(ByteBuffer byteArrayInput, DataOutputStream 
dataOutputStream,
-  KeyGenerator[] generator) throws IOException, KeyGenException {
+  @Override public void parseAndBitPack(ByteBuffer byteArrayInput,
--- End diff --

Please remove this method from interface as below new method will be used


---


[GitHub] carbondata pull request #2244: [CARBONDATA-2310] Refactored code to improve ...

2018-04-29 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2244#discussion_r184915943
  
--- Diff: 
hadoop/src/main/java/org/apache/carbondata/hadoop/CacheClient.java ---
@@ -16,21 +16,40 @@
  */
 package org.apache.carbondata.hadoop;
 
+import java.util.List;
+import java.util.Map;
+import java.util.concurrent.ConcurrentHashMap;
+
+import org.apache.carbondata.common.logging.LogService;
+import org.apache.carbondata.common.logging.LogServiceFactory;
 import org.apache.carbondata.core.cache.Cache;
 import org.apache.carbondata.core.cache.CacheProvider;
 import org.apache.carbondata.core.cache.CacheType;
+import org.apache.carbondata.core.datastore.SegmentTaskIndexStore;
 import org.apache.carbondata.core.datastore.TableSegmentUniqueIdentifier;
+import org.apache.carbondata.core.datastore.block.SegmentProperties;
 import org.apache.carbondata.core.datastore.block.SegmentTaskIndexWrapper;
+import org.apache.carbondata.core.metadata.AbsoluteTableIdentifier;
+import 
org.apache.carbondata.core.metadata.schema.table.column.ColumnSchema;
 
 /**
  * CacheClient : Holds all the Cache access clients for Btree, Dictionary
  */
 public class CacheClient {
 
+  private static final LogService LOGGER =
+  LogServiceFactory.getLogService(CacheClient.class.getName());
+
+  private final Object lock = new Object();
+
   // segment access client for driver LRU cache
   private CacheAccessClient
   segmentAccessClient;
 
+  private static Map
+  segmentProperties =
+  new ConcurrentHashMap<>();
--- End diff --

Move this to above line


---


[GitHub] carbondata pull request #2244: [CARBONDATA-2310] Refactored code to improve ...

2018-04-29 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2244#discussion_r184911430
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/datamap/dev/expr/DataMapExprWrapperImpl.java
 ---
@@ -86,4 +87,8 @@ public DataMapExprWrapperImpl(TableDataMap dataMap, 
FilterResolverIntf expressio
   @Override public DataMapLevel getDataMapType() {
 return dataMap.getDataMapFactory().getDataMapType();
   }
+
+  public DataMapFactory getDataMapFactory() {
--- End diff --

this is not an interface method...can we try to get the dataMapFactory in 
some other way and remove this method from here?


---


[GitHub] carbondata pull request #2244: [CARBONDATA-2310] Refactored code to improve ...

2018-04-29 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2244#discussion_r184915870
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/util/BlockletDataMapUtil.java ---
@@ -0,0 +1,198 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.carbondata.core.util;
+
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.HashMap;
+import java.util.HashSet;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+import java.util.TreeMap;
+
+import org.apache.carbondata.common.logging.LogService;
+import org.apache.carbondata.common.logging.LogServiceFactory;
+import org.apache.carbondata.core.constants.CarbonCommonConstants;
+import org.apache.carbondata.core.datamap.Segment;
+import 
org.apache.carbondata.core.datastore.filesystem.AbstractDFSCarbonFile;
+import org.apache.carbondata.core.datastore.filesystem.CarbonFile;
+import org.apache.carbondata.core.datastore.impl.FileFactory;
+import org.apache.carbondata.core.indexstore.BlockMetaInfo;
+import 
org.apache.carbondata.core.indexstore.TableBlockIndexUniqueIdentifier;
+import 
org.apache.carbondata.core.indexstore.blockletindex.BlockletDataMapDistributable;
+import 
org.apache.carbondata.core.indexstore.blockletindex.SegmentIndexFileStore;
+import org.apache.carbondata.core.metadata.blocklet.DataFileFooter;
+import org.apache.carbondata.core.util.path.CarbonTablePath;
+
+import org.apache.hadoop.fs.LocatedFileStatus;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.fs.PathFilter;
+import org.apache.hadoop.fs.RemoteIterator;
+
+
+public class BlockletDataMapUtil {
+
+  private static final LogService LOGGER =
+  LogServiceFactory.getLogService(BlockletDataMapUtil.class.getName());
+
+  public static Map getBlockMetaInfoMap(
+  TableBlockIndexUniqueIdentifier identifier, SegmentIndexFileStore 
indexFileStore,
+  Set filesRead, Map 
fileNameToMetaInfoMapping)
+  throws IOException {
+if (identifier.getMergeIndexFileName() != null
+&& indexFileStore.getFileData(identifier.getIndexFileName()) == 
null) {
+  CarbonFile indexMergeFile = FileFactory.getCarbonFile(
+  identifier.getIndexFilePath() + 
CarbonCommonConstants.FILE_SEPARATOR + identifier
+  .getMergeIndexFileName());
+  if (indexMergeFile.exists() && 
!filesRead.contains(indexMergeFile.getPath())) {
+indexFileStore.readAllIIndexOfSegment(new CarbonFile[] { 
indexMergeFile });
+filesRead.add(indexMergeFile.getPath());
+  }
+}
+if (indexFileStore.getFileData(identifier.getIndexFileName()) == null) 
{
+  indexFileStore.readAllIIndexOfSegment(new CarbonFile[] { 
FileFactory.getCarbonFile(
+  identifier.getIndexFilePath() + 
CarbonCommonConstants.FILE_SEPARATOR + identifier
+  .getIndexFileName()) });
+}
+DataFileFooterConverter fileFooterConverter = new 
DataFileFooterConverter();
+Map blockMetaInfoMap = new HashMap<>();
+List indexInfo = fileFooterConverter.getIndexInfo(
+identifier.getIndexFilePath() + 
CarbonCommonConstants.FILE_SEPARATOR + identifier
+.getIndexFileName(), 
indexFileStore.getFileData(identifier.getIndexFileName()));
+for (DataFileFooter footer : indexInfo) {
+  String blockPath = 
footer.getBlockInfo().getTableBlockInfo().getFilePath();
+  if (FileFactory.isFileExist(blockPath)) {
--- End diff --

Is this if check required?.. It will make a namenode call for every 
blocklet entry here so better to remove this


---


[GitHub] carbondata pull request #2244: [CARBONDATA-2310] Refactored code to improve ...

2018-04-29 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2244#discussion_r184914643
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/indexstore/blockletindex/BlockletDataMapFactory.java
 ---
@@ -277,4 +298,30 @@ public DataMapMeta getMeta() {
 }
 return blocklets;
   }
+
+  @Override public void cache(TableBlockIndexUniqueIdentifier 
tableBlockIndexUniqueIdentifier,
+  BlockletDataMapIndexWrapper blockletDataMapIndexWrapper) throws 
IOException, MemoryException {
+cache.put(tableBlockIndexUniqueIdentifier, 
blockletDataMapIndexWrapper);
+  }
+
+  @Override
+  public List getAllUncachedDistributables(
+  List distributables) throws IOException {
+List distributablesToBeLoaded = new 
ArrayList<>(distributables.size());
+for (DataMapDistributable distributable : distributables) {
+  Segment segment = distributable.getSegment();
+  Set 
tableBlockIndexUniqueIdentifiers =
+  getTableBlockIndexUniqueIdentifiers(segment);
+  // filter out the tableBlockIndexUniqueIdentifiers based on 
distributable
+  TableBlockIndexUniqueIdentifier validIdentifier = BlockletDataMapUtil
+  
.filterIdentifiersBasedOnDistributable(tableBlockIndexUniqueIdentifiers,
+  (BlockletDataMapDistributable) distributable);
+  ((BlockletDataMapDistributable) distributable)
+  .setTableBlockIndexUniqueIdentifier(validIdentifier);
--- End diff --

Move this line inside below if check


---


[GitHub] carbondata pull request #2244: [CARBONDATA-2310] Refactored code to improve ...

2018-04-29 Thread manishgupta88
Github user manishgupta88 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2244#discussion_r184914866
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/metadata/blocklet/DataFileFooter.java
 ---
@@ -52,7 +52,7 @@
   /**
* Information about leaf nodes of all columns in this file
*/
-  private List blockletList;
+  private transient List blockletList;
--- End diff --

why transient is required here?


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184916441
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/metadata/schema/table/column/ColumnSchema.java
 ---
@@ -43,6 +43,9 @@
*/
   private static final long serialVersionUID = 7676766554874863763L;
 
+  public void columnSchema() {
--- End diff --

What is the use of this method


---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread kumarvishal09
Github user kumarvishal09 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184916437
  
--- Diff: 
processing/src/main/java/org/apache/carbondata/processing/datatypes/ArrayDataType.java
 ---
@@ -171,9 +175,9 @@ public void fillCardinality(List 
dimCardWithComplex) {
   /**
* parse byte array and bit pack
*/
-  @Override
-  public void parseAndBitPack(ByteBuffer byteArrayInput, DataOutputStream 
dataOutputStream,
-  KeyGenerator[] generator) throws IOException, KeyGenException {
+  @Override public void parseAndBitPack(ByteBuffer byteArrayInput,
+  DataOutputStream dataOutputStream, KeyGenerator[] generator)
+  throws IOException, KeyGenException {
 int dataLength = byteArrayInput.getInt();
 
--- End diff --

remove extra space


---


[GitHub] carbondata pull request #2251: [CARBONDATA-2417] SDK writer goes to infinite...

2018-04-29 Thread ajantha-bhat
GitHub user ajantha-bhat opened a pull request:

https://github.com/apache/carbondata/pull/2251

[CARBONDATA-2417] SDK writer goes to infinite wait when consumer thread is 
dead

[CARBONDATA-2417]
problem: SDK writer goes to infinite wait when consumer thread is dead

root cause: due to bad record when exception happens at consumer thread
during write, this message  is not reached producer (sdk writer).
So, SDK keeps writing data assuming consumer will consume it. But as
consumer is dead. Queue becomes full and queue.put() will be blocked
forever.

Solution: If cannot be added to queue, check whether consumer is alive
or not after every 10 seconds.
If not alive throw exception, if alive try again


 - [ ] Any interfaces changed? no
 
 - [ ] Any backward compatibility impacted? no
 
 - [ ] Document update required? no

 - [ ] Testing done
  updated the testcase in TestNonTransactionalCarbonTable.scala
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA.  NA



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/ajantha-bhat/carbondata branch3

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2251.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2251


commit 7f19372fc32a6978ac5aa87f20f346e75a9b1e9d
Author: ajantha-bhat 
Date:   2018-04-28T12:47:35Z

[CARBONDATA-2417]

problem: SDK writer goes to infinite wait when cosumer thread is dead

root cause: due to bad record when exception happens at consumer thread
during write, this message  is not reached producer (sdk writer).
So, SDK keeps writing data assuming consumer will consume it. But as
consumer is dead. Queue becomes full and queue.put() will be blocked
forever.

Solution: If cannot be added to queue, check whether consumer is alive
or not after every 10 seconds.
If not alive throw exception, if alive try again




---


[GitHub] carbondata issue #2195: [CARBONDATA-2366] fixed concurrent datamap creation ...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2195
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5524/



---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread kumarvishal09
Github user kumarvishal09 commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184916011
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/scan/complextypes/PrimitiveQueryType.java
 ---
@@ -53,6 +55,21 @@ public PrimitiveQueryType(String name, String 
parentname, int blockIndex,
 this.name = name;
 this.parentname = parentname;
 this.isDirectDictionary = isDirectDictionary;
+this.isDictionary = true;
+  }
+
+
+  public PrimitiveQueryType(String name, String parentname, int blockIndex,
--- End diff --

One more extra constructor is not required. If dictionary is not null then 
isDictionary=true


---


[GitHub] carbondata issue #2202: [CARBONDATA-2347][LUCENE]change datamap factory inte...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2202
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4359/



---


[GitHub] carbondata issue #2202: [CARBONDATA-2347][LUCENE]change datamap factory inte...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2202
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5523/



---


[jira] [Created] (CARBONDATA-2417) SDK writer goes to infinite wait when consumer thread goes dead

2018-04-29 Thread Ajantha Bhat (JIRA)
Ajantha Bhat created CARBONDATA-2417:


 Summary: SDK writer goes to infinite wait when consumer thread 
goes dead
 Key: CARBONDATA-2417
 URL: https://issues.apache.org/jira/browse/CARBONDATA-2417
 Project: CarbonData
  Issue Type: Bug
Reporter: Ajantha Bhat
Assignee: Ajantha Bhat


problem: SDK writer goes to infinite wait when cosumer thread is dead

root cause: due to bad record when exception happens at consumer thread during 
write, this messageĀ  is not reached producer (sdk writer).
So, SDK keeps writing data assuming consumer will consume it. But as consumer 
is dead. Queue becomes full and queue.put() will be blocked forever.

Solution: If cannot be added to queue, check whether consumer is alive or not 
after every 10 seconds. 
If not alive throw exception, if alive try again



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata issue #2210: [CARBONDATA-2381] Improve compaction performance by ...

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2210
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4625/



---


[GitHub] carbondata issue #2172: [CARBONDATA-2333] Add validation for insert overwrit...

2018-04-29 Thread kunal642
Github user kunal642 commented on the issue:

https://github.com/apache/carbondata/pull/2172
  
retest this please


---


[GitHub] carbondata issue #2195: [CARBONDATA-2366] fixed concurrent datamap creation ...

2018-04-29 Thread kunal642
Github user kunal642 commented on the issue:

https://github.com/apache/carbondata/pull/2195
  
retest this please



---


[GitHub] carbondata issue #1812: [CARBONDATA-2033]Support user specified segments in ...

2018-04-29 Thread manishgupta88
Github user manishgupta88 commented on the issue:

https://github.com/apache/carbondata/pull/1812
  
retest sdv please


---


[GitHub] carbondata issue #2210: [CARBONDATA-2381] Improve compaction performance by ...

2018-04-29 Thread manishgupta88
Github user manishgupta88 commented on the issue:

https://github.com/apache/carbondata/pull/2210
  
retest sdv please


---


[GitHub] carbondata issue #2209: [CARBONDATA-2388][SDK]Avro Record Complex Type Imple...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5522/



---


[GitHub] carbondata issue #2209: [CARBONDATA-2388][SDK]Avro Record Complex Type Imple...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
Build Failed with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4358/



---


[GitHub] carbondata issue #2209: [CARBONDATA-2388][SDK]Avro Record Complex Type Imple...

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4624/



---


[GitHub] carbondata issue #2209: [CARBONDATA-2388][SDK]Avro Record Complex Type Imple...

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4623/



---


[GitHub] carbondata issue #2209: [CARBONDATA-2388][SDK]Avro Record Complex Type Imple...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5521/



---


[GitHub] carbondata issue #2209: [CARBONDATA-2388][SDK]Avro Record Complex Type Imple...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
Build Failed with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4357/



---


[GitHub] carbondata issue #2209: [CARBONDATA-2388][SDK]Avro Record Complex Type Imple...

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4622/



---


[GitHub] carbondata pull request #2209: [CARBONDATA-2388][SDK]Avro Record Complex Typ...

2018-04-29 Thread ajantha-bhat
Github user ajantha-bhat commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184905554
  
--- Diff: 
core/src/main/java/org/apache/carbondata/core/metadata/schema/SchemaReader.java 
---
@@ -81,10 +81,13 @@ public static TableInfo 
getTableInfo(AbsoluteTableIdentifier identifier)
   }
 
   public static TableInfo inferSchema(AbsoluteTableIdentifier identifier,
-  boolean isCarbonFileProvider) throws IOException {
+  boolean isCarbonFileProvider, TableInfo tableInfoFromCache) throws 
IOException {
 // This routine is going to infer schema from the carbondata file 
footer
 // Convert the ColumnSchema -> TableSchema -> TableInfo.
 // Return the TableInfo.
+if (tableInfoFromCache != null) {
--- End diff --

If tableInfoFromCache is found, No need to call inferschema itself, please 
handle this outside infer schema


---


[jira] [Assigned] (CARBONDATA-2313) Support Non Transactional carbon table

2018-04-29 Thread sounak chakraborty (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2313?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

sounak chakraborty reassigned CARBONDATA-2313:
--

Assignee: sounak chakraborty

> Support Non Transactional carbon table
> --
>
> Key: CARBONDATA-2313
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2313
> Project: CarbonData
>  Issue Type: New Feature
>Reporter: Ajantha Bhat
>Assignee: sounak chakraborty
>Priority: Major
> Attachments: carbon NonTranscational Table.pdf
>
>  Time Spent: 20h 20m
>  Remaining Estimate: 0h
>
> h5. Support unmanaged carbon table



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2209: [WIP][Non Transactional Table]Avro Record Com...

2018-04-29 Thread ajantha-bhat
Github user ajantha-bhat commented on a diff in the pull request:

https://github.com/apache/carbondata/pull/2209#discussion_r184904922
  
--- Diff: 
integration/spark-common-test/src/test/scala/org/apache/carbondata/spark/testsuite/createTable/TestNonTransactionalCarbonTable.scala
 ---
@@ -175,9 +175,31 @@ class TestNonTransactionalCarbonTable extends 
QueryTest with BeforeAndAfterAll {
 
   test("test create External Table with Schema with partition, should 
ignore schema and partition")
   {
-buildTestDataSingleFile()
+sql("DROP TABLE IF EXISTS sdkOutputTable")
+
+// with partition
+sql("CREATE EXTERNAL TABLE sdkOutputTable(name string) PARTITIONED BY 
(age int) STORED BY 'carbondata' LOCATION '/home/root1/avro/files' ")
+//
+//checkAnswer(sql("select * from sdkOutputTable"), Seq(Row("robot0", 
0, 0.0),
+//  Row("robot1", 1, 0.5),
+//  Row("robot2", 2, 1.0)))
--- End diff --

please revert back this test case and add new one


---


[GitHub] carbondata issue #2209: [WIP][Non Transactional Table]Avro Record Complex Ty...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5519/



---


[GitHub] carbondata issue #2209: [WIP][Non Transactional Table]Avro Record Complex Ty...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
Build Failed with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4355/



---


[GitHub] carbondata issue #2209: [WIP][Non Transactional Table]Avro Record Complex Ty...

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4621/



---


[GitHub] carbondata issue #2209: [WIP][Non Transactional Table]Avro Record Complex Ty...

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4620/



---


[GitHub] carbondata issue #2250: [HOTFIX] Fix lucene match limit code

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2250
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4619/



---


[GitHub] carbondata issue #2250: [HOTFIX] Fix lucene match limit code

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2250
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4353/



---


[GitHub] carbondata issue #2207: [WIP] Support flat folder for managed carbon table

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2207
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4618/



---


[GitHub] carbondata issue #2250: [HOTFIX] Fix lucene match limit code

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2250
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5517/



---


[GitHub] carbondata issue #2209: [WIP][Non Transactional Table]Avro Record Complex Ty...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5518/



---


[GitHub] carbondata issue #2209: [WIP][Non Transactional Table]Avro Record Complex Ty...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2209
  
Build Failed with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4354/



---


[GitHub] carbondata issue #2207: [WIP] Support flat folder for managed carbon table

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2207
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4617/



---


[GitHub] carbondata issue #2207: [WIP] Support flat folder for managed carbon table

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2207
  
Build Failed with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4352/



---


[GitHub] carbondata issue #2207: [WIP] Support flat folder for managed carbon table

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2207
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5516/



---


[GitHub] carbondata issue #2207: [WIP] Support flat folder for managed carbon table

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2207
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4616/



---


[GitHub] carbondata pull request #2250: [HOTFIX] Fix lucene match limit code

2018-04-29 Thread ravipesala
GitHub user ravipesala opened a pull request:

https://github.com/apache/carbondata/pull/2250

[HOTFIX] Fix lucene match limit code

Problem
Currently, Lucene match limit is set as Static in `MatchExpression` it 
cannot work in concurrent scenarios.
Solution:
Change to object variable and get the match max limit from expression.

Be sure to do all of the following checklist to help us incorporate 
your contribution quickly and easily:

 - [ ] Any interfaces changed?
 
 - [ ] Any backward compatibility impacted?
 
 - [ ] Document update required?

 - [ ] Testing done
Please provide details on 
- Whether new unit test cases have been added or why no new tests 
are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance 
test report.
- Any additional information to help reviewers in testing this 
change.
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/ravipesala/incubator-carbondata 
lucene-matchlimit-fix

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2250.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2250


commit 32eb738f50b53c76ce95630278ae86216bf4223a
Author: ravipesala 
Date:   2018-04-29T15:19:18Z

Fix lucene match limit code




---


[GitHub] carbondata issue #2207: [WIP] Support flat folder for managed carbon table

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2207
  
Build Failed with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4351/



---


[GitHub] carbondata issue #2207: [WIP] Support flat folder for managed carbon table

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2207
  
Build Failed with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4350/



---


[GitHub] carbondata issue #2207: [WIP] Support flat folder for managed carbon table

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2207
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5515/



---


[GitHub] carbondata issue #2249: [WIP]sortColumns Order we are getting wrong as we se...

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2249
  
SDV Build Success , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4615/



---


[GitHub] carbondata issue #2249: [WIP]sortColumns Order we are getting wrong as we se...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2249
  
Build Success with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4349/



---


[GitHub] carbondata issue #2249: [WIP]sortColumns Order we are getting wrong as we se...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2249
  
Build Success with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5514/



---


[GitHub] carbondata pull request #2249: [WIP]sortColumns Order we are getting wrong a...

2018-04-29 Thread rahulforallp
GitHub user rahulforallp opened a pull request:

https://github.com/apache/carbondata/pull/2249

[WIP]sortColumns Order we are getting wrong as we set for external table is 
fixed



Be sure to do all of the following checklist to help us incorporate 
your contribution quickly and easily:

 - [ ] Any interfaces changed?
 
 - [ ] Any backward compatibility impacted?
 
 - [ ] Document update required?

 - [ ] Testing done
Please provide details on 
- Whether new unit test cases have been added or why no new tests 
are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance 
test report.
- Any additional information to help reviewers in testing this 
change.
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/rahulforallp/incubator-carbondata sort_col_sdk

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2249.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2249






---


[GitHub] carbondata issue #2248: [CARBONDATA-2415][WIP] Support Refresh DataMap comma...

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2248
  
SDV Build Fail , Please check CI 
http://144.76.159.231:8080/job/ApacheSDVTests/4614/



---


[jira] [Resolved] (CARBONDATA-2371) Add Profiler output in EXPLAIN command

2018-04-29 Thread Ravindra Pesala (JIRA)

 [ 
https://issues.apache.org/jira/browse/CARBONDATA-2371?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravindra Pesala resolved CARBONDATA-2371.
-
Resolution: Fixed

> Add Profiler output in EXPLAIN command
> --
>
> Key: CARBONDATA-2371
> URL: https://issues.apache.org/jira/browse/CARBONDATA-2371
> Project: CarbonData
>  Issue Type: New Feature
>Reporter: Jacky Li
>Priority: Major
> Fix For: 1.4.0
>
>  Time Spent: 8h 40m
>  Remaining Estimate: 0h
>
> More information should give in EXPLAIN command to show the effeteness of 
> datamap



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] carbondata pull request #2197: [CARBONDATA-2371] Add Profiler output in EXPL...

2018-04-29 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2197


---


[GitHub] carbondata issue #2248: [CARBONDATA-2415][WIP] Support Refresh DataMap comma...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2248
  
Build Failed  with Spark 2.1.0, Please check CI 
http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5513/



---


[GitHub] carbondata issue #2248: [CARBONDATA-2415][WIP] Support Refresh DataMap comma...

2018-04-29 Thread CarbonDataQA
Github user CarbonDataQA commented on the issue:

https://github.com/apache/carbondata/pull/2248
  
Build Failed with Spark 2.2.1, Please check CI 
http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4348/



---


[GitHub] carbondata pull request #2248: [CARBONDATA-2415][WIP] Support Refresh DataMa...

2018-04-29 Thread jackylk
GitHub user jackylk opened a pull request:

https://github.com/apache/carbondata/pull/2248

[CARBONDATA-2415][WIP] Support Refresh DataMap command for all Index datamap

WIP

 - [ ] Any interfaces changed?
 
 - [ ] Any backward compatibility impacted?
 
 - [ ] Document update required?

 - [ ] Testing done
Please provide details on 
- Whether new unit test cases have been added or why no new tests 
are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance 
test report.
- Any additional information to help reviewers in testing this 
change.
   
 - [ ] For large changes, please consider breaking it into sub-tasks under 
an umbrella JIRA. 



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/jackylk/incubator-carbondata 
index-datamap-refactory

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/carbondata/pull/2248.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #2248






---


[GitHub] carbondata issue #2197: [CARBONDATA-2371] Add Profiler output in EXPLAIN com...

2018-04-29 Thread ravipesala
Github user ravipesala commented on the issue:

https://github.com/apache/carbondata/pull/2197
  
LGTM


---


[GitHub] carbondata pull request #2238: [HOTFIX][Streaming] Enhance streaming aggrega...

2018-04-29 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/carbondata/pull/2238


---


[GitHub] carbondata issue #2238: [HOTFIX][Streaming] Enhance streaming aggregate tabl...

2018-04-29 Thread chenliang613
Github user chenliang613 commented on the issue:

https://github.com/apache/carbondata/pull/2238
  
LGTM


---