[GitHub] [carbondata] akashrn5 opened a new pull request #3911: [wip]Fix update and delete issue when multiple partition columns are present

2020-09-03 Thread GitBox


akashrn5 opened a new pull request #3911:
URL: https://github.com/apache/carbondata/pull/3911


### Why is this PR needed?
After #3837 , when there are multiple partition columns are present in 
table, update and delete not happening as , it was considering the tuple id as 
external segment tuple ID , Because when multiple partitions are present TID 
contains # character. 

### What changes were proposed in this PR?
   To double check for external segment, check if the segment metadata details  
has path is not null.
   
### Does this PR introduce any user interface change?
- No
   
### Is any new testcase added?
 - Yes
   
   
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] QiangCai commented on a change in pull request #3902: [CARBONDATA-3961] reorder filter expression based on storage ordinal

2020-09-03 Thread GitBox


QiangCai commented on a change in pull request #3902:
URL: https://github.com/apache/carbondata/pull/3902#discussion_r483365586



##
File path: 
integration/spark/src/main/scala/org/apache/spark/sql/CarbonDatasourceHadoopRelation.scala
##
@@ -72,7 +72,9 @@ case class CarbonDatasourceHadoopRelation(
   projects: Seq[NamedExpression],
   filters: Array[Filter],
   partitions: Seq[PartitionSpec]): RDD[InternalRow] = {
-val filterExpression: Option[Expression] = filters.flatMap { filter =>
+val reorderedFilter = filters.map(CarbonFilters.reorderFilter(_, 
carbonTable)).sortBy(_._2)

Review comment:
   create new method CarbonFilters.reorderFilters and check 
CarbonProperties.isFilterReorderingEnabled once

##
File path: 
integration/spark/src/main/scala/org/apache/spark/sql/CarbonDatasourceHadoopRelation.scala
##
@@ -72,7 +72,9 @@ case class CarbonDatasourceHadoopRelation(
   projects: Seq[NamedExpression],
   filters: Array[Filter],
   partitions: Seq[PartitionSpec]): RDD[InternalRow] = {
-val filterExpression: Option[Expression] = filters.flatMap { filter =>
+val reorderedFilter = filters.map(CarbonFilters.reorderFilter(_, 
carbonTable)).sortBy(_._2)
+val filterExpression: Option[Expression] = reorderedFilter.flatMap { tuple 
=>

Review comment:
   case (filter, _) =>

##
File path: 
integration/spark/src/main/scala/org/apache/spark/sql/optimizer/CarbonFilters.scala
##
@@ -373,6 +375,148 @@ object CarbonFilters {
 val carbonTable = CarbonEnv.getCarbonTable(identifier)(sparkSession)
 getPartitions(partitionFilters, sparkSession, carbonTable)
   }
+
+  def getStorageOrdinal(filter: Filter, carbonTable: CarbonTable): Int = {
+val column = filter.references.map(carbonTable.getColumnByName)
+if (column.isEmpty) {
+  -1
+} else {
+  if (column.head.isDimension) {
+column.head.getOrdinal
+  } else {
+column.head.getOrdinal + carbonTable.getAllDimensions.size()
+  }
+}
+  }
+
+  def collectSimilarExpressions(filter: Filter, table: CarbonTable): 
Seq[(Filter, Int)] = {
+filter match {
+  case sources.And(left, right) =>
+collectSimilarExpressions(left, table) ++ 
collectSimilarExpressions(right, table)
+  case sources.Or(left, right) => collectSimilarExpressions(left, table) ++
+  collectSimilarExpressions(right, table)
+  case others => Seq((others, getStorageOrdinal(others, table)))
+}
+  }
+
+  /**
+   * This method will reorder the filter based on the Storage Ordinal of the 
column references.
+   *
+   * Example1:
+   * And   And
+   *  Or  And =>OrAnd
+   *  col3  col1  col2  col1col1  col3col1   col2
+   *
+   *  **Mixed expression filter reordered locally, but wont be reordered 
globally.**
+   *
+   * Example2:
+   * And   And
+   *  And  And   =>   AndAnd
+   *  col3  col1  col2  col1col1  col1col2   col3
+   *
+   * OrOr
+   *   Or  Or =>OrOr
+   *   col3  col1  col2  col1   col1  col1col2   col3
+   *
+   *  **Similar expression filters are reordered globally**
+   *
+   * @param filter the filter expression to be reordered
+   * @return The reordered filter with the current ordinal
+   */
+  def reorderFilter(filter: Filter, table: CarbonTable): (Filter, Int) = {
+val filterMap = mutable.HashMap[String, List[(Filter, Int)]]()
+// If the filter size is one or the user has disabled reordering then no 
need to reorder.
+if (filter.references.toSet.size == 1 ||
+!CarbonProperties.isFilterReorderingEnabled) {

Review comment:
   no need checking CarbonProperties.isFilterReorderingEnabled in each 
CarbonFilters.reorderFilter 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] CarbonDataQA1 commented on pull request #3902: [CARBONDATA-3961] reorder filter expression based on storage ordinal

2020-09-03 Thread GitBox


CarbonDataQA1 commented on pull request #3902:
URL: https://github.com/apache/carbondata/pull/3902#issuecomment-686530492


   Build Success with Spark 2.4.5, Please check CI 
http://121.244.95.60:12545/job/ApacheCarbon_PR_Builder_2.4.5/2231/
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] CarbonDataQA1 commented on pull request #3902: [CARBONDATA-3961] reorder filter expression based on storage ordinal

2020-09-03 Thread GitBox


CarbonDataQA1 commented on pull request #3902:
URL: https://github.com/apache/carbondata/pull/3902#issuecomment-686524218


   Build Success with Spark 2.3.4, Please check CI 
http://121.244.95.60:12545/job/ApacheCarbonPRBuilder2.3/3971/
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] CarbonDataQA1 commented on pull request #3906: [CARBONDATA-3968]Added test cases for hive read complex types and handled other issues

2020-09-03 Thread GitBox


CarbonDataQA1 commented on pull request #3906:
URL: https://github.com/apache/carbondata/pull/3906#issuecomment-686511376


   Build Failed  with Spark 2.4.5, Please check CI 
http://121.244.95.60:12545/job/ApacheCarbon_PR_Builder_2.4.5/2228/
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] CarbonDataQA1 commented on pull request #3906: [CARBONDATA-3968]Added test cases for hive read complex types and handled other issues

2020-09-03 Thread GitBox


CarbonDataQA1 commented on pull request #3906:
URL: https://github.com/apache/carbondata/pull/3906#issuecomment-686504682


   Build Failed  with Spark 2.3.4, Please check CI 
http://121.244.95.60:12545/job/ApacheCarbonPRBuilder2.3/3969/
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] CarbonDataQA1 commented on pull request #3876: TestingCI

2020-09-03 Thread GitBox


CarbonDataQA1 commented on pull request #3876:
URL: https://github.com/apache/carbondata/pull/3876#issuecomment-686504612


   Build Success with Spark 2.4.5, Please check CI 
http://121.244.95.60:12545/job/ApacheCarbon_PR_Builder_2.4.5/2229/
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] CarbonDataQA1 commented on pull request #3876: TestingCI

2020-09-03 Thread GitBox


CarbonDataQA1 commented on pull request #3876:
URL: https://github.com/apache/carbondata/pull/3876#issuecomment-686500746


   Build Success with Spark 2.3.4, Please check CI 
http://121.244.95.60:12545/job/ApacheCarbonPRBuilder2.3/3968/
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] kunal642 commented on a change in pull request #3902: [CARBONDATA-3961] reorder filter expression based on storage ordinal

2020-09-03 Thread GitBox


kunal642 commented on a change in pull request #3902:
URL: https://github.com/apache/carbondata/pull/3902#discussion_r482936991



##
File path: 
integration/spark/src/main/scala/org/apache/spark/sql/optimizer/CarbonFilters.scala
##
@@ -373,6 +375,164 @@ object CarbonFilters {
 val carbonTable = CarbonEnv.getCarbonTable(identifier)(sparkSession)
 getPartitions(partitionFilters, sparkSession, carbonTable)
   }
+
+  def getStorageOrdinal(filter: Filter, carbonTable: CarbonTable): Int = {
+val column = filter.references.map(carbonTable.getColumnByName)
+if (column.isEmpty) {
+  -1
+} else {
+  if (column.head.isDimension) {
+column.head.getOrdinal
+  } else {
+column.head.getOrdinal + carbonTable.getAllDimensions.size()
+  }
+}
+  }
+
+  def collectSimilarExpressions(filter: Filter, table: CarbonTable): 
Seq[(Filter, Int)] = {
+filter match {
+  case sources.And(left, right) =>
+collectSimilarExpressions(left, table) ++ 
collectSimilarExpressions(right, table)
+  case sources.Or(left, right) => collectSimilarExpressions(left, table) ++
+  collectSimilarExpressions(right, table)
+  case others => Seq((others, getStorageOrdinal(others, table)))
+}
+  }
+
+  /**
+   * This method will reorder the filter based on the Storage Ordinal of the 
column references.
+   *
+   * Example1:
+   * And   And
+   *  Or  And =>OrAnd
+   *  col3  col1  col2  col1col1  col3col1   col2
+   *
+   *  **Mixed expression filter reordered locally, but wont be reordered 
globally.**
+   *
+   * Example2:
+   * And   And
+   *  And  And   =>   AndAnd
+   *  col3  col1  col2  col1col1  col1col2   col3
+   *
+   * OrOr
+   *   Or  Or =>OrOr
+   *   col3  col1  col2  col1   col1  col1col2   col3
+   *
+   *  **Similar expression filters are reordered globally**
+   *
+   * @param filter the filter expression to be reordered
+   * @return The reordered filter with the current ordinal
+   */
+  def reorderFilter(filter: Filter, table: CarbonTable): (Filter, Int) = {
+val filterMap = mutable.HashMap[String, List[(Filter, Int)]]()
+// If the filter size is one then no need to reorder.
+val sortedFilter = if (filter.references.toSet.size == 1) {
+  (filter, -1)
+} else {
+  sortFilter(filter, filterMap, table)
+}
+// If filter has only AND/PR expression then sort the nodes globally using 
the filterMap.
+// Else sort the subnodes individually
+if (!filterMap.contains("OR") && filterMap.contains("AND") && 
filterMap("AND").nonEmpty) {
+  val sortedFilterAndOrdinal = filterMap("AND").sortBy(_._2)
+  (sortedFilterAndOrdinal.map(_._1).reduce(sources.And), 
sortedFilterAndOrdinal.head._2)
+} else if (!filterMap.contains("AND") && filterMap.contains("OR") &&
+   filterMap("OR").nonEmpty) {
+  val sortedFilterAndOrdinal = filterMap("OR").sortBy(_._2)
+  (sortedFilterAndOrdinal.map(_._1).reduce(sources.Or), 
sortedFilterAndOrdinal.head._2)
+} else {
+  sortedFilter
+}
+  }
+
+  def sortFilter(filter: Filter, filterMap: mutable.HashMap[String, 
List[(Filter, Int)]],
+  table: CarbonTable): (Filter, Int) = {
+filter match {
+  case sources.And(left, right) =>

Review comment:
   done some changes regarding this..please check
   





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] kunal642 commented on a change in pull request #3902: [CARBONDATA-3961] reorder filter expression based on storage ordinal

2020-09-03 Thread GitBox


kunal642 commented on a change in pull request #3902:
URL: https://github.com/apache/carbondata/pull/3902#discussion_r482926591



##
File path: 
integration/spark/src/main/scala/org/apache/spark/sql/optimizer/CarbonFilters.scala
##
@@ -373,6 +375,164 @@ object CarbonFilters {
 val carbonTable = CarbonEnv.getCarbonTable(identifier)(sparkSession)
 getPartitions(partitionFilters, sparkSession, carbonTable)
   }
+
+  def getStorageOrdinal(filter: Filter, carbonTable: CarbonTable): Int = {
+val column = filter.references.map(carbonTable.getColumnByName)
+if (column.isEmpty) {
+  -1
+} else {
+  if (column.head.isDimension) {
+column.head.getOrdinal
+  } else {
+column.head.getOrdinal + carbonTable.getAllDimensions.size()
+  }
+}
+  }
+
+  def collectSimilarExpressions(filter: Filter, table: CarbonTable): 
Seq[(Filter, Int)] = {
+filter match {
+  case sources.And(left, right) =>
+collectSimilarExpressions(left, table) ++ 
collectSimilarExpressions(right, table)
+  case sources.Or(left, right) => collectSimilarExpressions(left, table) ++
+  collectSimilarExpressions(right, table)
+  case others => Seq((others, getStorageOrdinal(others, table)))
+}
+  }
+
+  /**
+   * This method will reorder the filter based on the Storage Ordinal of the 
column references.
+   *
+   * Example1:
+   * And   And
+   *  Or  And =>OrAnd
+   *  col3  col1  col2  col1col1  col3col1   col2
+   *
+   *  **Mixed expression filter reordered locally, but wont be reordered 
globally.**
+   *
+   * Example2:
+   * And   And
+   *  And  And   =>   AndAnd
+   *  col3  col1  col2  col1col1  col1col2   col3
+   *
+   * OrOr
+   *   Or  Or =>OrOr
+   *   col3  col1  col2  col1   col1  col1col2   col3
+   *
+   *  **Similar expression filters are reordered globally**
+   *
+   * @param filter the filter expression to be reordered
+   * @return The reordered filter with the current ordinal
+   */
+  def reorderFilter(filter: Filter, table: CarbonTable): (Filter, Int) = {
+val filterMap = mutable.HashMap[String, List[(Filter, Int)]]()
+// If the filter size is one then no need to reorder.
+val sortedFilter = if (filter.references.toSet.size == 1) {
+  (filter, -1)
+} else {
+  sortFilter(filter, filterMap, table)
+}
+// If filter has only AND/PR expression then sort the nodes globally using 
the filterMap.

Review comment:
   done





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] kunal642 commented on a change in pull request #3902: [CARBONDATA-3961] reorder filter expression based on storage ordinal

2020-09-03 Thread GitBox


kunal642 commented on a change in pull request #3902:
URL: https://github.com/apache/carbondata/pull/3902#discussion_r482926526



##
File path: 
integration/spark/src/main/scala/org/apache/spark/sql/optimizer/CarbonFilters.scala
##
@@ -373,6 +375,146 @@ object CarbonFilters {
 val carbonTable = CarbonEnv.getCarbonTable(identifier)(sparkSession)
 getPartitions(partitionFilters, sparkSession, carbonTable)
   }
+
+  def getStorageOrdinal(filter: Filter, carbonTable: CarbonTable): Int = {
+val column = filter.references.map(carbonTable.getColumnByName)
+if (column.isEmpty) {
+  -1
+} else {
+  if (column.head.isDimension) {
+column.head.getOrdinal
+  } else {
+column.head.getOrdinal + carbonTable.getAllDimensions.size()
+  }
+}
+  }
+
+  def collectSimilarExpressions(filter: Filter, table: CarbonTable): 
Seq[(Filter, Int)] = {
+filter match {
+  case sources.And(left, right) =>
+collectSimilarExpressions(left, table) ++ 
collectSimilarExpressions(right, table)
+  case sources.Or(left, right) => collectSimilarExpressions(left, table) ++
+  collectSimilarExpressions(right, table)
+  case others => Seq((others, getStorageOrdinal(others, table)))
+}
+  }
+
+  /**
+   * This method will reorder the filter based on the Storage Ordinal of the 
column references.

Review comment:
   added a property which can disabled.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] Karan980 commented on pull request #3876: TestingCI

2020-09-03 Thread GitBox


Karan980 commented on pull request #3876:
URL: https://github.com/apache/carbondata/pull/3876#issuecomment-686434740


   retest this please



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] kunal642 commented on a change in pull request #3902: [CARBONDATA-3961] reorder filter expression based on storage ordinal

2020-09-03 Thread GitBox


kunal642 commented on a change in pull request #3902:
URL: https://github.com/apache/carbondata/pull/3902#discussion_r482912253



##
File path: 
integration/spark/src/main/scala/org/apache/spark/sql/CarbonDatasourceHadoopRelation.scala
##
@@ -72,7 +72,9 @@ case class CarbonDatasourceHadoopRelation(
   projects: Seq[NamedExpression],
   filters: Array[Filter],
   partitions: Seq[PartitionSpec]): RDD[InternalRow] = {
-val filterExpression: Option[Expression] = filters.flatMap { filter =>
+val reorderedFilter = filters.map(CarbonFilters.reorderFilter(_, 
carbonTable)).sortBy(_._2)

Review comment:
   Not in this PR, many things and the implementation has to be rewritten, 
case classes has to be added to replace Tuple with.
   
   I will do it in a seperate PR.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] QiangCai commented on a change in pull request #3908: [CARBONDATA-3967] cache partition on select to enable faster pruning

2020-09-03 Thread GitBox


QiangCai commented on a change in pull request #3908:
URL: https://github.com/apache/carbondata/pull/3908#discussion_r482855055



##
File path: 
integration/spark/src/main/scala/org/apache/spark/sql/hive/CarbonSessionUtil.scala
##
@@ -103,19 +107,61 @@ object CarbonSessionUtil {
*
* @param partitionFilters
* @param sparkSession
-   * @param identifier
+   * @param carbonTable
* @return
*/
-  def prunePartitionsByFilter(partitionFilters: Seq[Expression],
+  def pruneAndCachePartitionsByFilters(partitionFilters: Seq[Expression],
   sparkSession: SparkSession,
-  identifier: TableIdentifier): Seq[CatalogTablePartition] = {
-val allPartitions = 
sparkSession.sessionState.catalog.listPartitions(identifier)
+  carbonTable: CarbonTable): Seq[CatalogTablePartition] = {
+val allPartitions = 
PartitionCacheManager.get(PartitionCacheKey(carbonTable.getTableId,
+  carbonTable.getTablePath, CarbonUtil.getExpiration_time(carbonTable)))
 ExternalCatalogUtils.prunePartitionsByFilter(
-  sparkSession.sessionState.catalog.getTableMetadata(identifier),
+  
sparkSession.sessionState.catalog.getTableMetadata(TableIdentifier(carbonTable.getTableName,
+Some(carbonTable.getDatabaseName))),
+  allPartitions.partitionSpecs,
+  partitionFilters,
+  sparkSession.sessionState.conf.sessionLocalTimeZone
+)
+  }
+
+  def prunePartitionsByFilter(partitionFilters: Seq[Expression],

Review comment:
   can we remove this method?

##
File path: 
integration/spark/src/main/scala/org/apache/spark/util/PartitionCacheManger.scala
##
@@ -0,0 +1,143 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.spark.util
+
+import java.net.URI
+import java.util
+
+import scala.collection.JavaConverters._
+
+import org.apache.log4j.Logger
+import org.apache.spark.sql.catalyst.catalog.{CatalogStorageFormat, 
CatalogTablePartition}
+
+import org.apache.carbondata.common.logging.LogServiceFactory
+import org.apache.carbondata.core.cache.{Cache, Cacheable, CarbonLRUCache}
+import org.apache.carbondata.core.constants.CarbonCommonConstants
+import org.apache.carbondata.core.datastore.impl.FileFactory
+import org.apache.carbondata.core.metadata.SegmentFileStore
+import org.apache.carbondata.core.statusmanager.SegmentStatusManager
+import org.apache.carbondata.core.util.path.CarbonTablePath
+
+object  PartitionCacheManager extends Cache[PartitionCacheKey, 
CacheablePartitionSpec] {
+
+  private val CACHE = new CarbonLRUCache(
+CarbonCommonConstants.CARBON_PARTITION_MAX_DRIVER_LRU_CACHE_SIZE,
+CarbonCommonConstants.CARBON_MAX_LRU_CACHE_SIZE_DEFAULT)
+
+  val LOGGER: Logger = LogServiceFactory.getLogService(this.getClass.getName)
+
+  def get(identifier: PartitionCacheKey): CacheablePartitionSpec = {
+val cacheablePartitionSpec =
+  CACHE.get(identifier.tableId).asInstanceOf[CacheablePartitionSpec]
+val tableStatusModifiedTime = FileFactory
+  
.getCarbonFile(CarbonTablePath.getTableStatusFilePath(identifier.tablePath))
+  .getLastModifiedTime
+if (cacheablePartitionSpec != null) {
+  if (tableStatusModifiedTime > cacheablePartitionSpec.timestamp) {
+readPartitions(identifier, tableStatusModifiedTime)
+  } else {
+cacheablePartitionSpec
+  }
+} else {
+  readPartitions(identifier, tableStatusModifiedTime)
+}
+  }
+
+  override def getAll(keys: util.List[PartitionCacheKey]):
+  util.List[CacheablePartitionSpec] = {
+keys.asScala.map(get).toList.asJava
+  }
+
+  override def getIfPresent(key: PartitionCacheKey): CacheablePartitionSpec = {
+CACHE.get(key.tableId).asInstanceOf[CacheablePartitionSpec]
+  }
+
+  override def invalidate(partitionCacheKey: PartitionCacheKey): Unit = {
+CACHE.remove(partitionCacheKey.tableId)
+  }
+
+  private def readPartitions(identifier: PartitionCacheKey, 
tableStatusModifiedTime: Long) = {

Review comment:
   now it will reload all .segment files again. can we support incremental 
loading?





This is an automated message from the Apache Git Service.
To respond to the message, please log on to 

[GitHub] [carbondata] CarbonDataQA1 commented on pull request #3876: TestingCI

2020-09-03 Thread GitBox


CarbonDataQA1 commented on pull request #3876:
URL: https://github.com/apache/carbondata/pull/3876#issuecomment-686299588


   Build Success with Spark 2.3.4, Please check CI 
http://121.244.95.60:12545/job/ApacheCarbonPRBuilder2.3/3967/
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [carbondata] CarbonDataQA1 commented on pull request #3876: TestingCI

2020-09-03 Thread GitBox


CarbonDataQA1 commented on pull request #3876:
URL: https://github.com/apache/carbondata/pull/3876#issuecomment-686296235


   Build Success with Spark 2.4.5, Please check CI 
http://121.244.95.60:12545/job/ApacheCarbon_PR_Builder_2.4.5/2227/
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org