deniskuzZ commented on code in PR #4453:
URL: https://github.com/apache/hive/pull/4453#discussion_r1245115225


##########
iceberg/iceberg-handler/src/main/java/org/apache/iceberg/mr/hive/HiveTableUtil.java:
##########
@@ -182,24 +183,41 @@ private static List<DataFile> 
getDataFiles(RemoteIterator<LocatedFileStatus> fil
     return dataFiles;
   }
 
-  public static void appendFiles(URI fromURI, String format, Table icebergTbl, 
boolean isOverwrite, Configuration conf)
-      throws SemanticException {
+  public static void appendFiles(URI fromURI, String format, Table icebergTbl, 
boolean isOverwrite,
+      Map<String, String> partitionSpec, Configuration conf) throws 
SemanticException {
     try {
       Transaction transaction = icebergTbl.newTransaction();
       if (isOverwrite) {
         DeleteFiles delete = transaction.newDelete();
-        delete.deleteFromRowFilter(Expressions.alwaysTrue());
+        if (partitionSpec != null) {
+          for (Map.Entry<String, String> part : partitionSpec.entrySet()) {
+            delete.deleteFromRowFilter(Expressions.equal(part.getKey(),
+                
Conversions.fromPartitionString(icebergTbl.schema().findType(part.getKey()), 
part.getValue())));
+          }
+        } else {
+          delete.deleteFromRowFilter(Expressions.alwaysTrue());
+        }
         delete.commit();
       }
-      AppendFiles append = transaction.newAppend();
-      PartitionSpec spec = icebergTbl.spec();
+
       MetricsConfig metricsConfig = 
MetricsConfig.fromProperties(icebergTbl.properties());
+      PartitionSpec spec = icebergTbl.spec();
       String nameMappingString = 
icebergTbl.properties().get(TableProperties.DEFAULT_NAME_MAPPING);

Review Comment:
   up to you, but i think that is more readable
   ````   
   String nameMappingStr = 
icebergTbl.properties().get(TableProperties.DEFAULT_NAME_MAPPING);
   NameMapping nameMapping = null;
   if (nameMappingStr != null) {
      nameMapping = NameMappingParser.fromJson(nameMappingStr);
   }
   ````



##########
ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java:
##########
@@ -1183,8 +1183,14 @@ private boolean createDynPartSpec(ASTNode ast) {
       }
       return false;
     }
+

Review Comment:
   could we create
   ````
   public TableSpec(Hive db, HiveConf conf, ASTNode ast, Table tableHandle)
           throws SemanticException {
         this(db, conf, ast, true, false, tableHandle);
       }
   ````



##########
iceberg/iceberg-handler/src/main/java/org/apache/iceberg/mr/hive/HiveTableUtil.java:
##########
@@ -182,24 +183,41 @@ private static List<DataFile> 
getDataFiles(RemoteIterator<LocatedFileStatus> fil
     return dataFiles;
   }
 
-  public static void appendFiles(URI fromURI, String format, Table icebergTbl, 
boolean isOverwrite, Configuration conf)
-      throws SemanticException {
+  public static void appendFiles(URI fromURI, String format, Table icebergTbl, 
boolean isOverwrite,
+      Map<String, String> partitionSpec, Configuration conf) throws 
SemanticException {
     try {
       Transaction transaction = icebergTbl.newTransaction();
       if (isOverwrite) {
         DeleteFiles delete = transaction.newDelete();
-        delete.deleteFromRowFilter(Expressions.alwaysTrue());
+        if (partitionSpec != null) {
+          for (Map.Entry<String, String> part : partitionSpec.entrySet()) {
+            delete.deleteFromRowFilter(Expressions.equal(part.getKey(),
+                
Conversions.fromPartitionString(icebergTbl.schema().findType(part.getKey()), 
part.getValue())));

Review Comment:
   could we extract this to partKey var?
   icebergTbl.schema().findType(part.getKey()



##########
iceberg/iceberg-handler/src/main/java/org/apache/iceberg/mr/hive/HiveTableUtil.java:
##########
@@ -182,24 +183,41 @@ private static List<DataFile> 
getDataFiles(RemoteIterator<LocatedFileStatus> fil
     return dataFiles;
   }
 
-  public static void appendFiles(URI fromURI, String format, Table icebergTbl, 
boolean isOverwrite, Configuration conf)
-      throws SemanticException {
+  public static void appendFiles(URI fromURI, String format, Table icebergTbl, 
boolean isOverwrite,
+      Map<String, String> partitionSpec, Configuration conf) throws 
SemanticException {
     try {
       Transaction transaction = icebergTbl.newTransaction();
       if (isOverwrite) {
         DeleteFiles delete = transaction.newDelete();
-        delete.deleteFromRowFilter(Expressions.alwaysTrue());
+        if (partitionSpec != null) {
+          for (Map.Entry<String, String> part : partitionSpec.entrySet()) {
+            delete.deleteFromRowFilter(Expressions.equal(part.getKey(),
+                
Conversions.fromPartitionString(icebergTbl.schema().findType(part.getKey()), 
part.getValue())));
+          }
+        } else {
+          delete.deleteFromRowFilter(Expressions.alwaysTrue());
+        }
         delete.commit();
       }
-      AppendFiles append = transaction.newAppend();
-      PartitionSpec spec = icebergTbl.spec();
+
       MetricsConfig metricsConfig = 
MetricsConfig.fromProperties(icebergTbl.properties());
+      PartitionSpec spec = icebergTbl.spec();
       String nameMappingString = 
icebergTbl.properties().get(TableProperties.DEFAULT_NAME_MAPPING);
       NameMapping nameMapping = nameMappingString != null ? 
NameMappingParser.fromJson(nameMappingString) : null;
-      RemoteIterator<LocatedFileStatus> filesIterator = 
HiveTableUtil.getFilesIterator(new Path(fromURI), conf);
-      List<DataFile> dataFiles = HiveTableUtil.getDataFiles(filesIterator, 
Collections.emptyMap(),
-          format == null ? IOConstants.PARQUET : format, spec, metricsConfig, 
nameMapping, conf);
-      dataFiles.forEach(append::appendFile);
+      AppendFiles append = transaction.newAppend();
+      String actualFormat = format == null ? IOConstants.PARQUET.toLowerCase() 
: format.toLowerCase();
+      if (partitionSpec != null && !partitionSpec.isEmpty()) {
+        RemoteIterator<LocatedFileStatus> iterator = getFilesIterator(new 
Path(fromURI), conf);
+        List<DataFile> dataFiles =
+            getDataFiles(iterator, partitionSpec, actualFormat.toLowerCase(), 
spec, metricsConfig, nameMapping, conf);

Review Comment:
   to many `toLowerCase` :)



##########
ql/src/java/org/apache/hadoop/hive/ql/parse/LoadSemanticAnalyzer.java:
##########
@@ -291,18 +291,43 @@ private void analyzeLoad(ASTNode ast) throws 
SemanticException {
           ErrorMsg.INVALID_PATH.getMsg(), fromTree, e.getMessage()), e);
     }
 
+    Map<String, String> tmpPartSpec = null;
+    Table table;

Review Comment:
   String tableName = 
HiveTableName.withNoDefault(getUnescapedName((ASTNode)ast.getChild(0)));
   tableHandle = db.getTable(tableName);



##########
iceberg/iceberg-handler/src/main/java/org/apache/iceberg/mr/hive/HiveTableUtil.java:
##########
@@ -182,24 +183,41 @@ private static List<DataFile> 
getDataFiles(RemoteIterator<LocatedFileStatus> fil
     return dataFiles;
   }
 
-  public static void appendFiles(URI fromURI, String format, Table icebergTbl, 
boolean isOverwrite, Configuration conf)
-      throws SemanticException {
+  public static void appendFiles(URI fromURI, String format, Table icebergTbl, 
boolean isOverwrite,
+      Map<String, String> partitionSpec, Configuration conf) throws 
SemanticException {
     try {
       Transaction transaction = icebergTbl.newTransaction();
       if (isOverwrite) {
         DeleteFiles delete = transaction.newDelete();
-        delete.deleteFromRowFilter(Expressions.alwaysTrue());
+        if (partitionSpec != null) {
+          for (Map.Entry<String, String> part : partitionSpec.entrySet()) {
+            delete.deleteFromRowFilter(Expressions.equal(part.getKey(),
+                
Conversions.fromPartitionString(icebergTbl.schema().findType(part.getKey()), 
part.getValue())));
+          }
+        } else {
+          delete.deleteFromRowFilter(Expressions.alwaysTrue());
+        }
         delete.commit();
       }
-      AppendFiles append = transaction.newAppend();
-      PartitionSpec spec = icebergTbl.spec();
+
       MetricsConfig metricsConfig = 
MetricsConfig.fromProperties(icebergTbl.properties());
+      PartitionSpec spec = icebergTbl.spec();
       String nameMappingString = 
icebergTbl.properties().get(TableProperties.DEFAULT_NAME_MAPPING);
       NameMapping nameMapping = nameMappingString != null ? 
NameMappingParser.fromJson(nameMappingString) : null;
-      RemoteIterator<LocatedFileStatus> filesIterator = 
HiveTableUtil.getFilesIterator(new Path(fromURI), conf);
-      List<DataFile> dataFiles = HiveTableUtil.getDataFiles(filesIterator, 
Collections.emptyMap(),
-          format == null ? IOConstants.PARQUET : format, spec, metricsConfig, 
nameMapping, conf);
-      dataFiles.forEach(append::appendFile);
+      AppendFiles append = transaction.newAppend();
+      String actualFormat = format == null ? IOConstants.PARQUET.toLowerCase() 
: format.toLowerCase();

Review Comment:
   Optional.ofNullable(format).orElse(IOConstants.PARQUET).toLowerCase()



##########
iceberg/iceberg-handler/src/main/java/org/apache/iceberg/mr/hive/HiveTableUtil.java:
##########
@@ -182,24 +183,41 @@ private static List<DataFile> 
getDataFiles(RemoteIterator<LocatedFileStatus> fil
     return dataFiles;
   }
 
-  public static void appendFiles(URI fromURI, String format, Table icebergTbl, 
boolean isOverwrite, Configuration conf)
-      throws SemanticException {
+  public static void appendFiles(URI fromURI, String format, Table icebergTbl, 
boolean isOverwrite,
+      Map<String, String> partitionSpec, Configuration conf) throws 
SemanticException {
     try {
       Transaction transaction = icebergTbl.newTransaction();
       if (isOverwrite) {
         DeleteFiles delete = transaction.newDelete();
-        delete.deleteFromRowFilter(Expressions.alwaysTrue());
+        if (partitionSpec != null) {
+          for (Map.Entry<String, String> part : partitionSpec.entrySet()) {
+            delete.deleteFromRowFilter(Expressions.equal(part.getKey(),
+                
Conversions.fromPartitionString(icebergTbl.schema().findType(part.getKey()), 
part.getValue())));
+          }
+        } else {
+          delete.deleteFromRowFilter(Expressions.alwaysTrue());
+        }
         delete.commit();
       }
-      AppendFiles append = transaction.newAppend();
-      PartitionSpec spec = icebergTbl.spec();
+
       MetricsConfig metricsConfig = 
MetricsConfig.fromProperties(icebergTbl.properties());
+      PartitionSpec spec = icebergTbl.spec();
       String nameMappingString = 
icebergTbl.properties().get(TableProperties.DEFAULT_NAME_MAPPING);
       NameMapping nameMapping = nameMappingString != null ? 
NameMappingParser.fromJson(nameMappingString) : null;
-      RemoteIterator<LocatedFileStatus> filesIterator = 
HiveTableUtil.getFilesIterator(new Path(fromURI), conf);
-      List<DataFile> dataFiles = HiveTableUtil.getDataFiles(filesIterator, 
Collections.emptyMap(),
-          format == null ? IOConstants.PARQUET : format, spec, metricsConfig, 
nameMapping, conf);
-      dataFiles.forEach(append::appendFile);
+      AppendFiles append = transaction.newAppend();
+      String actualFormat = format == null ? IOConstants.PARQUET.toLowerCase() 
: format.toLowerCase();
+      if (partitionSpec != null && !partitionSpec.isEmpty()) {

Review Comment:
   remove if, code is the same and handle nulls in getDataFiles
   ````
   if(partitionSpec == null) {
   partitionSpec = Collections.emptyMap()
   }
   ````



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to