This is an automated email from the ASF dual-hosted git repository.

etudenhoefner pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/iceberg.git


The following commit(s) were added to refs/heads/main by this push:
     new 30fd752f24 Core, Spark: Remove deprecated code for 1.9.0 (#12336)
30fd752f24 is described below

commit 30fd752f24d796ce548dbeef550e875941ce6fb7
Author: Eduard Tudenhoefner <[email protected]>
AuthorDate: Thu Feb 20 15:15:03 2025 +0100

    Core, Spark: Remove deprecated code for 1.9.0 (#12336)
---
 .palantir/revapi.yml                               | 24 +++++++++++++
 .../java/org/apache/iceberg/UpdateStatistics.java  |  7 ++--
 .../java/org/apache/iceberg/MetadataUpdate.java    | 27 ++------------
 .../org/apache/iceberg/MetadataUpdateParser.java   |  8 +----
 .../java/org/apache/iceberg/SetStatistics.java     | 17 +--------
 .../java/org/apache/iceberg/TableMetadata.java     | 41 +---------------------
 .../org/apache/iceberg/TestRemoveSnapshots.java    |  2 +-
 .../java/org/apache/iceberg/TestSetStatistics.java |  4 +--
 .../java/org/apache/iceberg/TestTableMetadata.java |  4 ---
 .../org/apache/iceberg/TestUpdateRequirements.java |  2 +-
 .../iceberg/hadoop/TestCatalogUtilDropTable.java   |  2 +-
 .../extensions/TestExpireSnapshotsProcedure.java   |  4 +--
 .../extensions/TestRemoveOrphanFilesProcedure.java |  5 +--
 .../actions/ComputeTableStatsSparkAction.java      |  2 +-
 .../spark/actions/TestRemoveOrphanFilesAction.java |  2 +-
 .../spark/actions/TestRewriteTablePathsAction.java |  1 -
 .../spark/source/TestSparkMetadataColumns.java     |  5 +--
 .../spark/source/TestSparkReadProjection.java      |  3 +-
 .../apache/iceberg/spark/source/TestSparkScan.java |  8 ++---
 .../extensions/TestExpireSnapshotsProcedure.java   |  4 +--
 .../extensions/TestRemoveOrphanFilesProcedure.java |  5 +--
 .../actions/ComputeTableStatsSparkAction.java      |  2 +-
 .../spark/actions/TestRemoveOrphanFilesAction.java |  2 +-
 .../apache/iceberg/spark/source/TestSparkScan.java |  8 ++---
 24 files changed, 58 insertions(+), 131 deletions(-)

diff --git a/.palantir/revapi.yml b/.palantir/revapi.yml
index c10575a810..b15b2ce84b 100644
--- a/.palantir/revapi.yml
+++ b/.palantir/revapi.yml
@@ -1177,6 +1177,30 @@ acceptedBreaks:
       old: "method java.lang.String[] 
org.apache.iceberg.hadoop.Util::blockLocations(org.apache.iceberg.CombinedScanTask,\
         \ org.apache.hadoop.conf.Configuration)"
       justification: "Removing deprecated code"
+    - code: "java.method.removed"
+      old: "method org.apache.iceberg.TableMetadata 
org.apache.iceberg.TableMetadata::updateSchema(org.apache.iceberg.Schema,\
+        \ int)"
+      justification: "Removing deprecated code"
+    - code: "java.method.removed"
+      old: "method org.apache.iceberg.TableMetadata.Builder 
org.apache.iceberg.TableMetadata.Builder::addSchema(org.apache.iceberg.Schema,\
+        \ int)"
+      justification: "Removing deprecated code"
+    - code: "java.method.removed"
+      old: "method org.apache.iceberg.TableMetadata.Builder 
org.apache.iceberg.TableMetadata.Builder::setStatistics(long,\
+        \ org.apache.iceberg.StatisticsFile)"
+      justification: "Removing deprecated code"
+    - code: "java.method.removed"
+      old: "method org.apache.iceberg.UpdateStatistics 
org.apache.iceberg.SetStatistics::setStatistics(long,\
+        \ org.apache.iceberg.StatisticsFile)"
+      justification: "Removing deprecated code"
+    - code: "java.method.removed"
+      old: "method void 
org.apache.iceberg.MetadataUpdate.AddSchema::<init>(org.apache.iceberg.Schema,\
+        \ int)"
+      justification: "Removing deprecated code"
+    - code: "java.method.removed"
+      old: "method void 
org.apache.iceberg.MetadataUpdate.SetStatistics::<init>(long,\
+        \ org.apache.iceberg.StatisticsFile)"
+      justification: "Removing deprecated code"
     org.apache.iceberg:iceberg-parquet:
     - code: "java.class.visibilityReduced"
       old: "class org.apache.iceberg.data.parquet.BaseParquetReaders<T extends 
java.lang.Object>"
diff --git a/api/src/main/java/org/apache/iceberg/UpdateStatistics.java 
b/api/src/main/java/org/apache/iceberg/UpdateStatistics.java
index c595b16708..16e4a94bda 100644
--- a/api/src/main/java/org/apache/iceberg/UpdateStatistics.java
+++ b/api/src/main/java/org/apache/iceberg/UpdateStatistics.java
@@ -27,11 +27,12 @@ public interface UpdateStatistics extends 
PendingUpdate<List<StatisticsFile>> {
    * the snapshot if any exists.
    *
    * @return this for method chaining
-   * @deprecated since 1.8.0, will be removed 1.9.0 or 2.0.0, use {@link
-   *     #setStatistics(StatisticsFile)}.
+   * @deprecated since 1.8.0, will be removed in 2.0.0, use {@link 
#setStatistics(StatisticsFile)}.
    */
   @Deprecated
-  UpdateStatistics setStatistics(long snapshotId, StatisticsFile 
statisticsFile);
+  default UpdateStatistics setStatistics(long snapshotId, StatisticsFile 
statisticsFile) {
+    return setStatistics(statisticsFile);
+  }
 
   /**
    * Set the table's statistics file for given snapshot, replacing the 
previous statistics file for
diff --git a/core/src/main/java/org/apache/iceberg/MetadataUpdate.java 
b/core/src/main/java/org/apache/iceberg/MetadataUpdate.java
index 2f6c0c326b..3d0ad68629 100644
--- a/core/src/main/java/org/apache/iceberg/MetadataUpdate.java
+++ b/core/src/main/java/org/apache/iceberg/MetadataUpdate.java
@@ -84,21 +84,9 @@ public interface MetadataUpdate extends Serializable {
 
   class AddSchema implements MetadataUpdate {
     private final Schema schema;
-    private final int lastColumnId;
 
     public AddSchema(Schema schema) {
-      this(schema, schema.highestFieldId());
-    }
-
-    /**
-     * Set the schema
-     *
-     * @deprecated since 1.8.0, will be removed 1.9.0 or 2.0.0, use 
AddSchema(schema).
-     */
-    @Deprecated
-    public AddSchema(Schema schema, int lastColumnId) {
       this.schema = schema;
-      this.lastColumnId = lastColumnId;
     }
 
     public Schema schema() {
@@ -106,12 +94,12 @@ public interface MetadataUpdate extends Serializable {
     }
 
     public int lastColumnId() {
-      return lastColumnId;
+      return schema.highestFieldId();
     }
 
     @Override
     public void applyTo(TableMetadata.Builder metadataBuilder) {
-      metadataBuilder.addSchema(schema, lastColumnId);
+      metadataBuilder.addSchema(schema);
     }
 
     @Override
@@ -233,17 +221,6 @@ public interface MetadataUpdate extends Serializable {
   class SetStatistics implements MetadataUpdate {
     private final StatisticsFile statisticsFile;
 
-    /**
-     * Set statistics for a snapshot.
-     *
-     * @deprecated since 1.8.0, will be removed in 1.9.0 or 2.0.0, use
-     *     SetStatistics(statisticsFile).
-     */
-    @Deprecated
-    public SetStatistics(long snapshotId, StatisticsFile statisticsFile) {
-      this.statisticsFile = statisticsFile;
-    }
-
     public SetStatistics(StatisticsFile statisticsFile) {
       this.statisticsFile = statisticsFile;
     }
diff --git a/core/src/main/java/org/apache/iceberg/MetadataUpdateParser.java 
b/core/src/main/java/org/apache/iceberg/MetadataUpdateParser.java
index b985dbc089..9159da0b64 100644
--- a/core/src/main/java/org/apache/iceberg/MetadataUpdateParser.java
+++ b/core/src/main/java/org/apache/iceberg/MetadataUpdateParser.java
@@ -481,13 +481,7 @@ public class MetadataUpdateParser {
   private static MetadataUpdate readAddSchema(JsonNode node) {
     JsonNode schemaNode = JsonUtil.get(SCHEMA, node);
     Schema schema = SchemaParser.fromJson(schemaNode);
-    int lastColumnId;
-    if (node.has(LAST_COLUMN_ID)) {
-      lastColumnId = JsonUtil.getInt(LAST_COLUMN_ID, node);
-    } else {
-      lastColumnId = schema.highestFieldId();
-    }
-    return new MetadataUpdate.AddSchema(schema, lastColumnId);
+    return new MetadataUpdate.AddSchema(schema);
   }
 
   private static MetadataUpdate readSetCurrentSchema(JsonNode node) {
diff --git a/core/src/main/java/org/apache/iceberg/SetStatistics.java 
b/core/src/main/java/org/apache/iceberg/SetStatistics.java
index b5ec801e0a..ceb3fe91ba 100644
--- a/core/src/main/java/org/apache/iceberg/SetStatistics.java
+++ b/core/src/main/java/org/apache/iceberg/SetStatistics.java
@@ -21,7 +21,6 @@ package org.apache.iceberg;
 import java.util.List;
 import java.util.Map;
 import java.util.Optional;
-import org.apache.iceberg.relocated.com.google.common.base.Preconditions;
 import org.apache.iceberg.relocated.com.google.common.collect.Maps;
 
 public class SetStatistics implements UpdateStatistics {
@@ -32,20 +31,6 @@ public class SetStatistics implements UpdateStatistics {
     this.ops = ops;
   }
 
-  /**
-   * Set the statistics file for a snapshot.
-   *
-   * @deprecated since 1.8.0, will be removed in 1.9.0 or 2.0.0, use {@link
-   *     #setStatistics(StatisticsFile)}.
-   */
-  @Deprecated
-  @Override
-  public UpdateStatistics setStatistics(long snapshotId, StatisticsFile 
statisticsFile) {
-    Preconditions.checkArgument(snapshotId == statisticsFile.snapshotId());
-    statisticsToSet.put(statisticsFile.snapshotId(), 
Optional.of(statisticsFile));
-    return this;
-  }
-
   @Override
   public UpdateStatistics setStatistics(StatisticsFile statisticsFile) {
     statisticsToSet.put(statisticsFile.snapshotId(), 
Optional.of(statisticsFile));
@@ -75,7 +60,7 @@ public class SetStatistics implements UpdateStatistics {
     statisticsToSet.forEach(
         (snapshotId, statistics) -> {
           if (statistics.isPresent()) {
-            builder.setStatistics(snapshotId, statistics.get());
+            builder.setStatistics(statistics.get());
           } else {
             builder.removeStatistics(snapshotId);
           }
diff --git a/core/src/main/java/org/apache/iceberg/TableMetadata.java 
b/core/src/main/java/org/apache/iceberg/TableMetadata.java
index ea61f300b6..251f677725 100644
--- a/core/src/main/java/org/apache/iceberg/TableMetadata.java
+++ b/core/src/main/java/org/apache/iceberg/TableMetadata.java
@@ -592,16 +592,6 @@ public class TableMetadata implements Serializable {
     return nextRowId;
   }
 
-  /**
-   * Updates the schema
-   *
-   * @deprecated since 1.8.0, will be removed in 1.9.0 or 2.0.0, use 
updateSchema(schema).
-   */
-  @Deprecated
-  public TableMetadata updateSchema(Schema newSchema, int newLastColumnId) {
-    return new Builder(this).setCurrentSchema(newSchema, 
newLastColumnId).build();
-  }
-
   /** Updates the schema */
   public TableMetadata updateSchema(Schema newSchema) {
     return new Builder(this)
@@ -1141,17 +1131,6 @@ public class TableMetadata implements Serializable {
       return this;
     }
 
-    /**
-     * Add a new schema.
-     *
-     * @deprecated since 1.8.0, will be removed in 1.9.0 or 2.0.0, use 
AddSchema(schema).
-     */
-    @Deprecated
-    public Builder addSchema(Schema schema, int newLastColumnId) {
-      addSchemaInternal(schema, newLastColumnId);
-      return this;
-    }
-
     public Builder setDefaultPartitionSpec(PartitionSpec spec) {
       setDefaultPartitionSpec(addPartitionSpecInternal(spec));
       return this;
@@ -1366,24 +1345,6 @@ public class TableMetadata implements Serializable {
       return this;
     }
 
-    /**
-     * Set a statistics file for a snapshot.
-     *
-     * @deprecated since 1.8.0, will be removed 1.9.0 or 2.0.0, use 
setStatistics(statisticsFile).
-     */
-    @Deprecated
-    public Builder setStatistics(long snapshotId, StatisticsFile 
statisticsFile) {
-      Preconditions.checkNotNull(statisticsFile, "statisticsFile is null");
-      Preconditions.checkArgument(
-          snapshotId == statisticsFile.snapshotId(),
-          "snapshotId does not match: %s vs %s",
-          snapshotId,
-          statisticsFile.snapshotId());
-      statisticsFiles.put(statisticsFile.snapshotId(), 
ImmutableList.of(statisticsFile));
-      changes.add(new MetadataUpdate.SetStatistics(statisticsFile));
-      return this;
-    }
-
     public Builder setStatistics(StatisticsFile statisticsFile) {
       Preconditions.checkNotNull(statisticsFile, "statisticsFile is null");
       statisticsFiles.put(statisticsFile.snapshotId(), 
ImmutableList.of(statisticsFile));
@@ -1665,7 +1626,7 @@ public class TableMetadata implements Serializable {
         schemasById.put(newSchema.schemaId(), newSchema);
       }
 
-      changes.add(new MetadataUpdate.AddSchema(newSchema, lastColumnId));
+      changes.add(new MetadataUpdate.AddSchema(newSchema));
 
       this.lastAddedSchemaId = newSchemaId;
 
diff --git a/core/src/test/java/org/apache/iceberg/TestRemoveSnapshots.java 
b/core/src/test/java/org/apache/iceberg/TestRemoveSnapshots.java
index 6f0b2aed39..0fe76a4e51 100644
--- a/core/src/test/java/org/apache/iceberg/TestRemoveSnapshots.java
+++ b/core/src/test/java/org/apache/iceberg/TestRemoveSnapshots.java
@@ -1748,7 +1748,7 @@ public class TestRemoveSnapshots extends TestBase {
   }
 
   private void commitStats(Table table, StatisticsFile statisticsFile) {
-    table.updateStatistics().setStatistics(statisticsFile.snapshotId(), 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
   }
 
   private String statsFileLocation(String tableLocation) {
diff --git a/core/src/test/java/org/apache/iceberg/TestSetStatistics.java 
b/core/src/test/java/org/apache/iceberg/TestSetStatistics.java
index 03c644117a..4582e8f26c 100644
--- a/core/src/test/java/org/apache/iceberg/TestSetStatistics.java
+++ b/core/src/test/java/org/apache/iceberg/TestSetStatistics.java
@@ -80,7 +80,7 @@ public class TestSetStatistics extends TestBase {
                     ImmutableList.of(1, 2),
                     ImmutableMap.of("a-property", "some-property-value"))));
 
-    table.updateStatistics().setStatistics(snapshotId, 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
 
     TableMetadata metadata = readMetadata();
     assertThat(version()).isEqualTo(2);
@@ -100,7 +100,7 @@ public class TestSetStatistics extends TestBase {
         new GenericStatisticsFile(
             snapshotId, "/some/statistics/file.puffin", 100, 42, 
ImmutableList.of());
 
-    table.updateStatistics().setStatistics(snapshotId, 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
 
     TableMetadata metadata = readMetadata();
     assertThat(version()).isEqualTo(2);
diff --git a/core/src/test/java/org/apache/iceberg/TestTableMetadata.java 
b/core/src/test/java/org/apache/iceberg/TestTableMetadata.java
index 145ce6c835..1cd60fbbd1 100644
--- a/core/src/test/java/org/apache/iceberg/TestTableMetadata.java
+++ b/core/src/test/java/org/apache/iceberg/TestTableMetadata.java
@@ -1341,7 +1341,6 @@ public class TestTableMetadata {
     TableMetadata withStatistics =
         TableMetadata.buildFrom(meta)
             .setStatistics(
-                43,
                 new GenericStatisticsFile(
                     43, "/some/path/to/stats/file", 128, 27, 
ImmutableList.of()))
             .build();
@@ -1356,7 +1355,6 @@ public class TestTableMetadata {
     TableMetadata withStatisticsReplaced =
         TableMetadata.buildFrom(withStatistics)
             .setStatistics(
-                43,
                 new GenericStatisticsFile(
                     43, "/some/path/to/stats/file2", 128, 27, 
ImmutableList.of()))
             .build();
@@ -1378,11 +1376,9 @@ public class TestTableMetadata {
                 TableMetadata.newTableMetadata(
                     schema, PartitionSpec.unpartitioned(), null, 
ImmutableMap.of()))
             .setStatistics(
-                43,
                 new GenericStatisticsFile(
                     43, "/some/path/to/stats/file", 128, 27, 
ImmutableList.of()))
             .setStatistics(
-                44,
                 new GenericStatisticsFile(
                     44, "/some/path/to/stats/file2", 128, 27, 
ImmutableList.of()))
             .build();
diff --git a/core/src/test/java/org/apache/iceberg/TestUpdateRequirements.java 
b/core/src/test/java/org/apache/iceberg/TestUpdateRequirements.java
index 7d8cc471ba..3f1329d4f0 100644
--- a/core/src/test/java/org/apache/iceberg/TestUpdateRequirements.java
+++ b/core/src/test/java/org/apache/iceberg/TestUpdateRequirements.java
@@ -600,7 +600,7 @@ public class TestUpdateRequirements {
     List<UpdateRequirement> requirements =
         UpdateRequirements.forUpdateTable(
             metadata,
-            ImmutableList.of(new MetadataUpdate.SetStatistics(0L, 
mock(StatisticsFile.class))));
+            ImmutableList.of(new 
MetadataUpdate.SetStatistics(mock(StatisticsFile.class))));
     requirements.forEach(req -> req.validate(metadata));
 
     assertThat(requirements)
diff --git 
a/core/src/test/java/org/apache/iceberg/hadoop/TestCatalogUtilDropTable.java 
b/core/src/test/java/org/apache/iceberg/hadoop/TestCatalogUtilDropTable.java
index 6765b23d8a..a107a72ce6 100644
--- a/core/src/test/java/org/apache/iceberg/hadoop/TestCatalogUtilDropTable.java
+++ b/core/src/test/java/org/apache/iceberg/hadoop/TestCatalogUtilDropTable.java
@@ -66,7 +66,7 @@ public class TestCatalogUtilDropTable extends 
HadoopTableTestBase {
             table.currentSnapshot().sequenceNumber(),
             tableLocation + "/metadata/" + UUID.randomUUID() + ".stats",
             table.io());
-    table.updateStatistics().setStatistics(statisticsFile.snapshotId(), 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
 
     PartitionStatisticsFile partitionStatisticsFile =
         writePartitionStatsFile(
diff --git 
a/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
 
b/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
index d130ccbc2c..cb02f94dd0 100644
--- 
a/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
+++ 
b/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
@@ -455,7 +455,7 @@ public class TestExpireSnapshotsProcedure extends 
SparkExtensionsTestBase {
             table.currentSnapshot().sequenceNumber(),
             statsFileLocation1,
             table.io());
-    table.updateStatistics().setStatistics(statisticsFile1.snapshotId(), 
statisticsFile1).commit();
+    table.updateStatistics().setStatistics(statisticsFile1).commit();
 
     sql("INSERT INTO %s SELECT 20, 'def'", tableName);
     table.refresh();
@@ -466,7 +466,7 @@ public class TestExpireSnapshotsProcedure extends 
SparkExtensionsTestBase {
             table.currentSnapshot().sequenceNumber(),
             statsFileLocation2,
             table.io());
-    table.updateStatistics().setStatistics(statisticsFile2.snapshotId(), 
statisticsFile2).commit();
+    table.updateStatistics().setStatistics(statisticsFile2).commit();
 
     waitUntilAfter(table.currentSnapshot().timestampMillis());
 
diff --git 
a/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
 
b/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
index 2b99eecab3..fe549f20dc 100644
--- 
a/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
+++ 
b/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
@@ -493,10 +493,7 @@ public class TestRemoveOrphanFilesProcedure extends 
SparkExtensionsTestBase {
     }
 
     Transaction transaction = table.newTransaction();
-    transaction
-        .updateStatistics()
-        .setStatistics(statisticsFile.snapshotId(), statisticsFile)
-        .commit();
+    transaction.updateStatistics().setStatistics(statisticsFile).commit();
     transaction.commitTransaction();
 
     // wait to ensure files are old enough
diff --git 
a/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
 
b/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
index a508021c10..1a0c022ad2 100644
--- 
a/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
+++ 
b/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
@@ -104,7 +104,7 @@ public class ComputeTableStatsSparkAction extends 
BaseSparkAction<ComputeTableSt
         snapshotId());
     List<Blob> blobs = generateNDVBlobs();
     StatisticsFile statisticsFile = writeStatsFile(blobs);
-    table.updateStatistics().setStatistics(snapshotId(), 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
     return 
ImmutableComputeTableStats.Result.builder().statisticsFile(statisticsFile).build();
   }
 
diff --git 
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
 
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
index 60bc9ca60f..d322f1d67b 100644
--- 
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
+++ 
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
@@ -931,7 +931,7 @@ public abstract class TestRemoveOrphanFilesAction extends 
SparkTestBase {
     }
 
     Transaction transaction = table.newTransaction();
-    transaction.updateStatistics().setStatistics(snapshotId, 
statisticsFile).commit();
+    transaction.updateStatistics().setStatistics(statisticsFile).commit();
     transaction.commitTransaction();
 
     SparkActions.get()
diff --git 
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRewriteTablePathsAction.java
 
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRewriteTablePathsAction.java
index 2766078517..db1a068d4f 100644
--- 
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRewriteTablePathsAction.java
+++ 
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRewriteTablePathsAction.java
@@ -754,7 +754,6 @@ public class TestRewriteTablePathsAction extends 
SparkTestBase {
     TableMetadata withStatistics =
         TableMetadata.buildFrom(metadata)
             .setStatistics(
-                43,
                 new GenericStatisticsFile(
                     43, "/some/path/to/stats/file", 128, 27, 
ImmutableList.of()))
             .build();
diff --git 
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkMetadataColumns.java
 
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkMetadataColumns.java
index 0ba34a638a..1fd017025e 100644
--- 
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkMetadataColumns.java
+++ 
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkMetadataColumns.java
@@ -186,10 +186,7 @@ public class TestSparkMetadataColumns extends 
SparkTestBase {
 
     TableOperations ops = ((HasTableOperations) table).operations();
     TableMetadata base = ops.current();
-    ops.commit(
-        base,
-        base.updateSchema(manyColumnsSchema, 
manyColumnsSchema.highestFieldId())
-            .updatePartitionSpec(spec));
+    ops.commit(base, 
base.updateSchema(manyColumnsSchema).updatePartitionSpec(spec));
 
     Dataset<Row> df =
         spark
diff --git 
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReadProjection.java
 
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReadProjection.java
index 3a4b235c46..584a6b1c70 100644
--- 
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReadProjection.java
+++ 
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReadProjection.java
@@ -155,8 +155,7 @@ public class TestSparkReadProjection extends 
TestReadProjection {
       Schema expectedSchema = reassignIds(readSchema, idMapping);
 
       // Set the schema to the expected schema directly to simulate the table 
schema evolving
-      TestTables.replaceMetadata(
-          desc, TestTables.readMetadata(desc).updateSchema(expectedSchema, 
100));
+      TestTables.replaceMetadata(desc, 
TestTables.readMetadata(desc).updateSchema(expectedSchema));
 
       Dataset<Row> df =
           spark
diff --git 
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
 
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
index 45f5e925ca..a4f10d340e 100644
--- 
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
+++ 
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
@@ -206,7 +206,7 @@ public class TestSparkScan extends SparkTestBaseWithCatalog 
{
                     ImmutableList.of(1),
                     ImmutableMap.of("data_size", "4"))));
 
-    table.updateStatistics().setStatistics(snapshotId, 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
 
     checkColStatisticsNotReported(scan, 4L);
     withSQLConf(reportColStatsDisabled, () -> 
checkColStatisticsNotReported(scan, 4L));
@@ -259,7 +259,7 @@ public class TestSparkScan extends SparkTestBaseWithCatalog 
{
                     ImmutableList.of(1),
                     ImmutableMap.of("ndv", "4"))));
 
-    table.updateStatistics().setStatistics(snapshotId, 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
 
     checkColStatisticsNotReported(scan, 4L);
     withSQLConf(reportColStatsDisabled, () -> 
checkColStatisticsNotReported(scan, 4L));
@@ -320,7 +320,7 @@ public class TestSparkScan extends SparkTestBaseWithCatalog 
{
                     ImmutableList.of(1),
                     ImmutableMap.of("data_size", "2"))));
 
-    table.updateStatistics().setStatistics(snapshotId, 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
 
     checkColStatisticsNotReported(scan, 4L);
     withSQLConf(reportColStatsDisabled, () -> 
checkColStatisticsNotReported(scan, 4L));
@@ -380,7 +380,7 @@ public class TestSparkScan extends SparkTestBaseWithCatalog 
{
                     ImmutableList.of(2),
                     ImmutableMap.of("ndv", "2"))));
 
-    table.updateStatistics().setStatistics(snapshotId, 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
 
     checkColStatisticsNotReported(scan, 4L);
     withSQLConf(reportColStatsDisabled, () -> 
checkColStatisticsNotReported(scan, 4L));
diff --git 
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
 
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
index d3bbff5712..191216f64d 100644
--- 
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
+++ 
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
@@ -448,7 +448,7 @@ public class TestExpireSnapshotsProcedure extends 
ExtensionsTestBase {
             table.currentSnapshot().sequenceNumber(),
             statsFileLocation1,
             table.io());
-    table.updateStatistics().setStatistics(statisticsFile1.snapshotId(), 
statisticsFile1).commit();
+    table.updateStatistics().setStatistics(statisticsFile1).commit();
 
     sql("INSERT INTO %s SELECT 20, 'def'", tableName);
     table.refresh();
@@ -459,7 +459,7 @@ public class TestExpireSnapshotsProcedure extends 
ExtensionsTestBase {
             table.currentSnapshot().sequenceNumber(),
             statsFileLocation2,
             table.io());
-    table.updateStatistics().setStatistics(statisticsFile2.snapshotId(), 
statisticsFile2).commit();
+    table.updateStatistics().setStatistics(statisticsFile2).commit();
 
     waitUntilAfter(table.currentSnapshot().timestampMillis());
 
diff --git 
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
 
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
index 57d4550d52..a4aa4d8f73 100644
--- 
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
+++ 
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
@@ -481,10 +481,7 @@ public class TestRemoveOrphanFilesProcedure extends 
ExtensionsTestBase {
     }
 
     Transaction transaction = table.newTransaction();
-    transaction
-        .updateStatistics()
-        .setStatistics(statisticsFile.snapshotId(), statisticsFile)
-        .commit();
+    transaction.updateStatistics().setStatistics(statisticsFile).commit();
     transaction.commitTransaction();
 
     // wait to ensure files are old enough
diff --git 
a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
 
b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
index a508021c10..1a0c022ad2 100644
--- 
a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
+++ 
b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
@@ -104,7 +104,7 @@ public class ComputeTableStatsSparkAction extends 
BaseSparkAction<ComputeTableSt
         snapshotId());
     List<Blob> blobs = generateNDVBlobs();
     StatisticsFile statisticsFile = writeStatsFile(blobs);
-    table.updateStatistics().setStatistics(snapshotId(), 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
     return 
ImmutableComputeTableStats.Result.builder().statisticsFile(statisticsFile).build();
   }
 
diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
index a0016a5e42..022dfa1592 100644
--- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
+++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
@@ -944,7 +944,7 @@ public abstract class TestRemoveOrphanFilesAction extends 
TestBase {
     }
 
     Transaction transaction = table.newTransaction();
-    transaction.updateStatistics().setStatistics(snapshotId, 
statisticsFile).commit();
+    transaction.updateStatistics().setStatistics(statisticsFile).commit();
     transaction.commitTransaction();
 
     SparkActions.get()
diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
index dbb15ca5a7..334725ec8c 100644
--- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
+++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
@@ -224,7 +224,7 @@ public class TestSparkScan extends TestBaseWithCatalog {
                     ImmutableList.of(1),
                     ImmutableMap.of("data_size", "4"))));
 
-    table.updateStatistics().setStatistics(snapshotId, 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
 
     checkColStatisticsNotReported(scan, 4L);
     withSQLConf(reportColStatsDisabled, () -> 
checkColStatisticsNotReported(scan, 4L));
@@ -277,7 +277,7 @@ public class TestSparkScan extends TestBaseWithCatalog {
                     ImmutableList.of(1),
                     ImmutableMap.of("ndv", "4"))));
 
-    table.updateStatistics().setStatistics(snapshotId, 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
 
     checkColStatisticsNotReported(scan, 4L);
     withSQLConf(reportColStatsDisabled, () -> 
checkColStatisticsNotReported(scan, 4L));
@@ -338,7 +338,7 @@ public class TestSparkScan extends TestBaseWithCatalog {
                     ImmutableList.of(1),
                     ImmutableMap.of("data_size", "2"))));
 
-    table.updateStatistics().setStatistics(snapshotId, 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
 
     checkColStatisticsNotReported(scan, 4L);
     withSQLConf(reportColStatsDisabled, () -> 
checkColStatisticsNotReported(scan, 4L));
@@ -398,7 +398,7 @@ public class TestSparkScan extends TestBaseWithCatalog {
                     ImmutableList.of(2),
                     ImmutableMap.of("ndv", "2"))));
 
-    table.updateStatistics().setStatistics(snapshotId, 
statisticsFile).commit();
+    table.updateStatistics().setStatistics(statisticsFile).commit();
 
     checkColStatisticsNotReported(scan, 4L);
     withSQLConf(reportColStatsDisabled, () -> 
checkColStatisticsNotReported(scan, 4L));

Reply via email to