This is an automated email from the ASF dual-hosted git repository.
etudenhoefner pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/iceberg.git
The following commit(s) were added to refs/heads/main by this push:
new 30fd752f24 Core, Spark: Remove deprecated code for 1.9.0 (#12336)
30fd752f24 is described below
commit 30fd752f24d796ce548dbeef550e875941ce6fb7
Author: Eduard Tudenhoefner <[email protected]>
AuthorDate: Thu Feb 20 15:15:03 2025 +0100
Core, Spark: Remove deprecated code for 1.9.0 (#12336)
---
.palantir/revapi.yml | 24 +++++++++++++
.../java/org/apache/iceberg/UpdateStatistics.java | 7 ++--
.../java/org/apache/iceberg/MetadataUpdate.java | 27 ++------------
.../org/apache/iceberg/MetadataUpdateParser.java | 8 +----
.../java/org/apache/iceberg/SetStatistics.java | 17 +--------
.../java/org/apache/iceberg/TableMetadata.java | 41 +---------------------
.../org/apache/iceberg/TestRemoveSnapshots.java | 2 +-
.../java/org/apache/iceberg/TestSetStatistics.java | 4 +--
.../java/org/apache/iceberg/TestTableMetadata.java | 4 ---
.../org/apache/iceberg/TestUpdateRequirements.java | 2 +-
.../iceberg/hadoop/TestCatalogUtilDropTable.java | 2 +-
.../extensions/TestExpireSnapshotsProcedure.java | 4 +--
.../extensions/TestRemoveOrphanFilesProcedure.java | 5 +--
.../actions/ComputeTableStatsSparkAction.java | 2 +-
.../spark/actions/TestRemoveOrphanFilesAction.java | 2 +-
.../spark/actions/TestRewriteTablePathsAction.java | 1 -
.../spark/source/TestSparkMetadataColumns.java | 5 +--
.../spark/source/TestSparkReadProjection.java | 3 +-
.../apache/iceberg/spark/source/TestSparkScan.java | 8 ++---
.../extensions/TestExpireSnapshotsProcedure.java | 4 +--
.../extensions/TestRemoveOrphanFilesProcedure.java | 5 +--
.../actions/ComputeTableStatsSparkAction.java | 2 +-
.../spark/actions/TestRemoveOrphanFilesAction.java | 2 +-
.../apache/iceberg/spark/source/TestSparkScan.java | 8 ++---
24 files changed, 58 insertions(+), 131 deletions(-)
diff --git a/.palantir/revapi.yml b/.palantir/revapi.yml
index c10575a810..b15b2ce84b 100644
--- a/.palantir/revapi.yml
+++ b/.palantir/revapi.yml
@@ -1177,6 +1177,30 @@ acceptedBreaks:
old: "method java.lang.String[]
org.apache.iceberg.hadoop.Util::blockLocations(org.apache.iceberg.CombinedScanTask,\
\ org.apache.hadoop.conf.Configuration)"
justification: "Removing deprecated code"
+ - code: "java.method.removed"
+ old: "method org.apache.iceberg.TableMetadata
org.apache.iceberg.TableMetadata::updateSchema(org.apache.iceberg.Schema,\
+ \ int)"
+ justification: "Removing deprecated code"
+ - code: "java.method.removed"
+ old: "method org.apache.iceberg.TableMetadata.Builder
org.apache.iceberg.TableMetadata.Builder::addSchema(org.apache.iceberg.Schema,\
+ \ int)"
+ justification: "Removing deprecated code"
+ - code: "java.method.removed"
+ old: "method org.apache.iceberg.TableMetadata.Builder
org.apache.iceberg.TableMetadata.Builder::setStatistics(long,\
+ \ org.apache.iceberg.StatisticsFile)"
+ justification: "Removing deprecated code"
+ - code: "java.method.removed"
+ old: "method org.apache.iceberg.UpdateStatistics
org.apache.iceberg.SetStatistics::setStatistics(long,\
+ \ org.apache.iceberg.StatisticsFile)"
+ justification: "Removing deprecated code"
+ - code: "java.method.removed"
+ old: "method void
org.apache.iceberg.MetadataUpdate.AddSchema::<init>(org.apache.iceberg.Schema,\
+ \ int)"
+ justification: "Removing deprecated code"
+ - code: "java.method.removed"
+ old: "method void
org.apache.iceberg.MetadataUpdate.SetStatistics::<init>(long,\
+ \ org.apache.iceberg.StatisticsFile)"
+ justification: "Removing deprecated code"
org.apache.iceberg:iceberg-parquet:
- code: "java.class.visibilityReduced"
old: "class org.apache.iceberg.data.parquet.BaseParquetReaders<T extends
java.lang.Object>"
diff --git a/api/src/main/java/org/apache/iceberg/UpdateStatistics.java
b/api/src/main/java/org/apache/iceberg/UpdateStatistics.java
index c595b16708..16e4a94bda 100644
--- a/api/src/main/java/org/apache/iceberg/UpdateStatistics.java
+++ b/api/src/main/java/org/apache/iceberg/UpdateStatistics.java
@@ -27,11 +27,12 @@ public interface UpdateStatistics extends
PendingUpdate<List<StatisticsFile>> {
* the snapshot if any exists.
*
* @return this for method chaining
- * @deprecated since 1.8.0, will be removed 1.9.0 or 2.0.0, use {@link
- * #setStatistics(StatisticsFile)}.
+ * @deprecated since 1.8.0, will be removed in 2.0.0, use {@link
#setStatistics(StatisticsFile)}.
*/
@Deprecated
- UpdateStatistics setStatistics(long snapshotId, StatisticsFile
statisticsFile);
+ default UpdateStatistics setStatistics(long snapshotId, StatisticsFile
statisticsFile) {
+ return setStatistics(statisticsFile);
+ }
/**
* Set the table's statistics file for given snapshot, replacing the
previous statistics file for
diff --git a/core/src/main/java/org/apache/iceberg/MetadataUpdate.java
b/core/src/main/java/org/apache/iceberg/MetadataUpdate.java
index 2f6c0c326b..3d0ad68629 100644
--- a/core/src/main/java/org/apache/iceberg/MetadataUpdate.java
+++ b/core/src/main/java/org/apache/iceberg/MetadataUpdate.java
@@ -84,21 +84,9 @@ public interface MetadataUpdate extends Serializable {
class AddSchema implements MetadataUpdate {
private final Schema schema;
- private final int lastColumnId;
public AddSchema(Schema schema) {
- this(schema, schema.highestFieldId());
- }
-
- /**
- * Set the schema
- *
- * @deprecated since 1.8.0, will be removed 1.9.0 or 2.0.0, use
AddSchema(schema).
- */
- @Deprecated
- public AddSchema(Schema schema, int lastColumnId) {
this.schema = schema;
- this.lastColumnId = lastColumnId;
}
public Schema schema() {
@@ -106,12 +94,12 @@ public interface MetadataUpdate extends Serializable {
}
public int lastColumnId() {
- return lastColumnId;
+ return schema.highestFieldId();
}
@Override
public void applyTo(TableMetadata.Builder metadataBuilder) {
- metadataBuilder.addSchema(schema, lastColumnId);
+ metadataBuilder.addSchema(schema);
}
@Override
@@ -233,17 +221,6 @@ public interface MetadataUpdate extends Serializable {
class SetStatistics implements MetadataUpdate {
private final StatisticsFile statisticsFile;
- /**
- * Set statistics for a snapshot.
- *
- * @deprecated since 1.8.0, will be removed in 1.9.0 or 2.0.0, use
- * SetStatistics(statisticsFile).
- */
- @Deprecated
- public SetStatistics(long snapshotId, StatisticsFile statisticsFile) {
- this.statisticsFile = statisticsFile;
- }
-
public SetStatistics(StatisticsFile statisticsFile) {
this.statisticsFile = statisticsFile;
}
diff --git a/core/src/main/java/org/apache/iceberg/MetadataUpdateParser.java
b/core/src/main/java/org/apache/iceberg/MetadataUpdateParser.java
index b985dbc089..9159da0b64 100644
--- a/core/src/main/java/org/apache/iceberg/MetadataUpdateParser.java
+++ b/core/src/main/java/org/apache/iceberg/MetadataUpdateParser.java
@@ -481,13 +481,7 @@ public class MetadataUpdateParser {
private static MetadataUpdate readAddSchema(JsonNode node) {
JsonNode schemaNode = JsonUtil.get(SCHEMA, node);
Schema schema = SchemaParser.fromJson(schemaNode);
- int lastColumnId;
- if (node.has(LAST_COLUMN_ID)) {
- lastColumnId = JsonUtil.getInt(LAST_COLUMN_ID, node);
- } else {
- lastColumnId = schema.highestFieldId();
- }
- return new MetadataUpdate.AddSchema(schema, lastColumnId);
+ return new MetadataUpdate.AddSchema(schema);
}
private static MetadataUpdate readSetCurrentSchema(JsonNode node) {
diff --git a/core/src/main/java/org/apache/iceberg/SetStatistics.java
b/core/src/main/java/org/apache/iceberg/SetStatistics.java
index b5ec801e0a..ceb3fe91ba 100644
--- a/core/src/main/java/org/apache/iceberg/SetStatistics.java
+++ b/core/src/main/java/org/apache/iceberg/SetStatistics.java
@@ -21,7 +21,6 @@ package org.apache.iceberg;
import java.util.List;
import java.util.Map;
import java.util.Optional;
-import org.apache.iceberg.relocated.com.google.common.base.Preconditions;
import org.apache.iceberg.relocated.com.google.common.collect.Maps;
public class SetStatistics implements UpdateStatistics {
@@ -32,20 +31,6 @@ public class SetStatistics implements UpdateStatistics {
this.ops = ops;
}
- /**
- * Set the statistics file for a snapshot.
- *
- * @deprecated since 1.8.0, will be removed in 1.9.0 or 2.0.0, use {@link
- * #setStatistics(StatisticsFile)}.
- */
- @Deprecated
- @Override
- public UpdateStatistics setStatistics(long snapshotId, StatisticsFile
statisticsFile) {
- Preconditions.checkArgument(snapshotId == statisticsFile.snapshotId());
- statisticsToSet.put(statisticsFile.snapshotId(),
Optional.of(statisticsFile));
- return this;
- }
-
@Override
public UpdateStatistics setStatistics(StatisticsFile statisticsFile) {
statisticsToSet.put(statisticsFile.snapshotId(),
Optional.of(statisticsFile));
@@ -75,7 +60,7 @@ public class SetStatistics implements UpdateStatistics {
statisticsToSet.forEach(
(snapshotId, statistics) -> {
if (statistics.isPresent()) {
- builder.setStatistics(snapshotId, statistics.get());
+ builder.setStatistics(statistics.get());
} else {
builder.removeStatistics(snapshotId);
}
diff --git a/core/src/main/java/org/apache/iceberg/TableMetadata.java
b/core/src/main/java/org/apache/iceberg/TableMetadata.java
index ea61f300b6..251f677725 100644
--- a/core/src/main/java/org/apache/iceberg/TableMetadata.java
+++ b/core/src/main/java/org/apache/iceberg/TableMetadata.java
@@ -592,16 +592,6 @@ public class TableMetadata implements Serializable {
return nextRowId;
}
- /**
- * Updates the schema
- *
- * @deprecated since 1.8.0, will be removed in 1.9.0 or 2.0.0, use
updateSchema(schema).
- */
- @Deprecated
- public TableMetadata updateSchema(Schema newSchema, int newLastColumnId) {
- return new Builder(this).setCurrentSchema(newSchema,
newLastColumnId).build();
- }
-
/** Updates the schema */
public TableMetadata updateSchema(Schema newSchema) {
return new Builder(this)
@@ -1141,17 +1131,6 @@ public class TableMetadata implements Serializable {
return this;
}
- /**
- * Add a new schema.
- *
- * @deprecated since 1.8.0, will be removed in 1.9.0 or 2.0.0, use
AddSchema(schema).
- */
- @Deprecated
- public Builder addSchema(Schema schema, int newLastColumnId) {
- addSchemaInternal(schema, newLastColumnId);
- return this;
- }
-
public Builder setDefaultPartitionSpec(PartitionSpec spec) {
setDefaultPartitionSpec(addPartitionSpecInternal(spec));
return this;
@@ -1366,24 +1345,6 @@ public class TableMetadata implements Serializable {
return this;
}
- /**
- * Set a statistics file for a snapshot.
- *
- * @deprecated since 1.8.0, will be removed 1.9.0 or 2.0.0, use
setStatistics(statisticsFile).
- */
- @Deprecated
- public Builder setStatistics(long snapshotId, StatisticsFile
statisticsFile) {
- Preconditions.checkNotNull(statisticsFile, "statisticsFile is null");
- Preconditions.checkArgument(
- snapshotId == statisticsFile.snapshotId(),
- "snapshotId does not match: %s vs %s",
- snapshotId,
- statisticsFile.snapshotId());
- statisticsFiles.put(statisticsFile.snapshotId(),
ImmutableList.of(statisticsFile));
- changes.add(new MetadataUpdate.SetStatistics(statisticsFile));
- return this;
- }
-
public Builder setStatistics(StatisticsFile statisticsFile) {
Preconditions.checkNotNull(statisticsFile, "statisticsFile is null");
statisticsFiles.put(statisticsFile.snapshotId(),
ImmutableList.of(statisticsFile));
@@ -1665,7 +1626,7 @@ public class TableMetadata implements Serializable {
schemasById.put(newSchema.schemaId(), newSchema);
}
- changes.add(new MetadataUpdate.AddSchema(newSchema, lastColumnId));
+ changes.add(new MetadataUpdate.AddSchema(newSchema));
this.lastAddedSchemaId = newSchemaId;
diff --git a/core/src/test/java/org/apache/iceberg/TestRemoveSnapshots.java
b/core/src/test/java/org/apache/iceberg/TestRemoveSnapshots.java
index 6f0b2aed39..0fe76a4e51 100644
--- a/core/src/test/java/org/apache/iceberg/TestRemoveSnapshots.java
+++ b/core/src/test/java/org/apache/iceberg/TestRemoveSnapshots.java
@@ -1748,7 +1748,7 @@ public class TestRemoveSnapshots extends TestBase {
}
private void commitStats(Table table, StatisticsFile statisticsFile) {
- table.updateStatistics().setStatistics(statisticsFile.snapshotId(),
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
}
private String statsFileLocation(String tableLocation) {
diff --git a/core/src/test/java/org/apache/iceberg/TestSetStatistics.java
b/core/src/test/java/org/apache/iceberg/TestSetStatistics.java
index 03c644117a..4582e8f26c 100644
--- a/core/src/test/java/org/apache/iceberg/TestSetStatistics.java
+++ b/core/src/test/java/org/apache/iceberg/TestSetStatistics.java
@@ -80,7 +80,7 @@ public class TestSetStatistics extends TestBase {
ImmutableList.of(1, 2),
ImmutableMap.of("a-property", "some-property-value"))));
- table.updateStatistics().setStatistics(snapshotId,
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
TableMetadata metadata = readMetadata();
assertThat(version()).isEqualTo(2);
@@ -100,7 +100,7 @@ public class TestSetStatistics extends TestBase {
new GenericStatisticsFile(
snapshotId, "/some/statistics/file.puffin", 100, 42,
ImmutableList.of());
- table.updateStatistics().setStatistics(snapshotId,
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
TableMetadata metadata = readMetadata();
assertThat(version()).isEqualTo(2);
diff --git a/core/src/test/java/org/apache/iceberg/TestTableMetadata.java
b/core/src/test/java/org/apache/iceberg/TestTableMetadata.java
index 145ce6c835..1cd60fbbd1 100644
--- a/core/src/test/java/org/apache/iceberg/TestTableMetadata.java
+++ b/core/src/test/java/org/apache/iceberg/TestTableMetadata.java
@@ -1341,7 +1341,6 @@ public class TestTableMetadata {
TableMetadata withStatistics =
TableMetadata.buildFrom(meta)
.setStatistics(
- 43,
new GenericStatisticsFile(
43, "/some/path/to/stats/file", 128, 27,
ImmutableList.of()))
.build();
@@ -1356,7 +1355,6 @@ public class TestTableMetadata {
TableMetadata withStatisticsReplaced =
TableMetadata.buildFrom(withStatistics)
.setStatistics(
- 43,
new GenericStatisticsFile(
43, "/some/path/to/stats/file2", 128, 27,
ImmutableList.of()))
.build();
@@ -1378,11 +1376,9 @@ public class TestTableMetadata {
TableMetadata.newTableMetadata(
schema, PartitionSpec.unpartitioned(), null,
ImmutableMap.of()))
.setStatistics(
- 43,
new GenericStatisticsFile(
43, "/some/path/to/stats/file", 128, 27,
ImmutableList.of()))
.setStatistics(
- 44,
new GenericStatisticsFile(
44, "/some/path/to/stats/file2", 128, 27,
ImmutableList.of()))
.build();
diff --git a/core/src/test/java/org/apache/iceberg/TestUpdateRequirements.java
b/core/src/test/java/org/apache/iceberg/TestUpdateRequirements.java
index 7d8cc471ba..3f1329d4f0 100644
--- a/core/src/test/java/org/apache/iceberg/TestUpdateRequirements.java
+++ b/core/src/test/java/org/apache/iceberg/TestUpdateRequirements.java
@@ -600,7 +600,7 @@ public class TestUpdateRequirements {
List<UpdateRequirement> requirements =
UpdateRequirements.forUpdateTable(
metadata,
- ImmutableList.of(new MetadataUpdate.SetStatistics(0L,
mock(StatisticsFile.class))));
+ ImmutableList.of(new
MetadataUpdate.SetStatistics(mock(StatisticsFile.class))));
requirements.forEach(req -> req.validate(metadata));
assertThat(requirements)
diff --git
a/core/src/test/java/org/apache/iceberg/hadoop/TestCatalogUtilDropTable.java
b/core/src/test/java/org/apache/iceberg/hadoop/TestCatalogUtilDropTable.java
index 6765b23d8a..a107a72ce6 100644
--- a/core/src/test/java/org/apache/iceberg/hadoop/TestCatalogUtilDropTable.java
+++ b/core/src/test/java/org/apache/iceberg/hadoop/TestCatalogUtilDropTable.java
@@ -66,7 +66,7 @@ public class TestCatalogUtilDropTable extends
HadoopTableTestBase {
table.currentSnapshot().sequenceNumber(),
tableLocation + "/metadata/" + UUID.randomUUID() + ".stats",
table.io());
- table.updateStatistics().setStatistics(statisticsFile.snapshotId(),
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
PartitionStatisticsFile partitionStatisticsFile =
writePartitionStatsFile(
diff --git
a/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
b/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
index d130ccbc2c..cb02f94dd0 100644
---
a/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
+++
b/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
@@ -455,7 +455,7 @@ public class TestExpireSnapshotsProcedure extends
SparkExtensionsTestBase {
table.currentSnapshot().sequenceNumber(),
statsFileLocation1,
table.io());
- table.updateStatistics().setStatistics(statisticsFile1.snapshotId(),
statisticsFile1).commit();
+ table.updateStatistics().setStatistics(statisticsFile1).commit();
sql("INSERT INTO %s SELECT 20, 'def'", tableName);
table.refresh();
@@ -466,7 +466,7 @@ public class TestExpireSnapshotsProcedure extends
SparkExtensionsTestBase {
table.currentSnapshot().sequenceNumber(),
statsFileLocation2,
table.io());
- table.updateStatistics().setStatistics(statisticsFile2.snapshotId(),
statisticsFile2).commit();
+ table.updateStatistics().setStatistics(statisticsFile2).commit();
waitUntilAfter(table.currentSnapshot().timestampMillis());
diff --git
a/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
b/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
index 2b99eecab3..fe549f20dc 100644
---
a/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
+++
b/spark/v3.4/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
@@ -493,10 +493,7 @@ public class TestRemoveOrphanFilesProcedure extends
SparkExtensionsTestBase {
}
Transaction transaction = table.newTransaction();
- transaction
- .updateStatistics()
- .setStatistics(statisticsFile.snapshotId(), statisticsFile)
- .commit();
+ transaction.updateStatistics().setStatistics(statisticsFile).commit();
transaction.commitTransaction();
// wait to ensure files are old enough
diff --git
a/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
b/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
index a508021c10..1a0c022ad2 100644
---
a/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
+++
b/spark/v3.4/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
@@ -104,7 +104,7 @@ public class ComputeTableStatsSparkAction extends
BaseSparkAction<ComputeTableSt
snapshotId());
List<Blob> blobs = generateNDVBlobs();
StatisticsFile statisticsFile = writeStatsFile(blobs);
- table.updateStatistics().setStatistics(snapshotId(),
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
return
ImmutableComputeTableStats.Result.builder().statisticsFile(statisticsFile).build();
}
diff --git
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
index 60bc9ca60f..d322f1d67b 100644
---
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
+++
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
@@ -931,7 +931,7 @@ public abstract class TestRemoveOrphanFilesAction extends
SparkTestBase {
}
Transaction transaction = table.newTransaction();
- transaction.updateStatistics().setStatistics(snapshotId,
statisticsFile).commit();
+ transaction.updateStatistics().setStatistics(statisticsFile).commit();
transaction.commitTransaction();
SparkActions.get()
diff --git
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRewriteTablePathsAction.java
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRewriteTablePathsAction.java
index 2766078517..db1a068d4f 100644
---
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRewriteTablePathsAction.java
+++
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/actions/TestRewriteTablePathsAction.java
@@ -754,7 +754,6 @@ public class TestRewriteTablePathsAction extends
SparkTestBase {
TableMetadata withStatistics =
TableMetadata.buildFrom(metadata)
.setStatistics(
- 43,
new GenericStatisticsFile(
43, "/some/path/to/stats/file", 128, 27,
ImmutableList.of()))
.build();
diff --git
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkMetadataColumns.java
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkMetadataColumns.java
index 0ba34a638a..1fd017025e 100644
---
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkMetadataColumns.java
+++
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkMetadataColumns.java
@@ -186,10 +186,7 @@ public class TestSparkMetadataColumns extends
SparkTestBase {
TableOperations ops = ((HasTableOperations) table).operations();
TableMetadata base = ops.current();
- ops.commit(
- base,
- base.updateSchema(manyColumnsSchema,
manyColumnsSchema.highestFieldId())
- .updatePartitionSpec(spec));
+ ops.commit(base,
base.updateSchema(manyColumnsSchema).updatePartitionSpec(spec));
Dataset<Row> df =
spark
diff --git
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReadProjection.java
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReadProjection.java
index 3a4b235c46..584a6b1c70 100644
---
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReadProjection.java
+++
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReadProjection.java
@@ -155,8 +155,7 @@ public class TestSparkReadProjection extends
TestReadProjection {
Schema expectedSchema = reassignIds(readSchema, idMapping);
// Set the schema to the expected schema directly to simulate the table
schema evolving
- TestTables.replaceMetadata(
- desc, TestTables.readMetadata(desc).updateSchema(expectedSchema,
100));
+ TestTables.replaceMetadata(desc,
TestTables.readMetadata(desc).updateSchema(expectedSchema));
Dataset<Row> df =
spark
diff --git
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
index 45f5e925ca..a4f10d340e 100644
---
a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
+++
b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
@@ -206,7 +206,7 @@ public class TestSparkScan extends SparkTestBaseWithCatalog
{
ImmutableList.of(1),
ImmutableMap.of("data_size", "4"))));
- table.updateStatistics().setStatistics(snapshotId,
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
checkColStatisticsNotReported(scan, 4L);
withSQLConf(reportColStatsDisabled, () ->
checkColStatisticsNotReported(scan, 4L));
@@ -259,7 +259,7 @@ public class TestSparkScan extends SparkTestBaseWithCatalog
{
ImmutableList.of(1),
ImmutableMap.of("ndv", "4"))));
- table.updateStatistics().setStatistics(snapshotId,
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
checkColStatisticsNotReported(scan, 4L);
withSQLConf(reportColStatsDisabled, () ->
checkColStatisticsNotReported(scan, 4L));
@@ -320,7 +320,7 @@ public class TestSparkScan extends SparkTestBaseWithCatalog
{
ImmutableList.of(1),
ImmutableMap.of("data_size", "2"))));
- table.updateStatistics().setStatistics(snapshotId,
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
checkColStatisticsNotReported(scan, 4L);
withSQLConf(reportColStatsDisabled, () ->
checkColStatisticsNotReported(scan, 4L));
@@ -380,7 +380,7 @@ public class TestSparkScan extends SparkTestBaseWithCatalog
{
ImmutableList.of(2),
ImmutableMap.of("ndv", "2"))));
- table.updateStatistics().setStatistics(snapshotId,
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
checkColStatisticsNotReported(scan, 4L);
withSQLConf(reportColStatsDisabled, () ->
checkColStatisticsNotReported(scan, 4L));
diff --git
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
index d3bbff5712..191216f64d 100644
---
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
+++
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestExpireSnapshotsProcedure.java
@@ -448,7 +448,7 @@ public class TestExpireSnapshotsProcedure extends
ExtensionsTestBase {
table.currentSnapshot().sequenceNumber(),
statsFileLocation1,
table.io());
- table.updateStatistics().setStatistics(statisticsFile1.snapshotId(),
statisticsFile1).commit();
+ table.updateStatistics().setStatistics(statisticsFile1).commit();
sql("INSERT INTO %s SELECT 20, 'def'", tableName);
table.refresh();
@@ -459,7 +459,7 @@ public class TestExpireSnapshotsProcedure extends
ExtensionsTestBase {
table.currentSnapshot().sequenceNumber(),
statsFileLocation2,
table.io());
- table.updateStatistics().setStatistics(statisticsFile2.snapshotId(),
statisticsFile2).commit();
+ table.updateStatistics().setStatistics(statisticsFile2).commit();
waitUntilAfter(table.currentSnapshot().timestampMillis());
diff --git
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
index 57d4550d52..a4aa4d8f73 100644
---
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
+++
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestRemoveOrphanFilesProcedure.java
@@ -481,10 +481,7 @@ public class TestRemoveOrphanFilesProcedure extends
ExtensionsTestBase {
}
Transaction transaction = table.newTransaction();
- transaction
- .updateStatistics()
- .setStatistics(statisticsFile.snapshotId(), statisticsFile)
- .commit();
+ transaction.updateStatistics().setStatistics(statisticsFile).commit();
transaction.commitTransaction();
// wait to ensure files are old enough
diff --git
a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
index a508021c10..1a0c022ad2 100644
---
a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
+++
b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/actions/ComputeTableStatsSparkAction.java
@@ -104,7 +104,7 @@ public class ComputeTableStatsSparkAction extends
BaseSparkAction<ComputeTableSt
snapshotId());
List<Blob> blobs = generateNDVBlobs();
StatisticsFile statisticsFile = writeStatsFile(blobs);
- table.updateStatistics().setStatistics(snapshotId(),
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
return
ImmutableComputeTableStats.Result.builder().statisticsFile(statisticsFile).build();
}
diff --git
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
index a0016a5e42..022dfa1592 100644
---
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
+++
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/actions/TestRemoveOrphanFilesAction.java
@@ -944,7 +944,7 @@ public abstract class TestRemoveOrphanFilesAction extends
TestBase {
}
Transaction transaction = table.newTransaction();
- transaction.updateStatistics().setStatistics(snapshotId,
statisticsFile).commit();
+ transaction.updateStatistics().setStatistics(statisticsFile).commit();
transaction.commitTransaction();
SparkActions.get()
diff --git
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
index dbb15ca5a7..334725ec8c 100644
---
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
+++
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkScan.java
@@ -224,7 +224,7 @@ public class TestSparkScan extends TestBaseWithCatalog {
ImmutableList.of(1),
ImmutableMap.of("data_size", "4"))));
- table.updateStatistics().setStatistics(snapshotId,
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
checkColStatisticsNotReported(scan, 4L);
withSQLConf(reportColStatsDisabled, () ->
checkColStatisticsNotReported(scan, 4L));
@@ -277,7 +277,7 @@ public class TestSparkScan extends TestBaseWithCatalog {
ImmutableList.of(1),
ImmutableMap.of("ndv", "4"))));
- table.updateStatistics().setStatistics(snapshotId,
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
checkColStatisticsNotReported(scan, 4L);
withSQLConf(reportColStatsDisabled, () ->
checkColStatisticsNotReported(scan, 4L));
@@ -338,7 +338,7 @@ public class TestSparkScan extends TestBaseWithCatalog {
ImmutableList.of(1),
ImmutableMap.of("data_size", "2"))));
- table.updateStatistics().setStatistics(snapshotId,
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
checkColStatisticsNotReported(scan, 4L);
withSQLConf(reportColStatsDisabled, () ->
checkColStatisticsNotReported(scan, 4L));
@@ -398,7 +398,7 @@ public class TestSparkScan extends TestBaseWithCatalog {
ImmutableList.of(2),
ImmutableMap.of("ndv", "2"))));
- table.updateStatistics().setStatistics(snapshotId,
statisticsFile).commit();
+ table.updateStatistics().setStatistics(statisticsFile).commit();
checkColStatisticsNotReported(scan, 4L);
withSQLConf(reportColStatsDisabled, () ->
checkColStatisticsNotReported(scan, 4L));