This is an automated email from the ASF dual-hosted git repository.
apurtell pushed a commit to branch branch-2.5
in repository https://gitbox.apache.org/repos/asf/hbase.git
The following commit(s) were added to refs/heads/branch-2.5 by this push:
new dc781bd2b5b HBASE-29455 Spotless upgrade breaks builds with Java 8
(branch-2.5)
dc781bd2b5b is described below
commit dc781bd2b5b655a03c8e350c2c35e8182b741d4d
Author: Andrew Purtell <[email protected]>
AuthorDate: Wed Jul 16 14:44:00 2025 -0700
HBASE-29455 Spotless upgrade breaks builds with Java 8 (branch-2.5)
Revert "HBASE-29344 Update spotless to 2.44.4 (#7012)"
This reverts commit 17421e67502aca0e410d032f14711a1bdae8e50b.
---
.../org/apache/hadoop/hbase/client/AsyncTable.java | 2 +-
.../java/org/apache/hadoop/hbase/ipc/IPCUtil.java | 2 +-
.../wal/ProcedureWALPerformanceEvaluation.java | 1 -
.../org/apache/hadoop/hbase/master/HMaster.java | 2 +-
.../apache/hadoop/hbase/regionserver/CellSet.java | 2 +-
.../apache/hadoop/hbase/regionserver/HRegion.java | 1 -
.../hbase/regionserver/handler/HandlerUtil.java | 3 +--
.../normalizer/TestSimpleRegionNormalizer.java | 4 ++--
.../quotas/TestFileSystemUtilizationChore.java | 2 +-
.../hadoop/hbase/regionserver/TestBulkLoad.java | 24 +++++++++++-----------
.../hbase/regionserver/TestStripeStoreEngine.java | 2 +-
.../security/token/TestFsDelegationToken.java | 2 +-
pom.xml | 16 +--------------
13 files changed, 23 insertions(+), 40 deletions(-)
diff --git
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncTable.java
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncTable.java
index 9477618c730..2bd6b50df22 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncTable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncTable.java
@@ -212,7 +212,7 @@ public interface AsyncTable<C extends
ScanResultConsumerBase> {
Preconditions.checkNotNull(family, "family is null");
return increment(
new Increment(row).addColumn(family, qualifier,
amount).setDurability(durability))
- .thenApply(r -> Bytes.toLong(r.getValue(family, qualifier)));
+ .thenApply(r -> Bytes.toLong(r.getValue(family, qualifier)));
}
/**
diff --git
a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
index 3831a7f7831..d861ad0c3cc 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
@@ -202,7 +202,7 @@ class IPCUtil {
// connection refused; include the host:port in the error
return (IOException) new ConnectException(
"Call to " + getCallTarget(addr, regionInfo) + " failed on connection
exception: " + error)
- .initCause(error);
+ .initCause(error);
} else if (error instanceof SocketTimeoutException) {
return (IOException) new SocketTimeoutException(
"Call to " + getCallTarget(addr, regionInfo) + " failed because " +
error).initCause(error);
diff --git
a/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPerformanceEvaluation.java
b/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPerformanceEvaluation.java
index 5d9287d0758..f9bad936ee2 100644
---
a/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPerformanceEvaluation.java
+++
b/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPerformanceEvaluation.java
@@ -93,7 +93,6 @@ public class ProcedureWALPerformanceEvaluation
"Starting new log : " +
store.getActiveLogs().get(store.getActiveLogs().size() - 1));
}
}
-
///////////////////////////////
// HELPER CLASSES
///////////////////////////////
diff --git
a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
index 1e9b0c94e05..f9e75a0438f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
@@ -2151,7 +2151,7 @@ public class HMaster extends HRegionServer implements
MasterServices {
final Set<TableName> matchingTables = getTableDescriptors(new
LinkedList<>(),
ntfp.getNamespace(), ntfp.getRegex(), ntfp.getTableNames(),
false).stream()
- .map(TableDescriptor::getTableName).collect(Collectors.toSet());
+ .map(TableDescriptor::getTableName).collect(Collectors.toSet());
final Set<TableName> allEnabledTables =
tableStateManager.getTablesInStates(TableState.State.ENABLED);
final List<TableName> targetTables =
diff --git
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CellSet.java
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CellSet.java
index cbe8deb3211..4890c8a8449 100644
---
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CellSet.java
+++
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CellSet.java
@@ -44,7 +44,7 @@ public class CellSet implements NavigableSet<Cell> {
// is not already present.", this implementation "Adds the specified element
to this set EVEN
// if it is already present overwriting what was there previous".
// Otherwise, has same attributes as ConcurrentSkipListSet
- private final NavigableMap<Cell, Cell> delegatee;
+ private final NavigableMap<Cell, Cell> delegatee; ///
private final int numUniqueKeys;
diff --git
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
index d98af859702..ca0e7db716d 100644
---
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
+++
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
@@ -2106,7 +2106,6 @@ public class HRegion implements HeapSize,
PropagatingConfigurationObserver, Regi
// These methods are meant to be called periodically by the HRegionServer for
// upkeep.
//////////////////////////////////////////////////////////////////////////////
-
/**
* Do preparation for pending compaction.
*/
diff --git
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/handler/HandlerUtil.java
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/handler/HandlerUtil.java
index 58684e35c9e..43051893376 100644
---
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/handler/HandlerUtil.java
+++
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/handler/HandlerUtil.java
@@ -36,7 +36,6 @@ final class HandlerUtil {
return new RetryCounterFactory(
new RetryCounter.RetryConfig().setBackoffPolicy(new
RetryCounter.ExponentialBackoffPolicy())
.setSleepInterval(100).setMaxSleepTime(30000).setMaxAttempts(Integer.MAX_VALUE)
- .setTimeUnit(TimeUnit.MILLISECONDS).setJitter(0.01f))
- .create();
+ .setTimeUnit(TimeUnit.MILLISECONDS).setJitter(0.01f)).create();
}
}
diff --git
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizer.java
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizer.java
index e931716e77e..9c593035fee 100644
---
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizer.java
+++
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizer.java
@@ -144,7 +144,7 @@ public class TestSimpleRegionNormalizer {
setupMocksForNormalizer(regionSizes, regionInfos);
when(
masterServices.getAssignmentManager().getRegionStates().getRegionState(any(RegionInfo.class)))
- .thenReturn(RegionState.createForTesting(null, state));
+ .thenReturn(RegionState.createForTesting(null, state));
assertThat(normalizer.getMergeMinRegionCount(),
greaterThanOrEqualTo(regionInfos.size()));
List<NormalizationPlan> plans =
normalizer.computePlansForTable(tableDescriptor);
@@ -680,7 +680,7 @@ public class TestSimpleRegionNormalizer {
.thenReturn(sn);
when(
masterServices.getAssignmentManager().getRegionStates().getRegionState(any(RegionInfo.class)))
- .thenReturn(RegionState.createForTesting(null, RegionState.State.OPEN));
+ .thenReturn(RegionState.createForTesting(null,
RegionState.State.OPEN));
for (Map.Entry<byte[], Integer> region : regionSizes.entrySet()) {
RegionMetrics regionLoad = Mockito.mock(RegionMetrics.class);
diff --git
a/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestFileSystemUtilizationChore.java
b/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestFileSystemUtilizationChore.java
index b834ddf8c33..529088c9c0c 100644
---
a/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestFileSystemUtilizationChore.java
+++
b/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestFileSystemUtilizationChore.java
@@ -263,7 +263,7 @@ public class TestFileSystemUtilizationChore {
final FileSystemUtilizationChore chore = new
FileSystemUtilizationChore(rs);
doAnswer(
new ExpectedRegionSizeSummationAnswer(sum(Arrays.asList(r1HFileSizeSum,
r2HFileSizeSum))))
- .when(rs).reportRegionSizesForQuotas(any(RegionSizeStore.class));
+ .when(rs).reportRegionSizesForQuotas(any(RegionSizeStore.class));
final Region r1 = mockRegionWithHFileLinks(r1StoreFileSizes, r1HFileSizes);
final Region r2 = mockRegionWithHFileLinks(r2StoreFileSizes, r2HFileSizes);
diff --git
a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestBulkLoad.java
b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestBulkLoad.java
index 85a5b5d920f..a62ea0dcd97 100644
---
a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestBulkLoad.java
+++
b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestBulkLoad.java
@@ -67,18 +67,18 @@ public class TestBulkLoad extends TestBulkloadBase {
storeFileNames.add(storeFileName);
when(log.appendMarker(any(), any(),
argThat(bulkLogWalEdit(WALEdit.BULK_LOAD, tableName.toBytes(),
familyName, storeFileNames))))
- .thenAnswer(new Answer() {
- @Override
- public Object answer(InvocationOnMock invocation) {
- WALKeyImpl walKey = invocation.getArgument(1);
- MultiVersionConcurrencyControl mvcc = walKey.getMvcc();
- if (mvcc != null) {
- MultiVersionConcurrencyControl.WriteEntry we = mvcc.begin();
- walKey.setWriteEntry(we);
- }
- return 01L;
- };
- });
+ .thenAnswer(new Answer() {
+ @Override
+ public Object answer(InvocationOnMock invocation) {
+ WALKeyImpl walKey = invocation.getArgument(1);
+ MultiVersionConcurrencyControl mvcc = walKey.getMvcc();
+ if (mvcc != null) {
+ MultiVersionConcurrencyControl.WriteEntry we = mvcc.begin();
+ walKey.setWriteEntry(we);
+ }
+ return 01L;
+ };
+ });
testRegionWithFamiliesAndSpecifiedTableName(tableName,
family1).bulkLoadHFiles(familyPaths,
false, null);
verify(log).sync(anyLong());
diff --git
a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java
b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java
index 6ca392a9bfd..1493c3e04ce 100644
---
a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java
+++
b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java
@@ -81,7 +81,7 @@ public class TestStripeStoreEngine {
se.setCompactorOverride(mockCompactor);
when(
mockCompactor.compact(any(), anyInt(), anyLong(), any(), any(), any(),
any(), any(), any()))
- .thenReturn(new ArrayList<>());
+ .thenReturn(new ArrayList<>());
// Produce 3 L0 files.
HStoreFile sf = createFile();
diff --git
a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestFsDelegationToken.java
b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestFsDelegationToken.java
index 2385fdfb1e9..3efb944def4 100644
---
a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestFsDelegationToken.java
+++
b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestFsDelegationToken.java
@@ -74,7 +74,7 @@ public class TestFsDelegationToken {
.thenReturn(webhdfsToken);
when(
user.getToken(SWEBHDFS_TOKEN_KIND.toString(),
swebHdfsFileSystem.getCanonicalServiceName()))
- .thenReturn(swebhdfsToken);
+ .thenReturn(swebhdfsToken);
when(hdfsToken.getKind()).thenReturn(new Text("HDFS_DELEGATION_TOKEN"));
when(webhdfsToken.getKind()).thenReturn(WEBHDFS_TOKEN_KIND);
when(swebhdfsToken.getKind()).thenReturn(SWEBHDFS_TOKEN_KIND);
diff --git a/pom.xml b/pom.xml
index 611efbeb7f9..bfed437b176 100644
--- a/pom.xml
+++ b/pom.xml
@@ -659,7 +659,7 @@
<surefire.version>3.1.0</surefire.version>
<wagon.ssh.version>2.12</wagon.ssh.version>
<xml.maven.version>1.0.1</xml.maven.version>
- <spotless.version>2.44.4</spotless.version>
+ <spotless.version>2.27.2</spotless.version>
<maven-site.version>3.12.0</maven-site.version>
<!-- compression -->
<aircompressor.version>0.27</aircompressor.version>
@@ -2543,20 +2543,6 @@
<artifactId>spotless-maven-plugin</artifactId>
<version>${spotless.version}</version>
<configuration>
- <upToDateChecking>
- <!--
- we need to process some files twice with spotless so we need to
disable up to date
- check, see https://github.com/diffplug/spotless/issues/1767
- -->
- <enabled>false</enabled>
- </upToDateChecking>
- <!--
- On Windows, spotless might convert the line endings of these files
to LF.
- We're instructing it to respect the line ending definitions in
.gitattributes
- for these files.
- For more details, please refer to HBASE-29388.
- -->
- <lineEndings>GIT_ATTRIBUTES</lineEndings>
<!-- define a language-specific format -->
<java>
<excludes>