This is an automated email from the ASF dual-hosted git repository.
zhangduo pushed a commit to branch branch-2
in repository https://gitbox.apache.org/repos/asf/hbase.git
The following commit(s) were added to refs/heads/branch-2 by this push:
new fdf9665eaae HBASE-29344 Update spotless to 2.44.4 (#7010)
fdf9665eaae is described below
commit fdf9665eaae3da2761e68518a071fbd16ed9f264
Author: Duo Zhang <[email protected]>
AuthorDate: Thu May 22 10:09:19 2025 +0800
HBASE-29344 Update spotless to 2.44.4 (#7010)
Signed-off-by: Nick Dimiduk <[email protected]>
Signed-off-by: Nihal Jain <[email protected]>
---
.../apache/hadoop/hbase/backup/TestBackupBase.java | 2 +-
.../org/apache/hadoop/hbase/client/AsyncTable.java | 2 +-
.../java/org/apache/hadoop/hbase/ipc/IPCUtil.java | 2 +-
.../hbase/io/crypto/tls/TestHBaseTrustManager.java | 6 +++---
.../wal/ProcedureWALPerformanceEvaluation.java | 1 +
.../org/apache/hadoop/hbase/master/HMaster.java | 2 +-
.../apache/hadoop/hbase/regionserver/CellSet.java | 2 +-
.../apache/hadoop/hbase/regionserver/HRegion.java | 1 +
.../hbase/regionserver/handler/HandlerUtil.java | 3 ++-
.../normalizer/TestSimpleRegionNormalizer.java | 4 ++--
.../quotas/TestFileSystemUtilizationChore.java | 2 +-
.../hadoop/hbase/regionserver/TestBulkLoad.java | 24 +++++++++++-----------
.../hbase/regionserver/TestStripeStoreEngine.java | 2 +-
.../security/token/TestFsDelegationToken.java | 2 +-
pom.xml | 9 +++++++-
15 files changed, 37 insertions(+), 27 deletions(-)
diff --git
a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupBase.java
b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupBase.java
index db8c30bdba9..2a0be003cab 100644
---
a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupBase.java
+++
b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupBase.java
@@ -337,7 +337,7 @@ public class TestBackupBase {
if (useSecondCluster) {
BACKUP_REMOTE_ROOT_DIR = new Path(
new Path(TEST_UTIL2.getConfiguration().get("fs.defaultFS")) +
BACKUP_REMOTE_ROOT_DIR)
- .toString();
+ .toString();
LOG.info("REMOTE ROOTDIR " + BACKUP_REMOTE_ROOT_DIR);
}
createTables();
diff --git
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncTable.java
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncTable.java
index 385ffecd716..83308761fa2 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncTable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncTable.java
@@ -224,7 +224,7 @@ public interface AsyncTable<C extends
ScanResultConsumerBase> {
Preconditions.checkNotNull(family, "family is null");
return increment(
new Increment(row).addColumn(family, qualifier,
amount).setDurability(durability))
- .thenApply(r -> Bytes.toLong(r.getValue(family, qualifier)));
+ .thenApply(r -> Bytes.toLong(r.getValue(family, qualifier)));
}
/**
diff --git
a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
index 42094eb45e0..3058f0caee4 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/IPCUtil.java
@@ -218,7 +218,7 @@ class IPCUtil {
// connection refused; include the host:port in the error
return (IOException) new ConnectException(
"Call to " + getCallTarget(addr, regionInfo) + " failed on connection
exception: " + error)
- .initCause(error);
+ .initCause(error);
} else if (error instanceof SocketTimeoutException) {
return (IOException) new SocketTimeoutException(
"Call to " + getCallTarget(addr, regionInfo) + " failed because " +
error).initCause(error);
diff --git
a/hbase-common/src/test/java/org/apache/hadoop/hbase/io/crypto/tls/TestHBaseTrustManager.java
b/hbase-common/src/test/java/org/apache/hadoop/hbase/io/crypto/tls/TestHBaseTrustManager.java
index cfe197ae8d3..31a443758f9 100644
---
a/hbase-common/src/test/java/org/apache/hadoop/hbase/io/crypto/tls/TestHBaseTrustManager.java
+++
b/hbase-common/src/test/java/org/apache/hadoop/hbase/io/crypto/tls/TestHBaseTrustManager.java
@@ -153,9 +153,9 @@ public class TestHBaseTrustManager {
X509v3CertificateBuilder certificateBuilder =
new JcaX509v3CertificateBuilder(nameBuilder.build(), serialNumber,
notBefore, notAfter,
nameBuilder.build(), keyPair.getPublic())
- .addExtension(Extension.basicConstraints, true, new
BasicConstraints(0))
- .addExtension(Extension.keyUsage, true,
- new KeyUsage(KeyUsage.digitalSignature | KeyUsage.keyCertSign |
KeyUsage.cRLSign));
+ .addExtension(Extension.basicConstraints, true, new
BasicConstraints(0))
+ .addExtension(Extension.keyUsage, true,
+ new KeyUsage(KeyUsage.digitalSignature | KeyUsage.keyCertSign |
KeyUsage.cRLSign));
List<GeneralName> generalNames = new ArrayList<>();
if (ipAddress != null) {
diff --git
a/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPerformanceEvaluation.java
b/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPerformanceEvaluation.java
index f9bad936ee2..5d9287d0758 100644
---
a/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPerformanceEvaluation.java
+++
b/hbase-procedure/src/test/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPerformanceEvaluation.java
@@ -93,6 +93,7 @@ public class ProcedureWALPerformanceEvaluation
"Starting new log : " +
store.getActiveLogs().get(store.getActiveLogs().size() - 1));
}
}
+
///////////////////////////////
// HELPER CLASSES
///////////////////////////////
diff --git
a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
index 314f7b5cb35..817977ff896 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
@@ -2186,7 +2186,7 @@ public class HMaster extends HRegionServer implements
MasterServices {
final Set<TableName> matchingTables = getTableDescriptors(new
LinkedList<>(),
ntfp.getNamespace(), ntfp.getRegex(), ntfp.getTableNames(),
false).stream()
- .map(TableDescriptor::getTableName).collect(Collectors.toSet());
+ .map(TableDescriptor::getTableName).collect(Collectors.toSet());
final Set<TableName> allEnabledTables =
tableStateManager.getTablesInStates(TableState.State.ENABLED);
final List<TableName> targetTables =
diff --git
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CellSet.java
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CellSet.java
index c8d9b5b2ea6..f9c015596de 100644
---
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CellSet.java
+++
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CellSet.java
@@ -44,7 +44,7 @@ public class CellSet<T extends Cell> implements
NavigableSet<T> {
// is not already present.", this implementation "Adds the specified element
to this set EVEN
// if it is already present overwriting what was there previous".
// Otherwise, has same attributes as ConcurrentSkipListSet
- private final NavigableMap<T, T> delegatee; ///
+ private final NavigableMap<T, T> delegatee;
private final int numUniqueKeys;
diff --git
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
index 83c0018b98d..d9dbc950a0f 100644
---
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
+++
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
@@ -2199,6 +2199,7 @@ public class HRegion implements HeapSize,
PropagatingConfigurationObserver, Regi
// These methods are meant to be called periodically by the HRegionServer for
// upkeep.
//////////////////////////////////////////////////////////////////////////////
+
/**
* Do preparation for pending compaction.
*/
diff --git
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/handler/HandlerUtil.java
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/handler/HandlerUtil.java
index 43051893376..58684e35c9e 100644
---
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/handler/HandlerUtil.java
+++
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/handler/HandlerUtil.java
@@ -36,6 +36,7 @@ final class HandlerUtil {
return new RetryCounterFactory(
new RetryCounter.RetryConfig().setBackoffPolicy(new
RetryCounter.ExponentialBackoffPolicy())
.setSleepInterval(100).setMaxSleepTime(30000).setMaxAttempts(Integer.MAX_VALUE)
- .setTimeUnit(TimeUnit.MILLISECONDS).setJitter(0.01f)).create();
+ .setTimeUnit(TimeUnit.MILLISECONDS).setJitter(0.01f))
+ .create();
}
}
diff --git
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizer.java
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizer.java
index 9c593035fee..e931716e77e 100644
---
a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizer.java
+++
b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/normalizer/TestSimpleRegionNormalizer.java
@@ -144,7 +144,7 @@ public class TestSimpleRegionNormalizer {
setupMocksForNormalizer(regionSizes, regionInfos);
when(
masterServices.getAssignmentManager().getRegionStates().getRegionState(any(RegionInfo.class)))
- .thenReturn(RegionState.createForTesting(null, state));
+ .thenReturn(RegionState.createForTesting(null, state));
assertThat(normalizer.getMergeMinRegionCount(),
greaterThanOrEqualTo(regionInfos.size()));
List<NormalizationPlan> plans =
normalizer.computePlansForTable(tableDescriptor);
@@ -680,7 +680,7 @@ public class TestSimpleRegionNormalizer {
.thenReturn(sn);
when(
masterServices.getAssignmentManager().getRegionStates().getRegionState(any(RegionInfo.class)))
- .thenReturn(RegionState.createForTesting(null,
RegionState.State.OPEN));
+ .thenReturn(RegionState.createForTesting(null, RegionState.State.OPEN));
for (Map.Entry<byte[], Integer> region : regionSizes.entrySet()) {
RegionMetrics regionLoad = Mockito.mock(RegionMetrics.class);
diff --git
a/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestFileSystemUtilizationChore.java
b/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestFileSystemUtilizationChore.java
index 529088c9c0c..b834ddf8c33 100644
---
a/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestFileSystemUtilizationChore.java
+++
b/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestFileSystemUtilizationChore.java
@@ -263,7 +263,7 @@ public class TestFileSystemUtilizationChore {
final FileSystemUtilizationChore chore = new
FileSystemUtilizationChore(rs);
doAnswer(
new ExpectedRegionSizeSummationAnswer(sum(Arrays.asList(r1HFileSizeSum,
r2HFileSizeSum))))
- .when(rs).reportRegionSizesForQuotas(any(RegionSizeStore.class));
+ .when(rs).reportRegionSizesForQuotas(any(RegionSizeStore.class));
final Region r1 = mockRegionWithHFileLinks(r1StoreFileSizes, r1HFileSizes);
final Region r2 = mockRegionWithHFileLinks(r2StoreFileSizes, r2HFileSizes);
diff --git
a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestBulkLoad.java
b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestBulkLoad.java
index a62ea0dcd97..85a5b5d920f 100644
---
a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestBulkLoad.java
+++
b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestBulkLoad.java
@@ -67,18 +67,18 @@ public class TestBulkLoad extends TestBulkloadBase {
storeFileNames.add(storeFileName);
when(log.appendMarker(any(), any(),
argThat(bulkLogWalEdit(WALEdit.BULK_LOAD, tableName.toBytes(),
familyName, storeFileNames))))
- .thenAnswer(new Answer() {
- @Override
- public Object answer(InvocationOnMock invocation) {
- WALKeyImpl walKey = invocation.getArgument(1);
- MultiVersionConcurrencyControl mvcc = walKey.getMvcc();
- if (mvcc != null) {
- MultiVersionConcurrencyControl.WriteEntry we = mvcc.begin();
- walKey.setWriteEntry(we);
- }
- return 01L;
- };
- });
+ .thenAnswer(new Answer() {
+ @Override
+ public Object answer(InvocationOnMock invocation) {
+ WALKeyImpl walKey = invocation.getArgument(1);
+ MultiVersionConcurrencyControl mvcc = walKey.getMvcc();
+ if (mvcc != null) {
+ MultiVersionConcurrencyControl.WriteEntry we = mvcc.begin();
+ walKey.setWriteEntry(we);
+ }
+ return 01L;
+ };
+ });
testRegionWithFamiliesAndSpecifiedTableName(tableName,
family1).bulkLoadHFiles(familyPaths,
false, null);
verify(log).sync(anyLong());
diff --git
a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java
b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java
index 1493c3e04ce..6ca392a9bfd 100644
---
a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java
+++
b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestStripeStoreEngine.java
@@ -81,7 +81,7 @@ public class TestStripeStoreEngine {
se.setCompactorOverride(mockCompactor);
when(
mockCompactor.compact(any(), anyInt(), anyLong(), any(), any(), any(),
any(), any(), any()))
- .thenReturn(new ArrayList<>());
+ .thenReturn(new ArrayList<>());
// Produce 3 L0 files.
HStoreFile sf = createFile();
diff --git
a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestFsDelegationToken.java
b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestFsDelegationToken.java
index 3efb944def4..2385fdfb1e9 100644
---
a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestFsDelegationToken.java
+++
b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestFsDelegationToken.java
@@ -74,7 +74,7 @@ public class TestFsDelegationToken {
.thenReturn(webhdfsToken);
when(
user.getToken(SWEBHDFS_TOKEN_KIND.toString(),
swebHdfsFileSystem.getCanonicalServiceName()))
- .thenReturn(swebhdfsToken);
+ .thenReturn(swebhdfsToken);
when(hdfsToken.getKind()).thenReturn(new Text("HDFS_DELEGATION_TOKEN"));
when(webhdfsToken.getKind()).thenReturn(WEBHDFS_TOKEN_KIND);
when(swebhdfsToken.getKind()).thenReturn(SWEBHDFS_TOKEN_KIND);
diff --git a/pom.xml b/pom.xml
index ca2623374b4..e317dd12fbf 100644
--- a/pom.xml
+++ b/pom.xml
@@ -656,7 +656,7 @@
<surefire.version>3.1.0</surefire.version>
<wagon.ssh.version>2.12</wagon.ssh.version>
<xml.maven.version>1.0.1</xml.maven.version>
- <spotless.version>2.27.2</spotless.version>
+ <spotless.version>2.44.4</spotless.version>
<maven-site.version>3.12.0</maven-site.version>
<!-- compression -->
<aircompressor.version>0.27</aircompressor.version>
@@ -2530,6 +2530,13 @@
<artifactId>spotless-maven-plugin</artifactId>
<version>${spotless.version}</version>
<configuration>
+ <upToDateChecking>
+ <!--
+ we need to process some files twice with spotless so we need to
disable up to date
+ check, see https://github.com/diffplug/spotless/issues/1767
+ -->
+ <enabled>false</enabled>
+ </upToDateChecking>
<!-- define a language-specific format -->
<java>
<excludes>