This is an automated email from the ASF dual-hosted git repository. sureshanaparti pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/cloudstack.git
The following commit(s) were added to refs/heads/main by this push: new 7c32bd2506e Fixup main build errors (#9330) 7c32bd2506e is described below commit 7c32bd2506eb14348f2be22d6773d15d646ee809 Author: Vishesh <vishes...@gmail.com> AuthorDate: Thu Jul 4 13:00:37 2024 +0530 Fixup main build errors (#9330) * Fixup main build errors * Fixup flaky test * Address comments --- .../BasePrimaryDataStoreLifeCycleImpl.java | 14 +++--- .../api/response/QuotaResponseBuilderImplTest.java | 2 +- .../LibvirtCheckConvertInstanceCommandWrapper.java | 8 +--- .../LibvirtPrepareStorageClientCommandWrapper.java | 6 +-- ...ibvirtUnprepareStorageClientCommandWrapper.java | 6 +-- .../lifecycle/AdaptiveDataStoreLifeCycleImpl.java | 4 -- .../ElastistorPrimaryDataStoreLifeCycle.java | 5 -- .../lifecycle/DateraPrimaryDataStoreLifeCycle.java | 4 -- .../CloudStackPrimaryDataStoreLifeCycleImpl.java | 3 -- .../LinstorPrimaryDataStoreLifeCycleImpl.java | 4 -- .../NexentaPrimaryDataStoreLifeCycle.java | 3 -- .../ScaleIOPrimaryDataStoreLifeCycle.java | 4 -- .../datastore/manager/ScaleIOSDCManagerImpl.java | 53 +++++++++++----------- .../SolidFirePrimaryDataStoreLifeCycle.java | 5 -- .../SolidFireSharedPrimaryDataStoreLifeCycle.java | 5 -- .../StorPoolPrimaryDataStoreLifeCycle.java | 4 -- 16 files changed, 40 insertions(+), 90 deletions(-) diff --git a/engine/storage/volume/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/BasePrimaryDataStoreLifeCycleImpl.java b/engine/storage/volume/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/BasePrimaryDataStoreLifeCycleImpl.java index adc74a77d43..1ee4d40a567 100644 --- a/engine/storage/volume/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/BasePrimaryDataStoreLifeCycleImpl.java +++ b/engine/storage/volume/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/BasePrimaryDataStoreLifeCycleImpl.java @@ -25,7 +25,6 @@ import javax.inject.Inject; import org.apache.cloudstack.engine.subsystem.api.storage.ClusterScope; import org.apache.cloudstack.engine.subsystem.api.storage.DataStore; import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper; -import org.apache.log4j.Logger; import com.cloud.agent.AgentManager; import com.cloud.agent.api.Answer; @@ -39,9 +38,12 @@ import com.cloud.storage.StoragePool; import com.cloud.storage.StoragePoolHostVO; import com.cloud.storage.dao.StoragePoolHostDao; import com.cloud.utils.Pair; +import org.apache.logging.log4j.LogManager; +import org.apache.logging.log4j.Logger; public class BasePrimaryDataStoreLifeCycleImpl { - private static final Logger s_logger = Logger.getLogger(BasePrimaryDataStoreLifeCycleImpl.class); + protected Logger logger = LogManager.getLogger(getClass()); + @Inject AgentManager agentMgr; @Inject @@ -70,13 +72,13 @@ public class BasePrimaryDataStoreLifeCycleImpl { public void changeStoragePoolScopeToZone(DataStore store, ClusterScope clusterScope, HypervisorType hypervisorType) { List<HostVO> hosts = getPoolHostsList(clusterScope, hypervisorType); - s_logger.debug("Changing scope of the storage pool to Zone"); + logger.debug("Changing scope of the storage pool to Zone"); if (hosts != null) { for (HostVO host : hosts) { try { storageMgr.connectHostToSharedPool(host.getId(), store.getId()); } catch (Exception e) { - s_logger.warn("Unable to establish a connection between " + host + " and " + store, e); + logger.warn("Unable to establish a connection between " + host + " and " + store, e); } } } @@ -85,7 +87,7 @@ public class BasePrimaryDataStoreLifeCycleImpl { public void changeStoragePoolScopeToCluster(DataStore store, ClusterScope clusterScope, HypervisorType hypervisorType) { Pair<List<StoragePoolHostVO>, Integer> hostPoolRecords = storagePoolHostDao.listByPoolIdNotInCluster(clusterScope.getScopeId(), store.getId()); - s_logger.debug("Changing scope of the storage pool to Cluster"); + logger.debug("Changing scope of the storage pool to Cluster"); if (hostPoolRecords.second() > 0) { StoragePool pool = (StoragePool) store; for (StoragePoolHostVO host : hostPoolRecords.first()) { @@ -94,7 +96,7 @@ public class BasePrimaryDataStoreLifeCycleImpl { if (answer != null) { if (!answer.getResult()) { - s_logger.debug("Failed to delete storage pool: " + answer.getResult()); + logger.debug("Failed to delete storage pool: " + answer.getResult()); } else if (HypervisorType.KVM != hypervisorType) { break; } diff --git a/plugins/database/quota/src/test/java/org/apache/cloudstack/api/response/QuotaResponseBuilderImplTest.java b/plugins/database/quota/src/test/java/org/apache/cloudstack/api/response/QuotaResponseBuilderImplTest.java index 6bca4ea85bb..da02b6d3709 100644 --- a/plugins/database/quota/src/test/java/org/apache/cloudstack/api/response/QuotaResponseBuilderImplTest.java +++ b/plugins/database/quota/src/test/java/org/apache/cloudstack/api/response/QuotaResponseBuilderImplTest.java @@ -334,7 +334,7 @@ public class QuotaResponseBuilderImplTest extends TestCase { @Test public void validateEndDateOnCreatingNewQuotaTariffTestSetValidEndDate() { Date startDate = DateUtils.addDays(date, -100); - Date endDate = DateUtils.addMilliseconds(new Date(), 1); + Date endDate = DateUtils.addMinutes(new Date(), 1); quotaResponseBuilderSpy.validateEndDateOnCreatingNewQuotaTariff(quotaTariffVoMock, startDate, endDate); Mockito.verify(quotaTariffVoMock).setEndDate(Mockito.any(Date.class)); diff --git a/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtCheckConvertInstanceCommandWrapper.java b/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtCheckConvertInstanceCommandWrapper.java index d94fddeeb44..d3ebb28b106 100644 --- a/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtCheckConvertInstanceCommandWrapper.java +++ b/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtCheckConvertInstanceCommandWrapper.java @@ -18,8 +18,6 @@ // package com.cloud.hypervisor.kvm.resource.wrapper; -import org.apache.log4j.Logger; - import com.cloud.agent.api.Answer; import com.cloud.agent.api.CheckConvertInstanceAnswer; import com.cloud.agent.api.CheckConvertInstanceCommand; @@ -30,21 +28,19 @@ import com.cloud.resource.ResourceWrapper; @ResourceWrapper(handles = CheckConvertInstanceCommand.class) public class LibvirtCheckConvertInstanceCommandWrapper extends CommandWrapper<CheckConvertInstanceCommand, Answer, LibvirtComputingResource> { - private static final Logger s_logger = Logger.getLogger(LibvirtCheckConvertInstanceCommandWrapper.class); - @Override public Answer execute(CheckConvertInstanceCommand cmd, LibvirtComputingResource serverResource) { if (!serverResource.hostSupportsInstanceConversion()) { String msg = String.format("Cannot convert the instance from VMware as the virt-v2v binary is not found on host %s. " + "Please install virt-v2v%s on the host before attempting the instance conversion.", serverResource.getPrivateIp(), serverResource.isUbuntuHost()? ", nbdkit" : ""); - s_logger.info(msg); + logger.info(msg); return new CheckConvertInstanceAnswer(cmd, false, msg); } if (cmd.getCheckWindowsGuestConversionSupport() && !serverResource.hostSupportsWindowsGuestConversion()) { String msg = String.format("Cannot convert the instance from VMware as the virtio-win package is not found on host %s. " + "Please install virtio-win package on the host before attempting the windows guest instance conversion.", serverResource.getPrivateIp()); - s_logger.info(msg); + logger.info(msg); return new CheckConvertInstanceAnswer(cmd, false, msg); } diff --git a/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtPrepareStorageClientCommandWrapper.java b/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtPrepareStorageClientCommandWrapper.java index 79afd4696b0..b5cab17ecb1 100644 --- a/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtPrepareStorageClientCommandWrapper.java +++ b/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtPrepareStorageClientCommandWrapper.java @@ -21,8 +21,6 @@ package com.cloud.hypervisor.kvm.resource.wrapper; import java.util.Map; -import org.apache.log4j.Logger; - import com.cloud.agent.api.Answer; import com.cloud.agent.api.PrepareStorageClientAnswer; import com.cloud.agent.api.PrepareStorageClientCommand; @@ -35,15 +33,13 @@ import com.cloud.utils.Ternary; @ResourceWrapper(handles = PrepareStorageClientCommand.class) public class LibvirtPrepareStorageClientCommandWrapper extends CommandWrapper<PrepareStorageClientCommand, Answer, LibvirtComputingResource> { - private static final Logger s_logger = Logger.getLogger(LibvirtPrepareStorageClientCommandWrapper.class); - @Override public Answer execute(PrepareStorageClientCommand cmd, LibvirtComputingResource libvirtComputingResource) { final KVMStoragePoolManager storagePoolMgr = libvirtComputingResource.getStoragePoolMgr(); Ternary<Boolean, Map<String, String>, String> prepareStorageClientResult = storagePoolMgr.prepareStorageClient(cmd.getPoolType(), cmd.getPoolUuid(), cmd.getDetails()); if (!prepareStorageClientResult.first()) { String msg = prepareStorageClientResult.third(); - s_logger.debug("Unable to prepare storage client, due to: " + msg); + logger.debug("Unable to prepare storage client, due to: " + msg); return new PrepareStorageClientAnswer(cmd, false, msg); } Map<String, String> details = prepareStorageClientResult.second(); diff --git a/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtUnprepareStorageClientCommandWrapper.java b/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtUnprepareStorageClientCommandWrapper.java index f98782fe748..2f23a934003 100644 --- a/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtUnprepareStorageClientCommandWrapper.java +++ b/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtUnprepareStorageClientCommandWrapper.java @@ -19,8 +19,6 @@ package com.cloud.hypervisor.kvm.resource.wrapper; -import org.apache.log4j.Logger; - import com.cloud.agent.api.Answer; import com.cloud.agent.api.UnprepareStorageClientAnswer; import com.cloud.agent.api.UnprepareStorageClientCommand; @@ -33,15 +31,13 @@ import com.cloud.utils.Pair; @ResourceWrapper(handles = UnprepareStorageClientCommand.class) public class LibvirtUnprepareStorageClientCommandWrapper extends CommandWrapper<UnprepareStorageClientCommand, Answer, LibvirtComputingResource> { - private static final Logger s_logger = Logger.getLogger(LibvirtUnprepareStorageClientCommandWrapper.class); - @Override public Answer execute(UnprepareStorageClientCommand cmd, LibvirtComputingResource libvirtComputingResource) { final KVMStoragePoolManager storagePoolMgr = libvirtComputingResource.getStoragePoolMgr(); Pair<Boolean, String> unprepareStorageClientResult = storagePoolMgr.unprepareStorageClient(cmd.getPoolType(), cmd.getPoolUuid()); if (!unprepareStorageClientResult.first()) { String msg = unprepareStorageClientResult.second(); - s_logger.debug("Couldn't unprepare storage client, due to: " + msg); + logger.debug("Couldn't unprepare storage client, due to: " + msg); return new UnprepareStorageClientAnswer(cmd, false, msg); } return new UnprepareStorageClientAnswer(cmd, true); diff --git a/plugins/storage/volume/adaptive/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/AdaptiveDataStoreLifeCycleImpl.java b/plugins/storage/volume/adaptive/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/AdaptiveDataStoreLifeCycleImpl.java index 86c3bfa91de..f6ace68e2e8 100644 --- a/plugins/storage/volume/adaptive/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/AdaptiveDataStoreLifeCycleImpl.java +++ b/plugins/storage/volume/adaptive/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/AdaptiveDataStoreLifeCycleImpl.java @@ -57,8 +57,6 @@ import com.cloud.storage.Storage.StoragePoolType; import com.cloud.utils.crypt.DBEncryptionUtil; import com.cloud.utils.exception.CloudRuntimeException; import com.cloud.host.Host; -import org.apache.logging.log4j.LogManager; -import org.apache.logging.log4j.Logger; /** * Manages the lifecycle of a Managed Data Store in CloudStack @@ -66,8 +64,6 @@ import org.apache.logging.log4j.Logger; public class AdaptiveDataStoreLifeCycleImpl extends BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle { @Inject private PrimaryDataStoreDao _storagePoolDao; - protected Logger logger = LogManager.getLogger(getClass()); - @Inject PrimaryDataStoreHelper _dataStoreHelper; @Inject diff --git a/plugins/storage/volume/cloudbyte/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ElastistorPrimaryDataStoreLifeCycle.java b/plugins/storage/volume/cloudbyte/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ElastistorPrimaryDataStoreLifeCycle.java index 01058d717ed..5d7f2d86e14 100644 --- a/plugins/storage/volume/cloudbyte/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ElastistorPrimaryDataStoreLifeCycle.java +++ b/plugins/storage/volume/cloudbyte/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ElastistorPrimaryDataStoreLifeCycle.java @@ -26,9 +26,6 @@ import java.util.StringTokenizer; import javax.inject.Inject; -import org.apache.logging.log4j.Logger; -import org.apache.logging.log4j.LogManager; - import org.apache.cloudstack.engine.subsystem.api.storage.ClusterScope; import org.apache.cloudstack.engine.subsystem.api.storage.DataStore; import org.apache.cloudstack.engine.subsystem.api.storage.HostScope; @@ -67,8 +64,6 @@ import com.cloud.storage.dao.StoragePoolHostDao; import com.cloud.utils.exception.CloudRuntimeException; public class ElastistorPrimaryDataStoreLifeCycle extends BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle { - protected Logger logger = LogManager.getLogger(getClass()); - @Inject HostDao _hostDao; @Inject diff --git a/plugins/storage/volume/datera/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/DateraPrimaryDataStoreLifeCycle.java b/plugins/storage/volume/datera/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/DateraPrimaryDataStoreLifeCycle.java index 3944cdf7bbe..6d222b36b85 100644 --- a/plugins/storage/volume/datera/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/DateraPrimaryDataStoreLifeCycle.java +++ b/plugins/storage/volume/datera/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/DateraPrimaryDataStoreLifeCycle.java @@ -51,8 +51,6 @@ import org.apache.cloudstack.storage.datastore.db.PrimaryDataStoreDao; import org.apache.cloudstack.storage.datastore.db.StoragePoolVO; import org.apache.cloudstack.storage.datastore.util.DateraUtil; import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper; -import org.apache.logging.log4j.Logger; -import org.apache.logging.log4j.LogManager; import javax.inject.Inject; import java.util.ArrayList; @@ -60,8 +58,6 @@ import java.util.List; import java.util.Map; public class DateraPrimaryDataStoreLifeCycle extends BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle { - protected Logger logger = LogManager.getLogger(getClass()); - @Inject private CapacityManager _capacityMgr; @Inject diff --git a/plugins/storage/volume/default/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/CloudStackPrimaryDataStoreLifeCycleImpl.java b/plugins/storage/volume/default/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/CloudStackPrimaryDataStoreLifeCycleImpl.java index 9600f0618ac..bc66e2ff136 100644 --- a/plugins/storage/volume/default/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/CloudStackPrimaryDataStoreLifeCycleImpl.java +++ b/plugins/storage/volume/default/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/CloudStackPrimaryDataStoreLifeCycleImpl.java @@ -64,8 +64,6 @@ import org.apache.cloudstack.engine.subsystem.api.storage.ZoneScope; import org.apache.cloudstack.storage.datastore.db.PrimaryDataStoreDao; import org.apache.cloudstack.storage.datastore.db.StoragePoolVO; import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper; -import org.apache.logging.log4j.Logger; -import org.apache.logging.log4j.LogManager; import javax.inject.Inject; import java.util.ArrayList; @@ -74,7 +72,6 @@ import java.util.Map; import java.util.UUID; public class CloudStackPrimaryDataStoreLifeCycleImpl extends BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle { - protected Logger logger = LogManager.getLogger(getClass()); @Inject protected ResourceManager _resourceMgr; @Inject diff --git a/plugins/storage/volume/linstor/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/LinstorPrimaryDataStoreLifeCycleImpl.java b/plugins/storage/volume/linstor/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/LinstorPrimaryDataStoreLifeCycleImpl.java index 7b3f8dbb37e..ac2563cd436 100644 --- a/plugins/storage/volume/linstor/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/LinstorPrimaryDataStoreLifeCycleImpl.java +++ b/plugins/storage/volume/linstor/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/LinstorPrimaryDataStoreLifeCycleImpl.java @@ -51,12 +51,8 @@ import org.apache.cloudstack.storage.datastore.db.PrimaryDataStoreDao; import org.apache.cloudstack.storage.datastore.db.StoragePoolVO; import org.apache.cloudstack.storage.datastore.util.LinstorUtil; import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper; -import org.apache.logging.log4j.Logger; -import org.apache.logging.log4j.LogManager; public class LinstorPrimaryDataStoreLifeCycleImpl extends BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle { - protected Logger logger = LogManager.getLogger(getClass()); - @Inject private ClusterDao clusterDao; @Inject diff --git a/plugins/storage/volume/nexenta/src/main/java/org/apache/cloudstack/storage/datastore/lifecylce/NexentaPrimaryDataStoreLifeCycle.java b/plugins/storage/volume/nexenta/src/main/java/org/apache/cloudstack/storage/datastore/lifecylce/NexentaPrimaryDataStoreLifeCycle.java index 57cd4de1b97..62995196cac 100644 --- a/plugins/storage/volume/nexenta/src/main/java/org/apache/cloudstack/storage/datastore/lifecylce/NexentaPrimaryDataStoreLifeCycle.java +++ b/plugins/storage/volume/nexenta/src/main/java/org/apache/cloudstack/storage/datastore/lifecylce/NexentaPrimaryDataStoreLifeCycle.java @@ -33,8 +33,6 @@ import org.apache.cloudstack.engine.subsystem.api.storage.ZoneScope; import org.apache.cloudstack.storage.datastore.lifecycle.BasePrimaryDataStoreLifeCycleImpl; import org.apache.cloudstack.storage.datastore.util.NexentaUtil; import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper; -import org.apache.logging.log4j.Logger; -import org.apache.logging.log4j.LogManager; import com.cloud.agent.api.StoragePoolInfo; import com.cloud.dc.DataCenterVO; @@ -49,7 +47,6 @@ import com.cloud.storage.StoragePoolAutomation; public class NexentaPrimaryDataStoreLifeCycle extends BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle { - protected Logger logger = LogManager.getLogger(getClass()); @Inject private DataCenterDao zoneDao; diff --git a/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ScaleIOPrimaryDataStoreLifeCycle.java b/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ScaleIOPrimaryDataStoreLifeCycle.java index d37d0f17225..7bbe0331c07 100644 --- a/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ScaleIOPrimaryDataStoreLifeCycle.java +++ b/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ScaleIOPrimaryDataStoreLifeCycle.java @@ -47,8 +47,6 @@ import org.apache.cloudstack.storage.datastore.client.ScaleIOGatewayClient; import org.apache.cloudstack.storage.datastore.db.PrimaryDataStoreDao; import org.apache.cloudstack.storage.datastore.db.StoragePoolVO; import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper; -import org.apache.logging.log4j.Logger; -import org.apache.logging.log4j.LogManager; import com.cloud.agent.AgentManager; import com.cloud.agent.api.Answer; @@ -76,8 +74,6 @@ import com.cloud.utils.crypt.DBEncryptionUtil; import com.cloud.utils.exception.CloudRuntimeException; public class ScaleIOPrimaryDataStoreLifeCycle extends BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle { - protected Logger logger = LogManager.getLogger(getClass()); - @Inject private ClusterDao clusterDao; @Inject diff --git a/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/manager/ScaleIOSDCManagerImpl.java b/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/manager/ScaleIOSDCManagerImpl.java index b121a1da66f..92fe1d83761 100644 --- a/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/manager/ScaleIOSDCManagerImpl.java +++ b/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/manager/ScaleIOSDCManagerImpl.java @@ -30,7 +30,8 @@ import org.apache.cloudstack.storage.datastore.client.ScaleIOGatewayClientConnec import org.apache.cloudstack.storage.datastore.db.StoragePoolDetailsDao; import org.apache.commons.collections.MapUtils; import org.apache.commons.lang3.StringUtils; -import org.apache.log4j.Logger; +import org.apache.logging.log4j.LogManager; +import org.apache.logging.log4j.Logger; import org.springframework.stereotype.Component; import com.cloud.agent.AgentManager; @@ -51,7 +52,7 @@ import com.cloud.utils.exception.CloudRuntimeException; @Component public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { - private static final Logger LOGGER = Logger.getLogger(ScaleIOSDCManagerImpl.class); + private Logger logger = LogManager.getLogger(getClass()); @Inject AgentManager agentManager; @@ -79,14 +80,14 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { int connectedSdcsCount = getScaleIOClient(storagePoolId).getConnectedSdcsCount(); if (connectedSdcsCount < connectedClientsLimit) { - LOGGER.debug(String.format("Current connected SDCs count: %d - SDC connections are within the limit (%d) on PowerFlex Storage with pool id: %d", connectedSdcsCount, connectedClientsLimit, storagePoolId)); + logger.debug(String.format("Current connected SDCs count: %d - SDC connections are within the limit (%d) on PowerFlex Storage with pool id: %d", connectedSdcsCount, connectedClientsLimit, storagePoolId)); return true; } - LOGGER.debug(String.format("Current connected SDCs count: %d - SDC connections limit (%d) reached on PowerFlex Storage with pool id: %d", connectedSdcsCount, connectedClientsLimit, storagePoolId)); + logger.debug(String.format("Current connected SDCs count: %d - SDC connections limit (%d) reached on PowerFlex Storage with pool id: %d", connectedSdcsCount, connectedClientsLimit, storagePoolId)); return false; } catch (Exception e) { String errMsg = "Unable to check SDC connections for the PowerFlex storage pool with id: " + storagePoolId + " due to " + e.getMessage(); - LOGGER.warn(errMsg, e); + logger.warn(errMsg, e); return false; } } @@ -109,7 +110,7 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { int storagePoolMaxWaitSeconds = NumbersUtil.parseInt(configDao.getValue(Config.StoragePoolMaxWaitSeconds.key()), 3600); if (!hostIdStorageSystemIdLock.lock(storagePoolMaxWaitSeconds)) { - LOGGER.debug("Unable to prepare SDC, couldn't lock on " + hostIdStorageSystemIdLockString); + logger.debug("Unable to prepare SDC, couldn't lock on " + hostIdStorageSystemIdLockString); throw new CloudRuntimeException("Unable to prepare SDC, couldn't lock on " + hostIdStorageSystemIdLockString); } @@ -117,25 +118,25 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { long hostId = host.getId(); String sdcId = getConnectedSdc(poolId, hostId); if (StringUtils.isNotBlank(sdcId)) { - LOGGER.debug(String.format("SDC %s already connected for the pool: %d on host: %d, no need to prepare/start it", sdcId, poolId, hostId)); + logger.debug(String.format("SDC %s already connected for the pool: %d on host: %d, no need to prepare/start it", sdcId, poolId, hostId)); return sdcId; } String storageSystemIdLockString = String.format(POWERFLEX_SDC_SYSTEMID_LOCK_FORMAT, systemId); storageSystemIdLock = GlobalLock.getInternLock(storageSystemIdLockString); if (storageSystemIdLock == null) { - LOGGER.error("Unable to prepare SDC, couldn't get global lock on: " + storageSystemIdLockString); + logger.error("Unable to prepare SDC, couldn't get global lock on: " + storageSystemIdLockString); throw new CloudRuntimeException("Unable to prepare SDC, couldn't get global lock on " + storageSystemIdLockString); } if (!storageSystemIdLock.lock(storagePoolMaxWaitSeconds)) { - LOGGER.error("Unable to prepare SDC, couldn't lock on " + storageSystemIdLockString); + logger.error("Unable to prepare SDC, couldn't lock on " + storageSystemIdLockString); throw new CloudRuntimeException("Unable to prepare SDC, couldn't lock on " + storageSystemIdLockString); } if (!areSDCConnectionsWithinLimit(poolId)) { String errorMsg = String.format("Unable to check SDC connections or the connections limit reached for Powerflex storage (System ID: %s)", systemId); - LOGGER.error(errorMsg); + logger.error(errorMsg); throw new CloudRuntimeException(errorMsg); } @@ -174,7 +175,7 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { } private String prepareSDCOnHost(Host host, DataStore dataStore, String systemId) { - LOGGER.debug(String.format("Preparing SDC on the host %s (%s)", host.getId(), host.getName())); + logger.debug(String.format("Preparing SDC on the host %s (%s)", host.getId(), host.getName())); Map<String,String> details = new HashMap<>(); details.put(ScaleIOGatewayClient.STORAGE_POOL_SYSTEM_ID, systemId); PrepareStorageClientCommand cmd = new PrepareStorageClientCommand(((PrimaryDataStore) dataStore).getPoolType(), dataStore.getUuid(), details); @@ -186,25 +187,25 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { prepareStorageClientAnswer = (PrepareStorageClientAnswer) agentManager.send(host.getId(), cmd); } catch (AgentUnavailableException | OperationTimedoutException e) { String err = String.format("Failed to prepare SDC on the host %s, due to: %s", host.getName(), e.getMessage()); - LOGGER.error(err); + logger.error(err); throw new CloudRuntimeException(err); } if (prepareStorageClientAnswer == null) { String err = String.format("Unable to prepare SDC on the host %s", host.getName()); - LOGGER.error(err); + logger.error(err); throw new CloudRuntimeException(err); } if (!prepareStorageClientAnswer.getResult()) { String err = String.format("Unable to prepare SDC on the host %s, due to: %s", host.getName(), prepareStorageClientAnswer.getDetails()); - LOGGER.error(err); + logger.error(err); throw new CloudRuntimeException(err); } Map<String,String> poolDetails = prepareStorageClientAnswer.getDetailsMap(); if (MapUtils.isEmpty(poolDetails)) { - LOGGER.warn(String.format("PowerFlex storage SDC details not found on the host: %s, try (re)install SDC and restart agent", host.getId())); + logger.warn(String.format("PowerFlex storage SDC details not found on the host: %s, try (re)install SDC and restart agent", host.getId())); return null; } @@ -217,7 +218,7 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { } if (StringUtils.isBlank(sdcId)) { - LOGGER.warn(String.format("Couldn't retrieve PowerFlex storage SDC details from the host: %s, try (re)install SDC and restart agent", host.getId())); + logger.warn(String.format("Couldn't retrieve PowerFlex storage SDC details from the host: %s, try (re)install SDC and restart agent", host.getId())); return null; } @@ -241,7 +242,7 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { int storagePoolMaxWaitSeconds = NumbersUtil.parseInt(configDao.getValue(Config.StoragePoolMaxWaitSeconds.key()), 3600); if (!lock.lock(storagePoolMaxWaitSeconds)) { - LOGGER.debug("Unable to unprepare SDC, couldn't lock on " + hostIdStorageSystemIdLockString); + logger.debug("Unable to unprepare SDC, couldn't lock on " + hostIdStorageSystemIdLockString); throw new CloudRuntimeException("Unable to unprepare SDC, couldn't lock on " + hostIdStorageSystemIdLockString); } @@ -249,7 +250,7 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { long hostId = host.getId(); String sdcId = getConnectedSdc(poolId, hostId); if (StringUtils.isBlank(sdcId)) { - LOGGER.debug("SDC not connected, no need to unprepare it"); + logger.debug("SDC not connected, no need to unprepare it"); return true; } @@ -263,7 +264,7 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { } private boolean unprepareSDCOnHost(Host host, DataStore dataStore) { - LOGGER.debug(String.format("Unpreparing SDC on the host %s (%s)", host.getId(), host.getName())); + logger.debug(String.format("Unpreparing SDC on the host %s (%s)", host.getId(), host.getName())); UnprepareStorageClientCommand cmd = new UnprepareStorageClientCommand(((PrimaryDataStore) dataStore).getPoolType(), dataStore.getUuid()); int timeoutSeconds = 60; cmd.setWait(timeoutSeconds); @@ -273,13 +274,13 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { unprepareStorageClientAnswer = agentManager.send(host.getId(), cmd); } catch (AgentUnavailableException | OperationTimedoutException e) { String err = String.format("Failed to unprepare SDC on the host %s due to: %s", host.getName(), e.getMessage()); - LOGGER.error(err); + logger.error(err); return false; } if (!unprepareStorageClientAnswer.getResult()) { String err = String.format("Unable to unprepare SDC on the the host %s due to: %s", host.getName(), unprepareStorageClientAnswer.getDetails()); - LOGGER.error(err); + logger.error(err); return false; } return true; @@ -287,11 +288,11 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { private String getHostSdcId(String sdcGuid, long poolId) { try { - LOGGER.debug(String.format("Try to get host SDC Id for pool: %s, with SDC guid %s", poolId, sdcGuid)); + logger.debug(String.format("Try to get host SDC Id for pool: %s, with SDC guid %s", poolId, sdcGuid)); ScaleIOGatewayClient client = getScaleIOClient(poolId); return client.getSdcIdByGuid(sdcGuid); } catch (Exception e) { - LOGGER.error(String.format("Failed to get host SDC Id for pool: %s", poolId), e); + logger.error(String.format("Failed to get host SDC Id for pool: %s", poolId), e); throw new CloudRuntimeException(String.format("Failed to establish connection with PowerFlex Gateway to get host SDC Id for pool: %s", poolId)); } } @@ -308,14 +309,14 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { return poolHostVO.getLocalPath(); } } catch (Exception e) { - LOGGER.warn("Unable to get connected SDC for the host: " + hostId + " and storage pool: " + poolId + " due to " + e.getMessage(), e); + logger.warn("Unable to get connected SDC for the host: " + hostId + " and storage pool: " + poolId + " due to " + e.getMessage(), e); } return null; } private boolean hostSdcConnected(String sdcId, long poolId, int waitTimeInSecs) { - LOGGER.debug(String.format("Waiting (for %d secs) for the SDC %s of the pool id: %d to connect", waitTimeInSecs, sdcId, poolId)); + logger.debug(String.format("Waiting (for %d secs) for the SDC %s of the pool id: %d to connect", waitTimeInSecs, sdcId, poolId)); int timeBetweenTries = 1000; // Try more frequently (every sec) and return early if connected while (waitTimeInSecs > 0) { if (isHostSdcConnected(sdcId, poolId)) { @@ -335,7 +336,7 @@ public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager { final ScaleIOGatewayClient client = getScaleIOClient(poolId); return client.isSdcConnected(sdcId); } catch (Exception e) { - LOGGER.error("Failed to check host SDC connection", e); + logger.error("Failed to check host SDC connection", e); throw new CloudRuntimeException("Failed to establish connection with PowerFlex Gateway to check host SDC connection"); } } diff --git a/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFirePrimaryDataStoreLifeCycle.java b/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFirePrimaryDataStoreLifeCycle.java index 5080a413ac4..0b5393bd8d8 100644 --- a/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFirePrimaryDataStoreLifeCycle.java +++ b/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFirePrimaryDataStoreLifeCycle.java @@ -25,9 +25,6 @@ import java.util.UUID; import javax.inject.Inject; -import org.apache.logging.log4j.Logger; -import org.apache.logging.log4j.LogManager; - import org.apache.cloudstack.engine.subsystem.api.storage.ClusterScope; import org.apache.cloudstack.engine.subsystem.api.storage.DataStore; import org.apache.cloudstack.engine.subsystem.api.storage.HostScope; @@ -65,8 +62,6 @@ import com.cloud.utils.exception.CloudRuntimeException; import com.google.common.base.Preconditions; public class SolidFirePrimaryDataStoreLifeCycle extends BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle { - protected Logger logger = LogManager.getLogger(getClass()); - @Inject private CapacityManager _capacityMgr; @Inject private ClusterDao _clusterDao; @Inject private DataCenterDao _zoneDao; diff --git a/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFireSharedPrimaryDataStoreLifeCycle.java b/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFireSharedPrimaryDataStoreLifeCycle.java index 9742bfb21c2..4877e86bf9f 100644 --- a/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFireSharedPrimaryDataStoreLifeCycle.java +++ b/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFireSharedPrimaryDataStoreLifeCycle.java @@ -26,9 +26,6 @@ import java.util.Map; import javax.inject.Inject; -import org.apache.logging.log4j.Logger; -import org.apache.logging.log4j.LogManager; - import org.apache.cloudstack.context.CallContext; import org.apache.cloudstack.engine.subsystem.api.storage.ClusterScope; import org.apache.cloudstack.engine.subsystem.api.storage.DataStore; @@ -74,8 +71,6 @@ import com.cloud.utils.db.GlobalLock; import com.cloud.utils.exception.CloudRuntimeException; public class SolidFireSharedPrimaryDataStoreLifeCycle extends BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle { - protected Logger logger = LogManager.getLogger(getClass()); - @Inject private AccountDao accountDao; @Inject private AccountDetailsDao accountDetailsDao; @Inject private AgentManager agentMgr; diff --git a/plugins/storage/volume/storpool/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/StorPoolPrimaryDataStoreLifeCycle.java b/plugins/storage/volume/storpool/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/StorPoolPrimaryDataStoreLifeCycle.java index e13b7f7b706..4f2fdef1723 100644 --- a/plugins/storage/volume/storpool/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/StorPoolPrimaryDataStoreLifeCycle.java +++ b/plugins/storage/volume/storpool/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/StorPoolPrimaryDataStoreLifeCycle.java @@ -38,8 +38,6 @@ import org.apache.cloudstack.storage.datastore.util.StorPoolUtil; import org.apache.cloudstack.storage.datastore.util.StorPoolUtil.SpApiResponse; import org.apache.cloudstack.storage.datastore.util.StorPoolUtil.SpConnectionDesc; import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper; -import org.apache.logging.log4j.Logger; -import org.apache.logging.log4j.LogManager; import com.cloud.agent.api.StoragePoolInfo; import com.cloud.host.HostVO; @@ -62,8 +60,6 @@ import com.cloud.storage.dao.VMTemplatePoolDao; import com.cloud.utils.exception.CloudRuntimeException; public class StorPoolPrimaryDataStoreLifeCycle extends BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle { - protected Logger logger = LogManager.getLogger(getClass()); - @Inject protected PrimaryDataStoreHelper dataStoreHelper; @Inject