This is an automated email from the ASF dual-hosted git repository.
sureshanaparti pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/cloudstack.git
The following commit(s) were added to refs/heads/main by this push:
new 7c32bd2506e Fixup main build errors (#9330)
7c32bd2506e is described below
commit 7c32bd2506eb14348f2be22d6773d15d646ee809
Author: Vishesh <[email protected]>
AuthorDate: Thu Jul 4 13:00:37 2024 +0530
Fixup main build errors (#9330)
* Fixup main build errors
* Fixup flaky test
* Address comments
---
.../BasePrimaryDataStoreLifeCycleImpl.java | 14 +++---
.../api/response/QuotaResponseBuilderImplTest.java | 2 +-
.../LibvirtCheckConvertInstanceCommandWrapper.java | 8 +---
.../LibvirtPrepareStorageClientCommandWrapper.java | 6 +--
...ibvirtUnprepareStorageClientCommandWrapper.java | 6 +--
.../lifecycle/AdaptiveDataStoreLifeCycleImpl.java | 4 --
.../ElastistorPrimaryDataStoreLifeCycle.java | 5 --
.../lifecycle/DateraPrimaryDataStoreLifeCycle.java | 4 --
.../CloudStackPrimaryDataStoreLifeCycleImpl.java | 3 --
.../LinstorPrimaryDataStoreLifeCycleImpl.java | 4 --
.../NexentaPrimaryDataStoreLifeCycle.java | 3 --
.../ScaleIOPrimaryDataStoreLifeCycle.java | 4 --
.../datastore/manager/ScaleIOSDCManagerImpl.java | 53 +++++++++++-----------
.../SolidFirePrimaryDataStoreLifeCycle.java | 5 --
.../SolidFireSharedPrimaryDataStoreLifeCycle.java | 5 --
.../StorPoolPrimaryDataStoreLifeCycle.java | 4 --
16 files changed, 40 insertions(+), 90 deletions(-)
diff --git
a/engine/storage/volume/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/BasePrimaryDataStoreLifeCycleImpl.java
b/engine/storage/volume/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/BasePrimaryDataStoreLifeCycleImpl.java
index adc74a77d43..1ee4d40a567 100644
---
a/engine/storage/volume/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/BasePrimaryDataStoreLifeCycleImpl.java
+++
b/engine/storage/volume/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/BasePrimaryDataStoreLifeCycleImpl.java
@@ -25,7 +25,6 @@ import javax.inject.Inject;
import org.apache.cloudstack.engine.subsystem.api.storage.ClusterScope;
import org.apache.cloudstack.engine.subsystem.api.storage.DataStore;
import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper;
-import org.apache.log4j.Logger;
import com.cloud.agent.AgentManager;
import com.cloud.agent.api.Answer;
@@ -39,9 +38,12 @@ import com.cloud.storage.StoragePool;
import com.cloud.storage.StoragePoolHostVO;
import com.cloud.storage.dao.StoragePoolHostDao;
import com.cloud.utils.Pair;
+import org.apache.logging.log4j.LogManager;
+import org.apache.logging.log4j.Logger;
public class BasePrimaryDataStoreLifeCycleImpl {
- private static final Logger s_logger =
Logger.getLogger(BasePrimaryDataStoreLifeCycleImpl.class);
+ protected Logger logger = LogManager.getLogger(getClass());
+
@Inject
AgentManager agentMgr;
@Inject
@@ -70,13 +72,13 @@ public class BasePrimaryDataStoreLifeCycleImpl {
public void changeStoragePoolScopeToZone(DataStore store, ClusterScope
clusterScope, HypervisorType hypervisorType) {
List<HostVO> hosts = getPoolHostsList(clusterScope, hypervisorType);
- s_logger.debug("Changing scope of the storage pool to Zone");
+ logger.debug("Changing scope of the storage pool to Zone");
if (hosts != null) {
for (HostVO host : hosts) {
try {
storageMgr.connectHostToSharedPool(host.getId(),
store.getId());
} catch (Exception e) {
- s_logger.warn("Unable to establish a connection between "
+ host + " and " + store, e);
+ logger.warn("Unable to establish a connection between " +
host + " and " + store, e);
}
}
}
@@ -85,7 +87,7 @@ public class BasePrimaryDataStoreLifeCycleImpl {
public void changeStoragePoolScopeToCluster(DataStore store, ClusterScope
clusterScope, HypervisorType hypervisorType) {
Pair<List<StoragePoolHostVO>, Integer> hostPoolRecords =
storagePoolHostDao.listByPoolIdNotInCluster(clusterScope.getScopeId(),
store.getId());
- s_logger.debug("Changing scope of the storage pool to Cluster");
+ logger.debug("Changing scope of the storage pool to Cluster");
if (hostPoolRecords.second() > 0) {
StoragePool pool = (StoragePool) store;
for (StoragePoolHostVO host : hostPoolRecords.first()) {
@@ -94,7 +96,7 @@ public class BasePrimaryDataStoreLifeCycleImpl {
if (answer != null) {
if (!answer.getResult()) {
- s_logger.debug("Failed to delete storage pool: " +
answer.getResult());
+ logger.debug("Failed to delete storage pool: " +
answer.getResult());
} else if (HypervisorType.KVM != hypervisorType) {
break;
}
diff --git
a/plugins/database/quota/src/test/java/org/apache/cloudstack/api/response/QuotaResponseBuilderImplTest.java
b/plugins/database/quota/src/test/java/org/apache/cloudstack/api/response/QuotaResponseBuilderImplTest.java
index 6bca4ea85bb..da02b6d3709 100644
---
a/plugins/database/quota/src/test/java/org/apache/cloudstack/api/response/QuotaResponseBuilderImplTest.java
+++
b/plugins/database/quota/src/test/java/org/apache/cloudstack/api/response/QuotaResponseBuilderImplTest.java
@@ -334,7 +334,7 @@ public class QuotaResponseBuilderImplTest extends TestCase {
@Test
public void validateEndDateOnCreatingNewQuotaTariffTestSetValidEndDate() {
Date startDate = DateUtils.addDays(date, -100);
- Date endDate = DateUtils.addMilliseconds(new Date(), 1);
+ Date endDate = DateUtils.addMinutes(new Date(), 1);
quotaResponseBuilderSpy.validateEndDateOnCreatingNewQuotaTariff(quotaTariffVoMock,
startDate, endDate);
Mockito.verify(quotaTariffVoMock).setEndDate(Mockito.any(Date.class));
diff --git
a/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtCheckConvertInstanceCommandWrapper.java
b/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtCheckConvertInstanceCommandWrapper.java
index d94fddeeb44..d3ebb28b106 100644
---
a/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtCheckConvertInstanceCommandWrapper.java
+++
b/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtCheckConvertInstanceCommandWrapper.java
@@ -18,8 +18,6 @@
//
package com.cloud.hypervisor.kvm.resource.wrapper;
-import org.apache.log4j.Logger;
-
import com.cloud.agent.api.Answer;
import com.cloud.agent.api.CheckConvertInstanceAnswer;
import com.cloud.agent.api.CheckConvertInstanceCommand;
@@ -30,21 +28,19 @@ import com.cloud.resource.ResourceWrapper;
@ResourceWrapper(handles = CheckConvertInstanceCommand.class)
public class LibvirtCheckConvertInstanceCommandWrapper extends
CommandWrapper<CheckConvertInstanceCommand, Answer, LibvirtComputingResource> {
- private static final Logger s_logger =
Logger.getLogger(LibvirtCheckConvertInstanceCommandWrapper.class);
-
@Override
public Answer execute(CheckConvertInstanceCommand cmd,
LibvirtComputingResource serverResource) {
if (!serverResource.hostSupportsInstanceConversion()) {
String msg = String.format("Cannot convert the instance from
VMware as the virt-v2v binary is not found on host %s. " +
"Please install virt-v2v%s on the host before attempting
the instance conversion.", serverResource.getPrivateIp(),
serverResource.isUbuntuHost()? ", nbdkit" : "");
- s_logger.info(msg);
+ logger.info(msg);
return new CheckConvertInstanceAnswer(cmd, false, msg);
}
if (cmd.getCheckWindowsGuestConversionSupport() &&
!serverResource.hostSupportsWindowsGuestConversion()) {
String msg = String.format("Cannot convert the instance from
VMware as the virtio-win package is not found on host %s. " +
"Please install virtio-win package on the host before
attempting the windows guest instance conversion.",
serverResource.getPrivateIp());
- s_logger.info(msg);
+ logger.info(msg);
return new CheckConvertInstanceAnswer(cmd, false, msg);
}
diff --git
a/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtPrepareStorageClientCommandWrapper.java
b/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtPrepareStorageClientCommandWrapper.java
index 79afd4696b0..b5cab17ecb1 100644
---
a/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtPrepareStorageClientCommandWrapper.java
+++
b/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtPrepareStorageClientCommandWrapper.java
@@ -21,8 +21,6 @@ package com.cloud.hypervisor.kvm.resource.wrapper;
import java.util.Map;
-import org.apache.log4j.Logger;
-
import com.cloud.agent.api.Answer;
import com.cloud.agent.api.PrepareStorageClientAnswer;
import com.cloud.agent.api.PrepareStorageClientCommand;
@@ -35,15 +33,13 @@ import com.cloud.utils.Ternary;
@ResourceWrapper(handles = PrepareStorageClientCommand.class)
public class LibvirtPrepareStorageClientCommandWrapper extends
CommandWrapper<PrepareStorageClientCommand, Answer, LibvirtComputingResource> {
- private static final Logger s_logger =
Logger.getLogger(LibvirtPrepareStorageClientCommandWrapper.class);
-
@Override
public Answer execute(PrepareStorageClientCommand cmd,
LibvirtComputingResource libvirtComputingResource) {
final KVMStoragePoolManager storagePoolMgr =
libvirtComputingResource.getStoragePoolMgr();
Ternary<Boolean, Map<String, String>, String>
prepareStorageClientResult =
storagePoolMgr.prepareStorageClient(cmd.getPoolType(), cmd.getPoolUuid(),
cmd.getDetails());
if (!prepareStorageClientResult.first()) {
String msg = prepareStorageClientResult.third();
- s_logger.debug("Unable to prepare storage client, due to: " + msg);
+ logger.debug("Unable to prepare storage client, due to: " + msg);
return new PrepareStorageClientAnswer(cmd, false, msg);
}
Map<String, String> details = prepareStorageClientResult.second();
diff --git
a/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtUnprepareStorageClientCommandWrapper.java
b/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtUnprepareStorageClientCommandWrapper.java
index f98782fe748..2f23a934003 100644
---
a/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtUnprepareStorageClientCommandWrapper.java
+++
b/plugins/hypervisors/kvm/src/main/java/com/cloud/hypervisor/kvm/resource/wrapper/LibvirtUnprepareStorageClientCommandWrapper.java
@@ -19,8 +19,6 @@
package com.cloud.hypervisor.kvm.resource.wrapper;
-import org.apache.log4j.Logger;
-
import com.cloud.agent.api.Answer;
import com.cloud.agent.api.UnprepareStorageClientAnswer;
import com.cloud.agent.api.UnprepareStorageClientCommand;
@@ -33,15 +31,13 @@ import com.cloud.utils.Pair;
@ResourceWrapper(handles = UnprepareStorageClientCommand.class)
public class LibvirtUnprepareStorageClientCommandWrapper extends
CommandWrapper<UnprepareStorageClientCommand, Answer, LibvirtComputingResource>
{
- private static final Logger s_logger =
Logger.getLogger(LibvirtUnprepareStorageClientCommandWrapper.class);
-
@Override
public Answer execute(UnprepareStorageClientCommand cmd,
LibvirtComputingResource libvirtComputingResource) {
final KVMStoragePoolManager storagePoolMgr =
libvirtComputingResource.getStoragePoolMgr();
Pair<Boolean, String> unprepareStorageClientResult =
storagePoolMgr.unprepareStorageClient(cmd.getPoolType(), cmd.getPoolUuid());
if (!unprepareStorageClientResult.first()) {
String msg = unprepareStorageClientResult.second();
- s_logger.debug("Couldn't unprepare storage client, due to: " +
msg);
+ logger.debug("Couldn't unprepare storage client, due to: " + msg);
return new UnprepareStorageClientAnswer(cmd, false, msg);
}
return new UnprepareStorageClientAnswer(cmd, true);
diff --git
a/plugins/storage/volume/adaptive/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/AdaptiveDataStoreLifeCycleImpl.java
b/plugins/storage/volume/adaptive/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/AdaptiveDataStoreLifeCycleImpl.java
index 86c3bfa91de..f6ace68e2e8 100644
---
a/plugins/storage/volume/adaptive/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/AdaptiveDataStoreLifeCycleImpl.java
+++
b/plugins/storage/volume/adaptive/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/AdaptiveDataStoreLifeCycleImpl.java
@@ -57,8 +57,6 @@ import com.cloud.storage.Storage.StoragePoolType;
import com.cloud.utils.crypt.DBEncryptionUtil;
import com.cloud.utils.exception.CloudRuntimeException;
import com.cloud.host.Host;
-import org.apache.logging.log4j.LogManager;
-import org.apache.logging.log4j.Logger;
/**
* Manages the lifecycle of a Managed Data Store in CloudStack
@@ -66,8 +64,6 @@ import org.apache.logging.log4j.Logger;
public class AdaptiveDataStoreLifeCycleImpl extends
BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle {
@Inject
private PrimaryDataStoreDao _storagePoolDao;
- protected Logger logger = LogManager.getLogger(getClass());
-
@Inject
PrimaryDataStoreHelper _dataStoreHelper;
@Inject
diff --git
a/plugins/storage/volume/cloudbyte/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ElastistorPrimaryDataStoreLifeCycle.java
b/plugins/storage/volume/cloudbyte/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ElastistorPrimaryDataStoreLifeCycle.java
index 01058d717ed..5d7f2d86e14 100644
---
a/plugins/storage/volume/cloudbyte/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ElastistorPrimaryDataStoreLifeCycle.java
+++
b/plugins/storage/volume/cloudbyte/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ElastistorPrimaryDataStoreLifeCycle.java
@@ -26,9 +26,6 @@ import java.util.StringTokenizer;
import javax.inject.Inject;
-import org.apache.logging.log4j.Logger;
-import org.apache.logging.log4j.LogManager;
-
import org.apache.cloudstack.engine.subsystem.api.storage.ClusterScope;
import org.apache.cloudstack.engine.subsystem.api.storage.DataStore;
import org.apache.cloudstack.engine.subsystem.api.storage.HostScope;
@@ -67,8 +64,6 @@ import com.cloud.storage.dao.StoragePoolHostDao;
import com.cloud.utils.exception.CloudRuntimeException;
public class ElastistorPrimaryDataStoreLifeCycle extends
BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle {
- protected Logger logger = LogManager.getLogger(getClass());
-
@Inject
HostDao _hostDao;
@Inject
diff --git
a/plugins/storage/volume/datera/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/DateraPrimaryDataStoreLifeCycle.java
b/plugins/storage/volume/datera/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/DateraPrimaryDataStoreLifeCycle.java
index 3944cdf7bbe..6d222b36b85 100644
---
a/plugins/storage/volume/datera/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/DateraPrimaryDataStoreLifeCycle.java
+++
b/plugins/storage/volume/datera/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/DateraPrimaryDataStoreLifeCycle.java
@@ -51,8 +51,6 @@ import
org.apache.cloudstack.storage.datastore.db.PrimaryDataStoreDao;
import org.apache.cloudstack.storage.datastore.db.StoragePoolVO;
import org.apache.cloudstack.storage.datastore.util.DateraUtil;
import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper;
-import org.apache.logging.log4j.Logger;
-import org.apache.logging.log4j.LogManager;
import javax.inject.Inject;
import java.util.ArrayList;
@@ -60,8 +58,6 @@ import java.util.List;
import java.util.Map;
public class DateraPrimaryDataStoreLifeCycle extends
BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle {
- protected Logger logger = LogManager.getLogger(getClass());
-
@Inject
private CapacityManager _capacityMgr;
@Inject
diff --git
a/plugins/storage/volume/default/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/CloudStackPrimaryDataStoreLifeCycleImpl.java
b/plugins/storage/volume/default/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/CloudStackPrimaryDataStoreLifeCycleImpl.java
index 9600f0618ac..bc66e2ff136 100644
---
a/plugins/storage/volume/default/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/CloudStackPrimaryDataStoreLifeCycleImpl.java
+++
b/plugins/storage/volume/default/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/CloudStackPrimaryDataStoreLifeCycleImpl.java
@@ -64,8 +64,6 @@ import
org.apache.cloudstack.engine.subsystem.api.storage.ZoneScope;
import org.apache.cloudstack.storage.datastore.db.PrimaryDataStoreDao;
import org.apache.cloudstack.storage.datastore.db.StoragePoolVO;
import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper;
-import org.apache.logging.log4j.Logger;
-import org.apache.logging.log4j.LogManager;
import javax.inject.Inject;
import java.util.ArrayList;
@@ -74,7 +72,6 @@ import java.util.Map;
import java.util.UUID;
public class CloudStackPrimaryDataStoreLifeCycleImpl extends
BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle {
- protected Logger logger = LogManager.getLogger(getClass());
@Inject
protected ResourceManager _resourceMgr;
@Inject
diff --git
a/plugins/storage/volume/linstor/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/LinstorPrimaryDataStoreLifeCycleImpl.java
b/plugins/storage/volume/linstor/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/LinstorPrimaryDataStoreLifeCycleImpl.java
index 7b3f8dbb37e..ac2563cd436 100644
---
a/plugins/storage/volume/linstor/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/LinstorPrimaryDataStoreLifeCycleImpl.java
+++
b/plugins/storage/volume/linstor/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/LinstorPrimaryDataStoreLifeCycleImpl.java
@@ -51,12 +51,8 @@ import
org.apache.cloudstack.storage.datastore.db.PrimaryDataStoreDao;
import org.apache.cloudstack.storage.datastore.db.StoragePoolVO;
import org.apache.cloudstack.storage.datastore.util.LinstorUtil;
import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper;
-import org.apache.logging.log4j.Logger;
-import org.apache.logging.log4j.LogManager;
public class LinstorPrimaryDataStoreLifeCycleImpl extends
BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle {
- protected Logger logger = LogManager.getLogger(getClass());
-
@Inject
private ClusterDao clusterDao;
@Inject
diff --git
a/plugins/storage/volume/nexenta/src/main/java/org/apache/cloudstack/storage/datastore/lifecylce/NexentaPrimaryDataStoreLifeCycle.java
b/plugins/storage/volume/nexenta/src/main/java/org/apache/cloudstack/storage/datastore/lifecylce/NexentaPrimaryDataStoreLifeCycle.java
index 57cd4de1b97..62995196cac 100644
---
a/plugins/storage/volume/nexenta/src/main/java/org/apache/cloudstack/storage/datastore/lifecylce/NexentaPrimaryDataStoreLifeCycle.java
+++
b/plugins/storage/volume/nexenta/src/main/java/org/apache/cloudstack/storage/datastore/lifecylce/NexentaPrimaryDataStoreLifeCycle.java
@@ -33,8 +33,6 @@ import
org.apache.cloudstack.engine.subsystem.api.storage.ZoneScope;
import
org.apache.cloudstack.storage.datastore.lifecycle.BasePrimaryDataStoreLifeCycleImpl;
import org.apache.cloudstack.storage.datastore.util.NexentaUtil;
import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper;
-import org.apache.logging.log4j.Logger;
-import org.apache.logging.log4j.LogManager;
import com.cloud.agent.api.StoragePoolInfo;
import com.cloud.dc.DataCenterVO;
@@ -49,7 +47,6 @@ import com.cloud.storage.StoragePoolAutomation;
public class NexentaPrimaryDataStoreLifeCycle
extends BasePrimaryDataStoreLifeCycleImpl
implements PrimaryDataStoreLifeCycle {
- protected Logger logger = LogManager.getLogger(getClass());
@Inject
private DataCenterDao zoneDao;
diff --git
a/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ScaleIOPrimaryDataStoreLifeCycle.java
b/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ScaleIOPrimaryDataStoreLifeCycle.java
index d37d0f17225..7bbe0331c07 100644
---
a/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ScaleIOPrimaryDataStoreLifeCycle.java
+++
b/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/ScaleIOPrimaryDataStoreLifeCycle.java
@@ -47,8 +47,6 @@ import
org.apache.cloudstack.storage.datastore.client.ScaleIOGatewayClient;
import org.apache.cloudstack.storage.datastore.db.PrimaryDataStoreDao;
import org.apache.cloudstack.storage.datastore.db.StoragePoolVO;
import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper;
-import org.apache.logging.log4j.Logger;
-import org.apache.logging.log4j.LogManager;
import com.cloud.agent.AgentManager;
import com.cloud.agent.api.Answer;
@@ -76,8 +74,6 @@ import com.cloud.utils.crypt.DBEncryptionUtil;
import com.cloud.utils.exception.CloudRuntimeException;
public class ScaleIOPrimaryDataStoreLifeCycle extends
BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle {
- protected Logger logger = LogManager.getLogger(getClass());
-
@Inject
private ClusterDao clusterDao;
@Inject
diff --git
a/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/manager/ScaleIOSDCManagerImpl.java
b/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/manager/ScaleIOSDCManagerImpl.java
index b121a1da66f..92fe1d83761 100644
---
a/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/manager/ScaleIOSDCManagerImpl.java
+++
b/plugins/storage/volume/scaleio/src/main/java/org/apache/cloudstack/storage/datastore/manager/ScaleIOSDCManagerImpl.java
@@ -30,7 +30,8 @@ import
org.apache.cloudstack.storage.datastore.client.ScaleIOGatewayClientConnec
import org.apache.cloudstack.storage.datastore.db.StoragePoolDetailsDao;
import org.apache.commons.collections.MapUtils;
import org.apache.commons.lang3.StringUtils;
-import org.apache.log4j.Logger;
+import org.apache.logging.log4j.LogManager;
+import org.apache.logging.log4j.Logger;
import org.springframework.stereotype.Component;
import com.cloud.agent.AgentManager;
@@ -51,7 +52,7 @@ import com.cloud.utils.exception.CloudRuntimeException;
@Component
public class ScaleIOSDCManagerImpl implements ScaleIOSDCManager {
- private static final Logger LOGGER =
Logger.getLogger(ScaleIOSDCManagerImpl.class);
+ private Logger logger = LogManager.getLogger(getClass());
@Inject
AgentManager agentManager;
@@ -79,14 +80,14 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
int connectedSdcsCount =
getScaleIOClient(storagePoolId).getConnectedSdcsCount();
if (connectedSdcsCount < connectedClientsLimit) {
- LOGGER.debug(String.format("Current connected SDCs count: %d -
SDC connections are within the limit (%d) on PowerFlex Storage with pool id:
%d", connectedSdcsCount, connectedClientsLimit, storagePoolId));
+ logger.debug(String.format("Current connected SDCs count: %d -
SDC connections are within the limit (%d) on PowerFlex Storage with pool id:
%d", connectedSdcsCount, connectedClientsLimit, storagePoolId));
return true;
}
- LOGGER.debug(String.format("Current connected SDCs count: %d - SDC
connections limit (%d) reached on PowerFlex Storage with pool id: %d",
connectedSdcsCount, connectedClientsLimit, storagePoolId));
+ logger.debug(String.format("Current connected SDCs count: %d - SDC
connections limit (%d) reached on PowerFlex Storage with pool id: %d",
connectedSdcsCount, connectedClientsLimit, storagePoolId));
return false;
} catch (Exception e) {
String errMsg = "Unable to check SDC connections for the PowerFlex
storage pool with id: " + storagePoolId + " due to " + e.getMessage();
- LOGGER.warn(errMsg, e);
+ logger.warn(errMsg, e);
return false;
}
}
@@ -109,7 +110,7 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
int storagePoolMaxWaitSeconds =
NumbersUtil.parseInt(configDao.getValue(Config.StoragePoolMaxWaitSeconds.key()),
3600);
if (!hostIdStorageSystemIdLock.lock(storagePoolMaxWaitSeconds)) {
- LOGGER.debug("Unable to prepare SDC, couldn't lock on " +
hostIdStorageSystemIdLockString);
+ logger.debug("Unable to prepare SDC, couldn't lock on " +
hostIdStorageSystemIdLockString);
throw new CloudRuntimeException("Unable to prepare SDC,
couldn't lock on " + hostIdStorageSystemIdLockString);
}
@@ -117,25 +118,25 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
long hostId = host.getId();
String sdcId = getConnectedSdc(poolId, hostId);
if (StringUtils.isNotBlank(sdcId)) {
- LOGGER.debug(String.format("SDC %s already connected for the
pool: %d on host: %d, no need to prepare/start it", sdcId, poolId, hostId));
+ logger.debug(String.format("SDC %s already connected for the
pool: %d on host: %d, no need to prepare/start it", sdcId, poolId, hostId));
return sdcId;
}
String storageSystemIdLockString =
String.format(POWERFLEX_SDC_SYSTEMID_LOCK_FORMAT, systemId);
storageSystemIdLock =
GlobalLock.getInternLock(storageSystemIdLockString);
if (storageSystemIdLock == null) {
- LOGGER.error("Unable to prepare SDC, couldn't get global lock
on: " + storageSystemIdLockString);
+ logger.error("Unable to prepare SDC, couldn't get global lock
on: " + storageSystemIdLockString);
throw new CloudRuntimeException("Unable to prepare SDC,
couldn't get global lock on " + storageSystemIdLockString);
}
if (!storageSystemIdLock.lock(storagePoolMaxWaitSeconds)) {
- LOGGER.error("Unable to prepare SDC, couldn't lock on " +
storageSystemIdLockString);
+ logger.error("Unable to prepare SDC, couldn't lock on " +
storageSystemIdLockString);
throw new CloudRuntimeException("Unable to prepare SDC,
couldn't lock on " + storageSystemIdLockString);
}
if (!areSDCConnectionsWithinLimit(poolId)) {
String errorMsg = String.format("Unable to check SDC
connections or the connections limit reached for Powerflex storage (System ID:
%s)", systemId);
- LOGGER.error(errorMsg);
+ logger.error(errorMsg);
throw new CloudRuntimeException(errorMsg);
}
@@ -174,7 +175,7 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
}
private String prepareSDCOnHost(Host host, DataStore dataStore, String
systemId) {
- LOGGER.debug(String.format("Preparing SDC on the host %s (%s)",
host.getId(), host.getName()));
+ logger.debug(String.format("Preparing SDC on the host %s (%s)",
host.getId(), host.getName()));
Map<String,String> details = new HashMap<>();
details.put(ScaleIOGatewayClient.STORAGE_POOL_SYSTEM_ID, systemId);
PrepareStorageClientCommand cmd = new
PrepareStorageClientCommand(((PrimaryDataStore) dataStore).getPoolType(),
dataStore.getUuid(), details);
@@ -186,25 +187,25 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
prepareStorageClientAnswer = (PrepareStorageClientAnswer)
agentManager.send(host.getId(), cmd);
} catch (AgentUnavailableException | OperationTimedoutException e) {
String err = String.format("Failed to prepare SDC on the host %s,
due to: %s", host.getName(), e.getMessage());
- LOGGER.error(err);
+ logger.error(err);
throw new CloudRuntimeException(err);
}
if (prepareStorageClientAnswer == null) {
String err = String.format("Unable to prepare SDC on the host %s",
host.getName());
- LOGGER.error(err);
+ logger.error(err);
throw new CloudRuntimeException(err);
}
if (!prepareStorageClientAnswer.getResult()) {
String err = String.format("Unable to prepare SDC on the host %s,
due to: %s", host.getName(), prepareStorageClientAnswer.getDetails());
- LOGGER.error(err);
+ logger.error(err);
throw new CloudRuntimeException(err);
}
Map<String,String> poolDetails =
prepareStorageClientAnswer.getDetailsMap();
if (MapUtils.isEmpty(poolDetails)) {
- LOGGER.warn(String.format("PowerFlex storage SDC details not found
on the host: %s, try (re)install SDC and restart agent", host.getId()));
+ logger.warn(String.format("PowerFlex storage SDC details not found
on the host: %s, try (re)install SDC and restart agent", host.getId()));
return null;
}
@@ -217,7 +218,7 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
}
if (StringUtils.isBlank(sdcId)) {
- LOGGER.warn(String.format("Couldn't retrieve PowerFlex storage SDC
details from the host: %s, try (re)install SDC and restart agent",
host.getId()));
+ logger.warn(String.format("Couldn't retrieve PowerFlex storage SDC
details from the host: %s, try (re)install SDC and restart agent",
host.getId()));
return null;
}
@@ -241,7 +242,7 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
int storagePoolMaxWaitSeconds =
NumbersUtil.parseInt(configDao.getValue(Config.StoragePoolMaxWaitSeconds.key()),
3600);
if (!lock.lock(storagePoolMaxWaitSeconds)) {
- LOGGER.debug("Unable to unprepare SDC, couldn't lock on " +
hostIdStorageSystemIdLockString);
+ logger.debug("Unable to unprepare SDC, couldn't lock on " +
hostIdStorageSystemIdLockString);
throw new CloudRuntimeException("Unable to unprepare SDC,
couldn't lock on " + hostIdStorageSystemIdLockString);
}
@@ -249,7 +250,7 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
long hostId = host.getId();
String sdcId = getConnectedSdc(poolId, hostId);
if (StringUtils.isBlank(sdcId)) {
- LOGGER.debug("SDC not connected, no need to unprepare it");
+ logger.debug("SDC not connected, no need to unprepare it");
return true;
}
@@ -263,7 +264,7 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
}
private boolean unprepareSDCOnHost(Host host, DataStore dataStore) {
- LOGGER.debug(String.format("Unpreparing SDC on the host %s (%s)",
host.getId(), host.getName()));
+ logger.debug(String.format("Unpreparing SDC on the host %s (%s)",
host.getId(), host.getName()));
UnprepareStorageClientCommand cmd = new
UnprepareStorageClientCommand(((PrimaryDataStore) dataStore).getPoolType(),
dataStore.getUuid());
int timeoutSeconds = 60;
cmd.setWait(timeoutSeconds);
@@ -273,13 +274,13 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
unprepareStorageClientAnswer = agentManager.send(host.getId(),
cmd);
} catch (AgentUnavailableException | OperationTimedoutException e) {
String err = String.format("Failed to unprepare SDC on the host %s
due to: %s", host.getName(), e.getMessage());
- LOGGER.error(err);
+ logger.error(err);
return false;
}
if (!unprepareStorageClientAnswer.getResult()) {
String err = String.format("Unable to unprepare SDC on the the
host %s due to: %s", host.getName(), unprepareStorageClientAnswer.getDetails());
- LOGGER.error(err);
+ logger.error(err);
return false;
}
return true;
@@ -287,11 +288,11 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
private String getHostSdcId(String sdcGuid, long poolId) {
try {
- LOGGER.debug(String.format("Try to get host SDC Id for pool: %s,
with SDC guid %s", poolId, sdcGuid));
+ logger.debug(String.format("Try to get host SDC Id for pool: %s,
with SDC guid %s", poolId, sdcGuid));
ScaleIOGatewayClient client = getScaleIOClient(poolId);
return client.getSdcIdByGuid(sdcGuid);
} catch (Exception e) {
- LOGGER.error(String.format("Failed to get host SDC Id for pool:
%s", poolId), e);
+ logger.error(String.format("Failed to get host SDC Id for pool:
%s", poolId), e);
throw new CloudRuntimeException(String.format("Failed to establish
connection with PowerFlex Gateway to get host SDC Id for pool: %s", poolId));
}
}
@@ -308,14 +309,14 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
return poolHostVO.getLocalPath();
}
} catch (Exception e) {
- LOGGER.warn("Unable to get connected SDC for the host: " + hostId
+ " and storage pool: " + poolId + " due to " + e.getMessage(), e);
+ logger.warn("Unable to get connected SDC for the host: " + hostId
+ " and storage pool: " + poolId + " due to " + e.getMessage(), e);
}
return null;
}
private boolean hostSdcConnected(String sdcId, long poolId, int
waitTimeInSecs) {
- LOGGER.debug(String.format("Waiting (for %d secs) for the SDC %s of
the pool id: %d to connect", waitTimeInSecs, sdcId, poolId));
+ logger.debug(String.format("Waiting (for %d secs) for the SDC %s of
the pool id: %d to connect", waitTimeInSecs, sdcId, poolId));
int timeBetweenTries = 1000; // Try more frequently (every sec) and
return early if connected
while (waitTimeInSecs > 0) {
if (isHostSdcConnected(sdcId, poolId)) {
@@ -335,7 +336,7 @@ public class ScaleIOSDCManagerImpl implements
ScaleIOSDCManager {
final ScaleIOGatewayClient client = getScaleIOClient(poolId);
return client.isSdcConnected(sdcId);
} catch (Exception e) {
- LOGGER.error("Failed to check host SDC connection", e);
+ logger.error("Failed to check host SDC connection", e);
throw new CloudRuntimeException("Failed to establish connection
with PowerFlex Gateway to check host SDC connection");
}
}
diff --git
a/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFirePrimaryDataStoreLifeCycle.java
b/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFirePrimaryDataStoreLifeCycle.java
index 5080a413ac4..0b5393bd8d8 100644
---
a/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFirePrimaryDataStoreLifeCycle.java
+++
b/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFirePrimaryDataStoreLifeCycle.java
@@ -25,9 +25,6 @@ import java.util.UUID;
import javax.inject.Inject;
-import org.apache.logging.log4j.Logger;
-import org.apache.logging.log4j.LogManager;
-
import org.apache.cloudstack.engine.subsystem.api.storage.ClusterScope;
import org.apache.cloudstack.engine.subsystem.api.storage.DataStore;
import org.apache.cloudstack.engine.subsystem.api.storage.HostScope;
@@ -65,8 +62,6 @@ import com.cloud.utils.exception.CloudRuntimeException;
import com.google.common.base.Preconditions;
public class SolidFirePrimaryDataStoreLifeCycle extends
BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle {
- protected Logger logger = LogManager.getLogger(getClass());
-
@Inject private CapacityManager _capacityMgr;
@Inject private ClusterDao _clusterDao;
@Inject private DataCenterDao _zoneDao;
diff --git
a/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFireSharedPrimaryDataStoreLifeCycle.java
b/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFireSharedPrimaryDataStoreLifeCycle.java
index 9742bfb21c2..4877e86bf9f 100644
---
a/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFireSharedPrimaryDataStoreLifeCycle.java
+++
b/plugins/storage/volume/solidfire/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/SolidFireSharedPrimaryDataStoreLifeCycle.java
@@ -26,9 +26,6 @@ import java.util.Map;
import javax.inject.Inject;
-import org.apache.logging.log4j.Logger;
-import org.apache.logging.log4j.LogManager;
-
import org.apache.cloudstack.context.CallContext;
import org.apache.cloudstack.engine.subsystem.api.storage.ClusterScope;
import org.apache.cloudstack.engine.subsystem.api.storage.DataStore;
@@ -74,8 +71,6 @@ import com.cloud.utils.db.GlobalLock;
import com.cloud.utils.exception.CloudRuntimeException;
public class SolidFireSharedPrimaryDataStoreLifeCycle extends
BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle {
- protected Logger logger = LogManager.getLogger(getClass());
-
@Inject private AccountDao accountDao;
@Inject private AccountDetailsDao accountDetailsDao;
@Inject private AgentManager agentMgr;
diff --git
a/plugins/storage/volume/storpool/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/StorPoolPrimaryDataStoreLifeCycle.java
b/plugins/storage/volume/storpool/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/StorPoolPrimaryDataStoreLifeCycle.java
index e13b7f7b706..4f2fdef1723 100644
---
a/plugins/storage/volume/storpool/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/StorPoolPrimaryDataStoreLifeCycle.java
+++
b/plugins/storage/volume/storpool/src/main/java/org/apache/cloudstack/storage/datastore/lifecycle/StorPoolPrimaryDataStoreLifeCycle.java
@@ -38,8 +38,6 @@ import
org.apache.cloudstack.storage.datastore.util.StorPoolUtil;
import org.apache.cloudstack.storage.datastore.util.StorPoolUtil.SpApiResponse;
import
org.apache.cloudstack.storage.datastore.util.StorPoolUtil.SpConnectionDesc;
import org.apache.cloudstack.storage.volume.datastore.PrimaryDataStoreHelper;
-import org.apache.logging.log4j.Logger;
-import org.apache.logging.log4j.LogManager;
import com.cloud.agent.api.StoragePoolInfo;
import com.cloud.host.HostVO;
@@ -62,8 +60,6 @@ import com.cloud.storage.dao.VMTemplatePoolDao;
import com.cloud.utils.exception.CloudRuntimeException;
public class StorPoolPrimaryDataStoreLifeCycle extends
BasePrimaryDataStoreLifeCycleImpl implements PrimaryDataStoreLifeCycle {
- protected Logger logger = LogManager.getLogger(getClass());
-
@Inject
protected PrimaryDataStoreHelper dataStoreHelper;
@Inject