[jira] [Resolved] (CLOUDSTACK-4858) Disable copy snapshot to secondary storage snapshot.backup.rightafter

2017-05-09 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-4858?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller resolved CLOUDSTACK-4858.
--
   Resolution: Fixed
 Assignee: Nathan Johnson
Fix Version/s: 4.9.0

> Disable copy snapshot to secondary storage snapshot.backup.rightafter
> -
>
> Key: CLOUDSTACK-4858
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-4858
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: API
>Affects Versions: 4.2.0
> Environment: kvm, rbd
>Reporter: Sebastian Igerl
>Assignee: Nathan Johnson
> Fix For: 4.9.0
>
>
> "snapshot.backup.rightafterbackup"  is set to false but still the snapshot is 
> copied to secondary storage.
> Primary storage is Ceph/RDB, Secondary is NFS. I restarted both the 
> management server and the nodes.
> see : 
> http://www.marshut.com/wpviz/disable-copy-snapshot-to-secondary-storage-in-cs-4-2.html



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Updated] (CLOUDSTACK-4858) Disable copy snapshot to secondary storage snapshot.backup.rightafter

2017-05-09 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-4858?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller updated CLOUDSTACK-4858:
-
Fix Version/s: 4.10.0.0

> Disable copy snapshot to secondary storage snapshot.backup.rightafter
> -
>
> Key: CLOUDSTACK-4858
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-4858
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: API
>Affects Versions: 4.2.0
> Environment: kvm, rbd
>Reporter: Sebastian Igerl
>Assignee: Nathan Johnson
> Fix For: 4.9.0, 4.10.0.0
>
>
> "snapshot.backup.rightafterbackup"  is set to false but still the snapshot is 
> copied to secondary storage.
> Primary storage is Ceph/RDB, Secondary is NFS. I restarted both the 
> management server and the nodes.
> see : 
> http://www.marshut.com/wpviz/disable-copy-snapshot-to-secondary-storage-in-cs-4-2.html



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Updated] (CLOUDSTACK-9661) Allow Carrier Grade NAT IP ranges (RFC 6598) to be used within a VPC

2016-12-09 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9661?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller updated CLOUDSTACK-9661:
-
Affects Version/s: 4.8.0
   4.9.0
Fix Version/s: (was: Future)

> Allow Carrier Grade NAT IP ranges (RFC 6598) to be used within a VPC
> 
>
> Key: CLOUDSTACK-9661
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9661
> Project: CloudStack
>  Issue Type: Improvement
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: Management Server
>Affects Versions: 4.8.0, 4.9.0, 4.10.0.0
>Reporter: Simon Weller
>Assignee: Simon Weller
>Priority: Minor
>
> In a carrier environment, IP space is very precious. RFC 6598 specifies a 
> IANA-Reserved IPv4 Prefix for Shared Address Space that is used in carrier 
> networks for handling Carrier Grade NAT (CGN). Since this range is not 
> publicly routable, it can be used within a carriers network in a similar way 
> to RFC 1918 space.
> This PR adds support for RFC 6598 space.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CLOUDSTACK-9661) Allow Carrier Grade NAT IP ranges (RFC 6598) to be used within a VPC

2016-12-09 Thread Simon Weller (JIRA)
Simon Weller created CLOUDSTACK-9661:


 Summary: Allow Carrier Grade NAT IP ranges (RFC 6598) to be used 
within a VPC
 Key: CLOUDSTACK-9661
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9661
 Project: CloudStack
  Issue Type: Improvement
  Security Level: Public (Anyone can view this level - this is the default.)
  Components: Management Server
Affects Versions: 4.10.0.0
Reporter: Simon Weller
Assignee: Simon Weller
Priority: Minor
 Fix For: Future


In a carrier environment, IP space is very precious. RFC 6598 specifies a 
IANA-Reserved IPv4 Prefix for Shared Address Space that is used in carrier 
networks for handling Carrier Grade NAT (CGN). Since this range is not publicly 
routable, it can be used within a carriers network in a similar way to RFC 1918 
space.

This PR adds support for RFC 6598 space.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (CLOUDSTACK-9461) When RBD backup snapshot to secondary occurs, secondary storage image format is RAW

2016-08-18 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9461?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller reassigned CLOUDSTACK-9461:


Assignee: Simon Weller

> When RBD backup snapshot to secondary occurs, secondary storage image format 
> is RAW
> ---
>
> Key: CLOUDSTACK-9461
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9461
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: Storage Controller
>Affects Versions: 4.7.0, 4.8.0, 4.9.0
>Reporter: Simon Weller
>Assignee: Simon Weller
>Priority: Minor
> Fix For: Future
>
>
> RBD snapshot backup to secondary storage should use QCOW2, not RAW.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CLOUDSTACK-9461) When RBD backup snapshot to secondary occurs, secondary storage image format is RAW

2016-08-18 Thread Simon Weller (JIRA)
Simon Weller created CLOUDSTACK-9461:


 Summary: When RBD backup snapshot to secondary occurs, secondary 
storage image format is RAW
 Key: CLOUDSTACK-9461
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9461
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public (Anyone can view this level - this is the default.)
  Components: Storage Controller
Affects Versions: 4.9.0, 4.8.0, 4.7.0
Reporter: Simon Weller
Priority: Minor
 Fix For: Future


RBD snapshot backup to secondary storage should use QCOW2, not RAW.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CLOUDSTACK-4858) Disable copy snapshot to secondary storage snapshot.backup.rightafter

2016-08-08 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-4858?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15411728#comment-15411728
 ] 

Simon Weller commented on CLOUDSTACK-4858:
--

We have a working PR to fix this in the works. We're working on extending the 
test set around it and then we'll issue it against ACS.

- Si

> Disable copy snapshot to secondary storage snapshot.backup.rightafter
> -
>
> Key: CLOUDSTACK-4858
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-4858
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: API
>Affects Versions: 4.2.0
> Environment: kvm, rbd
>Reporter: Sebastian Igerl
>
> "snapshot.backup.rightafterbackup"  is set to false but still the snapshot is 
> copied to secondary storage.
> Primary storage is Ceph/RDB, Secondary is NFS. I restarted both the 
> management server and the nodes.
> see : 
> http://www.marshut.com/wpviz/disable-copy-snapshot-to-secondary-storage-in-cs-4-2.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CLOUDSTACK-9434) NPE on attempting account/domain cleanup automation

2016-07-21 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9434?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller updated CLOUDSTACK-9434:
-
Attachment: dump.txt

MySQL query logs for the NPE

> NPE on attempting account/domain cleanup automation
> ---
>
> Key: CLOUDSTACK-9434
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9434
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: Automation
>Affects Versions: 4.8.0
> Environment: Centos 7.2, KVM
>Reporter: Simon Weller
>Priority: Minor
> Attachments: dump.txt
>
>
> When attempting to delete a domain, an NPE is thrown trying to cleanup 
> associated accounts.
> 2016-07-21 08:48:58,119 INFO  [c.c.u.AccountManagerImpl] 
> (AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Found 5 removed accounts to 
> cleanup
> 2016-07-21 08:48:58,119 DEBUG [c.c.u.AccountManagerImpl] 
> (AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Cleaning up 12
> 2016-07-21 08:48:58,125 DEBUG [c.c.u.AccountManagerImpl] 
> (AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Successfully deleted 
> snapshots directories for all volumes under account 12 across all zones
> 2016-07-21 08:48:58,129 DEBUG [c.c.u.AccountManagerImpl] 
> (AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Expunging # of vms 
> (accountId=12): 0
> 2016-07-21 08:48:58,134 INFO  [c.c.u.AccountManagerImpl] 
> (AccountChecker-1:ctx-d349b03d) (logid:61884c2d) deleteAccount: Deleted 0 
> network groups for account 12
> 2016-07-21 08:48:58,135 INFO  [c.c.u.AccountManagerImpl] 
> (AccountChecker-1:ctx-d349b03d) (logid:61884c2d) deleteAccount: Deleted 0 
> affinity groups for account 12
> 2016-07-21 08:48:58,135 DEBUG [c.c.u.AccountManagerImpl] 
> (AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Deleting networks for 
> account 12
> 2016-07-21 08:48:58,138 DEBUG [c.c.u.AccountManagerImpl] 
> (AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Deleting vpcs for account 12
> 2016-07-21 08:48:58,140 DEBUG [c.c.u.AccountManagerImpl] 
> (AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Deleting site-to-site VPN 
> customer gateways for account 12
> 2016-07-21 08:48:58,148 DEBUG [c.c.u.d.T.Transaction] 
> (AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Rolling back the 
> transaction: Time = 2 Name =  AccountChecker-1; called by 
> -TransactionLegacy.rollback:879-TransactionLegacy.removeUpTo:822-TransactionLegacy.close:646-Transaction.execute:43-Transaction.execute:47-ConfigurationManagerImpl.releaseAccountSpecificVirtualRanges:4941-GeneratedMethodAccessor764.invoke:-1-DelegatingMethodAccessorImpl.invoke:43-Method.invoke:606-AopUtils.invokeJoinpointUsingReflection:317-ReflectiveMethodInvocation.invokeJoinpoint:183-ReflectiveMethodInvocation.proceed:150
> 2016-07-21 08:48:58,150 WARN  [c.c.u.AccountManagerImpl] 
> (AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Failed to cleanup account 
> Acct[9f367fca-d4c6-4603-b487-afd4e3d46dfc-admin-2177] due to 
> java.lang.NullPointerException
> at 
> com.cloud.configuration.ConfigurationManagerImpl.releasePublicIpRange(ConfigurationManagerImpl.java:3473)
> at 
> com.cloud.configuration.ConfigurationManagerImpl$12.doInTransactionWithoutResult(ConfigurationManagerImpl.java:4945)
> at 
> com.cloud.utils.db.TransactionCallbackNoReturn.doInTransaction(TransactionCallbackNoReturn.java:25)
> at 
> com.cloud.utils.db.Transaction$2.doInTransaction(Transaction.java:50)
> at com.cloud.utils.db.Transaction.execute(Transaction.java:40)
> at com.cloud.utils.db.Transaction.execute(Transaction.java:47)
> at 
> com.cloud.configuration.ConfigurationManagerImpl.releaseAccountSpecificVirtualRanges(ConfigurationManagerImpl.java:4941)
> at sun.reflect.GeneratedMethodAccessor764.invoke(Unknown Source)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at 
> org.springframework.aop.support.AopUtils.invokeJoinpointUsingReflection(AopUtils.java:317)
> at 
> org.springframework.aop.framework.ReflectiveMethodInvocation.invokeJoinpoint(ReflectiveMethodInvocation.java:183)
> at 
> org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:150)
> at 
> org.springframework.aop.interceptor.ExposeInvocationInterceptor.invoke(ExposeInvocationInterceptor.java:91)
> at 
> org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:172)
> at 
> org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:204)
> at 
> com.sun.proxy.$Proxy117.releaseAccountSpecificVirtualRanges(Unknown 

[jira] [Created] (CLOUDSTACK-9434) NPE on attempting account/domain cleanup automation

2016-07-21 Thread Simon Weller (JIRA)
Simon Weller created CLOUDSTACK-9434:


 Summary: NPE on attempting account/domain cleanup automation
 Key: CLOUDSTACK-9434
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9434
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public (Anyone can view this level - this is the default.)
  Components: Automation
Affects Versions: 4.8.0
 Environment: Centos 7.2, KVM
Reporter: Simon Weller
Priority: Minor


When attempting to delete a domain, an NPE is thrown trying to cleanup 
associated accounts.

2016-07-21 08:48:58,119 INFO  [c.c.u.AccountManagerImpl] 
(AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Found 5 removed accounts to 
cleanup
2016-07-21 08:48:58,119 DEBUG [c.c.u.AccountManagerImpl] 
(AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Cleaning up 12
2016-07-21 08:48:58,125 DEBUG [c.c.u.AccountManagerImpl] 
(AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Successfully deleted snapshots 
directories for all volumes under account 12 across all zones
2016-07-21 08:48:58,129 DEBUG [c.c.u.AccountManagerImpl] 
(AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Expunging # of vms 
(accountId=12): 0
2016-07-21 08:48:58,134 INFO  [c.c.u.AccountManagerImpl] 
(AccountChecker-1:ctx-d349b03d) (logid:61884c2d) deleteAccount: Deleted 0 
network groups for account 12
2016-07-21 08:48:58,135 INFO  [c.c.u.AccountManagerImpl] 
(AccountChecker-1:ctx-d349b03d) (logid:61884c2d) deleteAccount: Deleted 0 
affinity groups for account 12
2016-07-21 08:48:58,135 DEBUG [c.c.u.AccountManagerImpl] 
(AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Deleting networks for account 
12
2016-07-21 08:48:58,138 DEBUG [c.c.u.AccountManagerImpl] 
(AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Deleting vpcs for account 12
2016-07-21 08:48:58,140 DEBUG [c.c.u.AccountManagerImpl] 
(AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Deleting site-to-site VPN 
customer gateways for account 12
2016-07-21 08:48:58,148 DEBUG [c.c.u.d.T.Transaction] 
(AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Rolling back the transaction: 
Time = 2 Name =  AccountChecker-1; called by 
-TransactionLegacy.rollback:879-TransactionLegacy.removeUpTo:822-TransactionLegacy.close:646-Transaction.execute:43-Transaction.execute:47-ConfigurationManagerImpl.releaseAccountSpecificVirtualRanges:4941-GeneratedMethodAccessor764.invoke:-1-DelegatingMethodAccessorImpl.invoke:43-Method.invoke:606-AopUtils.invokeJoinpointUsingReflection:317-ReflectiveMethodInvocation.invokeJoinpoint:183-ReflectiveMethodInvocation.proceed:150
2016-07-21 08:48:58,150 WARN  [c.c.u.AccountManagerImpl] 
(AccountChecker-1:ctx-d349b03d) (logid:61884c2d) Failed to cleanup account 
Acct[9f367fca-d4c6-4603-b487-afd4e3d46dfc-admin-2177] due to 
java.lang.NullPointerException
at 
com.cloud.configuration.ConfigurationManagerImpl.releasePublicIpRange(ConfigurationManagerImpl.java:3473)
at 
com.cloud.configuration.ConfigurationManagerImpl$12.doInTransactionWithoutResult(ConfigurationManagerImpl.java:4945)
at 
com.cloud.utils.db.TransactionCallbackNoReturn.doInTransaction(TransactionCallbackNoReturn.java:25)
at com.cloud.utils.db.Transaction$2.doInTransaction(Transaction.java:50)
at com.cloud.utils.db.Transaction.execute(Transaction.java:40)
at com.cloud.utils.db.Transaction.execute(Transaction.java:47)
at 
com.cloud.configuration.ConfigurationManagerImpl.releaseAccountSpecificVirtualRanges(ConfigurationManagerImpl.java:4941)
at sun.reflect.GeneratedMethodAccessor764.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.springframework.aop.support.AopUtils.invokeJoinpointUsingReflection(AopUtils.java:317)
at 
org.springframework.aop.framework.ReflectiveMethodInvocation.invokeJoinpoint(ReflectiveMethodInvocation.java:183)
at 
org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:150)
at 
org.springframework.aop.interceptor.ExposeInvocationInterceptor.invoke(ExposeInvocationInterceptor.java:91)
at 
org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:172)
at 
org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:204)
at com.sun.proxy.$Proxy117.releaseAccountSpecificVirtualRanges(Unknown 
Source)
at 
com.cloud.user.AccountManagerImpl.cleanupAccount(AccountManagerImpl.java:869)
at 
com.cloud.user.AccountManagerImpl$AccountCleanupTask.runInContext(AccountManagerImpl.java:1722)
at 
org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49)
at 

[jira] [Commented] (CLOUDSTACK-9209) Database HA not working on a fresh install on Ubuntu 12.04

2016-06-28 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9209?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15353668#comment-15353668
 ] 

Simon Weller commented on CLOUDSTACK-9209:
--

Tom,

This issue is fixed in master (future 4.9).

You can change a couple of config files in your 4.8 install to fix this. Please 
see 
https://github.com/apache/cloudstack/pull/1428/commits/c22659d76d73f00f41c13776c490e17a50aacd20
 for the changes.

- Si



> Database HA not working on a fresh install on Ubuntu 12.04
> --
>
> Key: CLOUDSTACK-9209
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9209
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: Install and Setup
>Affects Versions: 4.7.0
> Environment: Two Ubuntu 12.04 Management Servers, Mysql Master Master 
> Replication, Official community repository: 
> http://cloudstack.apt-get.eu/ubuntu/dists/precise/4.7/
>Reporter: Thomas
>Priority: Blocker
>
> The cloudstack-management service will not come up because of this error:
> 2016-01-05 16:15:00,899 INFO  [c.c.u.d.T.Transaction] (main:null) (logid:) Is 
> Data Base High Availiability enabled? Ans : true
> 2016-01-05 16:15:00,924 INFO  [c.c.u.d.T.Transaction] (main:null) (logid:) 
> The slaves configured for Cloud Data base is/are : 192.168.0.90
> 2016-01-05 16:15:00,974 ERROR [c.c.u.d.Merovingian2] (main:null) (logid:) 
> Unable to get a new db connection
> java.sql.SQLException: Invalid load balancing strategy 
> 'com.cloud.utils.db.StaticStrategy'.
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:934)
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:931)
> at com.mysql.jdbc.Util.loadExtensions(Util.java:588)
> at 
> com.mysql.jdbc.LoadBalancingConnectionProxy.(LoadBalancingConnectionProxy.java:279)
> at 
> com.mysql.jdbc.FailoverConnectionProxy.(FailoverConnectionProxy.java:66)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connectFailover(NonRegisteringDriver.java:382)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:301)
> at java.sql.DriverManager.getConnection(DriverManager.java:571)
> at java.sql.DriverManager.getConnection(DriverManager.java:215)
> at 
> org.apache.commons.dbcp.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:75)
> at 
> org.apache.commons.dbcp.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:582)
> at 
> org.apache.commons.pool.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:1188)
> at 
> org.apache.commons.dbcp.PoolingDataSource.getConnection(PoolingDataSource.java:106)
> at 
> com.cloud.utils.db.TransactionLegacy.getStandaloneConnectionWithException(TransactionLegacy.java:202)
> at com.cloud.utils.db.Merovingian2.(Merovingian2.java:68)
> at 
> com.cloud.utils.db.Merovingian2.createLockMaster(Merovingian2.java:88)
> at 
> com.cloud.server.LockMasterListener.(LockMasterListener.java:33)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
> Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> at 
> org.springframework.beans.BeanUtils.instantiateClass(BeanUtils.java:148)
> at 
> org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:121)
> at 
> org.springframework.beans.factory.support.ConstructorResolver.autowireConstructor(ConstructorResolver.java:277)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.autowireConstructor(AbstractAutowireCapableBeanFactory.java:1077)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:981)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:487)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:458)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory$1.getObject(AbstractBeanFactory.java:293)
> at 
> org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:223)
> at 
> 

[jira] [Commented] (CLOUDSTACK-9370) Failed to create VPC: Unable to start VPC VR (VM DomainRouter) due to error in finalizeStart, not retrying

2016-04-27 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9370?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15260136#comment-15260136
 ] 

Simon Weller commented on CLOUDSTACK-9370:
--

According to the 4.8 docs, OVS and KVM VPC doesn't currently work. I'm not 
aware of any work that has changed this in master.

http://docs.cloudstack.apache.org/en/latest/networking/ovs-plugin.html#using-the-ovs-plugin-with-vpc


> Failed to create VPC: Unable to start  VPC VR (VM DomainRouter) due to error 
> in finalizeStart, not retrying
> ---
>
> Key: CLOUDSTACK-9370
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9370
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: Virtual Router
>Affects Versions: 4.9.0
> Environment: Centos El7
> KVM
> OpenvSwitch (VLAN 0)
> NFS (primary/secondary) 
>Reporter: Mani Prashanth Varma Manthena
>Priority: Critical
> Fix For: 4.9.0
>
>
> I am unable to create VPCs on latest cloudstack master due to the following 
> error:
> {noformat:title=Root Cause Error in Agent log}
> 2016-04-27 02:31:03,134 DEBUG [kvm.resource.LibvirtComputingResource] 
> (agentRequest-Handler-2:null) (logid:6b2d4faa) [INFO] update_config.py :: 
> Processing incoming file => ip_associations.json[INFO] Processing JSON file 
> ip_associations.jsonTraceback (most recent call last):  File 
> "/opt/cloud/bin/update_config.py", line 140, in process_file()  
> File "/opt/cloud/bin/update_config.py", line 52, in process_file
> qf.load(None)  File "/opt/cloud/bin/merge.py", line 258, in loadproc = 
> updateDataBag(self)  File "/opt/cloud/bin/merge.py", line 91, in __init__
> self.process()  File "/opt/cloud/bin/merge.py", line 103, in processdbag 
> = self.processIP(self.db.getDataBag())  File "/opt/cloud/bin/merge.py", line 
> 190, in processIPdbag = cs_ip.merge(dbag, ip)  File 
> "/opt/cloud/bin/cs_ip.py", line 32, in mergeip['device'] = 'eth' + 
> str(ip['nic_dev_id'])KeyError: 'nic_dev_id'
> 2016-04-27 02:31:03,135 DEBUG 
> [resource.virtualnetwork.VirtualRoutingResource] 
> (agentRequest-Handler-2:null) (logid:6b2d4faa) Processing ScriptConfigItem, 
> executing update_config.py ip_associations.json took 911ms
> {noformat}
> {noformat:title=Root Cause Error in Management Server log}
> 2016-04-27 02:30:19,975 DEBUG [c.c.a.m.ClusteredAgentAttache] 
> (Work-Job-Executor-10:ctx-1279b068 job-1159/job-1160 ctx-c31efe73) 
> (logid:6b2d4faa) Seq 9-332421947495286159: Forwarding Seq 
> 9-332421947495286159:  { Cmd , MgmtId: 275619427298304, via: 
> 9(ovs-2.mvdcvtb16.us.alcatel-lucent.com), Ver: v1, Flags: 100111, 
> [{"com.cloud.agent.api.StartCommand":{"vm":{"id":252,"name":"r-252-VM","type":"DomainRouter","cpus":1,"minSpeed":500,"maxSpeed":500,"minRam":268435456,"maxRam":268435456,"arch":"x86_64","os":"Debian
>  GNU/Linux 5.0 (64-bit)","platformEmulator":"Debian GNU/Linux 5","bootArgs":" 
> vpccidr=10.1.1.1/16 domain=cs2cloud.internal dns1=128.251.10.29 template=domP 
> name=r-252-VM eth0ip=169.254.1.123 eth0mask=255.255.0.0 type=vpcrouter 
> disable_rp_filter=true 
> baremetalnotificationsecuritykey=0oLpL4swbL6Yu_xsuRdyjwmmyPHAU1V-iMpmMNKO00vNIP5bxronvhQZ_qehiEZ99Eo9avCHg9uLh1cbiz7pQA
>  
> baremetalnotificationapikey=wEax_CyEaKZHn8ZkPBQLQaibjSWZ0OYJuEQA3l2RUA41GXZxaie9P6oQPeNlzjIGl-fDpKWp9MkAEQOJYvE4vA
>  host=10.31.59.151 
> 

[jira] [Updated] (CLOUDSTACK-9300) MySQL HA feature StaticStrategy throws exception

2016-04-20 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller updated CLOUDSTACK-9300:
-
Fix Version/s: 4.9.0

> MySQL HA feature StaticStrategy throws exception
> 
>
> Key: CLOUDSTACK-9300
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9300
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>Affects Versions: 4.7.0, 4.7.1, 4.8.0, Future
> Environment: Centos 7
>Reporter: Simon Weller
>Assignee: Simon Weller
>Priority: Minor
> Fix For: 4.9.0
>
>
> 2016-03-03 12:00:13,204 INFO  [c.c.u.d.T.Transaction] 
> (localhost-startStop-1:null) (logid:) Is Data Base High Availiability 
> enabled? Ans : true
> 2016-03-03 12:00:13,239 INFO  [c.c.u.d.T.Transaction] 
> (localhost-startStop-1:null) (logid:) The slaves configured for Cloud Data 
> base is/are : localhost,localhost
> 2016-03-03 12:00:13,303 ERROR [c.c.u.d.Merovingian2] 
> (localhost-startStop-1:null) (logid:) Unable to get a new db connection
> java.sql.SQLException: Invalid load balancing strategy 
> 'com.cloud.utils.db.StaticStrategy'.
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:927)
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:924)
> at com.mysql.jdbc.Util.loadExtensions(Util.java:602)
> at 
> com.mysql.jdbc.LoadBalancingConnectionProxy.(LoadBalancingConnectionProxy.java:280)
> at 
> com.mysql.jdbc.FailoverConnectionProxy.(FailoverConnectionProxy.java:67)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connectFailover(NonRegisteringDriver.java:433)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:346)
> at java.sql.DriverManager.getConnection(DriverManager.java:571)
> at java.sql.DriverManager.getConnection(DriverManager.java:215)
> at 
> org.apache.commons.dbcp.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:75)
> at 
> org.apache.commons.dbcp.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:582)
> at 
> org.apache.commons.pool.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:1188)
> at 
> org.apache.commons.dbcp.PoolingDataSource.getConnection(PoolingDataSource.java:106)
> at 
> com.cloud.utils.db.TransactionLegacy.getStandaloneConnectionWithException(TransactionLegacy.java:202)
> at com.cloud.utils.db.Merovingian2.(Merovingian2.java:68)
> at 
> com.cloud.utils.db.Merovingian2.createLockMaster(Merovingian2.java:88)
> at 
> com.cloud.server.LockMasterListener.(LockMasterListener.java:33)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
> Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> at 
> org.springframework.beans.BeanUtils.instantiateClass(BeanUtils.java:148)
> at 
> org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:121)
> at 
> org.springframework.beans.factory.support.ConstructorResolver.autowireConstructor(ConstructorResolver.java:277)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.autowireConstructor(AbstractAutowireCapableBeanFactory.java:1077)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:981)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:487)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:458)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory$1.getObject(AbstractBeanFactory.java:293)
> at 
> org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:223)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:290)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:191)
> at 
> org.springframework.beans.factory.support.DefaultListableBeanFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:636)
> at 
> 

[jira] [Assigned] (CLOUDSTACK-9300) MySQL HA feature StaticStrategy throws exception

2016-04-20 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller reassigned CLOUDSTACK-9300:


Assignee: Simon Weller

> MySQL HA feature StaticStrategy throws exception
> 
>
> Key: CLOUDSTACK-9300
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9300
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>Affects Versions: 4.7.0, 4.7.1, 4.8.0, Future
> Environment: Centos 7
>Reporter: Simon Weller
>Assignee: Simon Weller
>Priority: Minor
>
> 2016-03-03 12:00:13,204 INFO  [c.c.u.d.T.Transaction] 
> (localhost-startStop-1:null) (logid:) Is Data Base High Availiability 
> enabled? Ans : true
> 2016-03-03 12:00:13,239 INFO  [c.c.u.d.T.Transaction] 
> (localhost-startStop-1:null) (logid:) The slaves configured for Cloud Data 
> base is/are : localhost,localhost
> 2016-03-03 12:00:13,303 ERROR [c.c.u.d.Merovingian2] 
> (localhost-startStop-1:null) (logid:) Unable to get a new db connection
> java.sql.SQLException: Invalid load balancing strategy 
> 'com.cloud.utils.db.StaticStrategy'.
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:927)
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:924)
> at com.mysql.jdbc.Util.loadExtensions(Util.java:602)
> at 
> com.mysql.jdbc.LoadBalancingConnectionProxy.(LoadBalancingConnectionProxy.java:280)
> at 
> com.mysql.jdbc.FailoverConnectionProxy.(FailoverConnectionProxy.java:67)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connectFailover(NonRegisteringDriver.java:433)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:346)
> at java.sql.DriverManager.getConnection(DriverManager.java:571)
> at java.sql.DriverManager.getConnection(DriverManager.java:215)
> at 
> org.apache.commons.dbcp.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:75)
> at 
> org.apache.commons.dbcp.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:582)
> at 
> org.apache.commons.pool.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:1188)
> at 
> org.apache.commons.dbcp.PoolingDataSource.getConnection(PoolingDataSource.java:106)
> at 
> com.cloud.utils.db.TransactionLegacy.getStandaloneConnectionWithException(TransactionLegacy.java:202)
> at com.cloud.utils.db.Merovingian2.(Merovingian2.java:68)
> at 
> com.cloud.utils.db.Merovingian2.createLockMaster(Merovingian2.java:88)
> at 
> com.cloud.server.LockMasterListener.(LockMasterListener.java:33)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
> Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> at 
> org.springframework.beans.BeanUtils.instantiateClass(BeanUtils.java:148)
> at 
> org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:121)
> at 
> org.springframework.beans.factory.support.ConstructorResolver.autowireConstructor(ConstructorResolver.java:277)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.autowireConstructor(AbstractAutowireCapableBeanFactory.java:1077)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:981)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:487)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:458)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory$1.getObject(AbstractBeanFactory.java:293)
> at 
> org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:223)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:290)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:191)
> at 
> org.springframework.beans.factory.support.DefaultListableBeanFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:636)
> at 
> 

[jira] [Commented] (CLOUDSTACK-9349) Unable to detach root volume when using Hypervisor Type KVM

2016-04-14 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9349?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15241212#comment-15241212
 ] 

Simon Weller commented on CLOUDSTACK-9349:
--

Here's the commit: 
https://github.com/myENA/cloudstack/commit/d0a02640dfd4878da81a2e59588c4b5ff2a06401

We're going to complete some further testing prior to submitting a PR.



> Unable to detach root volume when using Hypervisor Type KVM
> ---
>
> Key: CLOUDSTACK-9349
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9349
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: Volumes
>Affects Versions: 4.5.1, 4.6.2, 4.7.1, 4.8.0, 4.9.0
> Environment: Centos 7
>Reporter: Simon Weller
>Priority: Minor
> Fix For: 4.7.2
>
>
> Back in 4.5, support was added in CLOUDSTACK-6284 for detaching root volumes. 
> The original support was meant to work with Xen, VMware and KVM.
> After chatting with fuflo in the Cloudstack irc channel, it was pointed out 
> that a constraint was not correctly modified in VolumeApiServiceImpl.java to 
> allow the detach to occur when vm.getHypervisorType() == HypervisorType.KVM.
> This is a very useful feature, as it allows us to simulate a snapshot revert 
> with Ceph by using createVolume sourced from a snapshot, then detaching and 
> reattaching the root volume (new root volume needs to be attached as 
> device=0).
> I'm going to propose a PR for this shortly



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CLOUDSTACK-9349) Unable to detach root volume when using Hypervisor Type KVM

2016-04-14 Thread Simon Weller (JIRA)
Simon Weller created CLOUDSTACK-9349:


 Summary: Unable to detach root volume when using Hypervisor Type 
KVM
 Key: CLOUDSTACK-9349
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9349
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public (Anyone can view this level - this is the default.)
  Components: Volumes
Affects Versions: 4.8.0, 4.7.1, 4.6.2, 4.5.1, 4.9.0
 Environment: Centos 7
Reporter: Simon Weller
Priority: Minor
 Fix For: 4.7.2


Back in 4.5, support was added in CLOUDSTACK-6284 for detaching root volumes. 
The original support was meant to work with Xen, VMware and KVM.

After chatting with fuflo in the Cloudstack irc channel, it was pointed out 
that a constraint was not correctly modified in VolumeApiServiceImpl.java to 
allow the detach to occur when vm.getHypervisorType() == HypervisorType.KVM.

This is a very useful feature, as it allows us to simulate a snapshot revert 
with Ceph by using createVolume sourced from a snapshot, then detaching and 
reattaching the root volume (new root volume needs to be attached as device=0).

I'm going to propose a PR for this shortly



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CLOUDSTACK-9209) Database HA not working on a fresh install on Ubuntu 12.04

2016-03-22 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9209?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15206689#comment-15206689
 ] 

Simon Weller commented on CLOUDSTACK-9209:
--

Please see my PR tested on Centos 7 that addresses this. I expect the Ubuntu 
issue is the same, but I have no way of testing it.
https://github.com/apache/cloudstack/pull/1428
Usage server issue is address here: 
https://github.com/apache/cloudstack/pull/1433

> Database HA not working on a fresh install on Ubuntu 12.04
> --
>
> Key: CLOUDSTACK-9209
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9209
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: Install and Setup
>Affects Versions: 4.7.0
> Environment: Two Ubuntu 12.04 Management Servers, Mysql Master Master 
> Replication, Official community repository: 
> http://cloudstack.apt-get.eu/ubuntu/dists/precise/4.7/
>Reporter: Thomas
>Priority: Blocker
>
> The cloudstack-management service will not come up because of this error:
> 2016-01-05 16:15:00,899 INFO  [c.c.u.d.T.Transaction] (main:null) (logid:) Is 
> Data Base High Availiability enabled? Ans : true
> 2016-01-05 16:15:00,924 INFO  [c.c.u.d.T.Transaction] (main:null) (logid:) 
> The slaves configured for Cloud Data base is/are : 192.168.0.90
> 2016-01-05 16:15:00,974 ERROR [c.c.u.d.Merovingian2] (main:null) (logid:) 
> Unable to get a new db connection
> java.sql.SQLException: Invalid load balancing strategy 
> 'com.cloud.utils.db.StaticStrategy'.
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:934)
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:931)
> at com.mysql.jdbc.Util.loadExtensions(Util.java:588)
> at 
> com.mysql.jdbc.LoadBalancingConnectionProxy.(LoadBalancingConnectionProxy.java:279)
> at 
> com.mysql.jdbc.FailoverConnectionProxy.(FailoverConnectionProxy.java:66)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connectFailover(NonRegisteringDriver.java:382)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:301)
> at java.sql.DriverManager.getConnection(DriverManager.java:571)
> at java.sql.DriverManager.getConnection(DriverManager.java:215)
> at 
> org.apache.commons.dbcp.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:75)
> at 
> org.apache.commons.dbcp.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:582)
> at 
> org.apache.commons.pool.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:1188)
> at 
> org.apache.commons.dbcp.PoolingDataSource.getConnection(PoolingDataSource.java:106)
> at 
> com.cloud.utils.db.TransactionLegacy.getStandaloneConnectionWithException(TransactionLegacy.java:202)
> at com.cloud.utils.db.Merovingian2.(Merovingian2.java:68)
> at 
> com.cloud.utils.db.Merovingian2.createLockMaster(Merovingian2.java:88)
> at 
> com.cloud.server.LockMasterListener.(LockMasterListener.java:33)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
> Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> at 
> org.springframework.beans.BeanUtils.instantiateClass(BeanUtils.java:148)
> at 
> org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:121)
> at 
> org.springframework.beans.factory.support.ConstructorResolver.autowireConstructor(ConstructorResolver.java:277)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.autowireConstructor(AbstractAutowireCapableBeanFactory.java:1077)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:981)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:487)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:458)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory$1.getObject(AbstractBeanFactory.java:293)
> at 
> org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:223)
> at 
> 

[jira] [Commented] (CLOUDSTACK-9079) ipsec service is not running after restarting virtual router

2016-03-22 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9079?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15206681#comment-15206681
 ] 

Simon Weller commented on CLOUDSTACK-9079:
--

Is this the same issue as https://issues.apache.org/jira/browse/CLOUDSTACK-9296?


> ipsec service is not running after restarting virtual router
> 
>
> Key: CLOUDSTACK-9079
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9079
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: SystemVM, Virtual Router
>Affects Versions: 4.6.0
> Environment: ACS 4.5.1 upgraded to 4.6.0
>Reporter: Erik Weber
>Priority: Critical
>  Labels: systemvm, virtualrouter, vpn
>
> after upgrading to 4.6.0 the ipsec service does no longer automatically start 
> and as an effect vpn does not work.
> manually starting the ipsec service, or doing any vpn related configuration 
> (adding/deleting users) remedies the problem.
> Steps to reproduce:
> 1) Install ACS 4.6.0
> 2) Configure a VPC (I haven't tested with normal isolated networks)
> 3) Enable remote access VPN, configure a user
> 4) Confirm that VPN works
> 5) Restart the VR
> 6) Confirm that VPN does NOT work, either by trying to connect or by issuing 
> 'service ipsec status' on the VR



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CLOUDSTACK-9285) Cloudstack 4.8 can't connect to XEN and KVM hosts

2016-03-04 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15180338#comment-15180338
 ] 

Simon Weller commented on CLOUDSTACK-9285:
--

Pull request now available to address issue: 
https://github.com/apache/cloudstack/pull/1429

> Cloudstack 4.8 can't connect to XEN and KVM hosts
> -
>
> Key: CLOUDSTACK-9285
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9285
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: KVM, Xen
>Affects Versions: 4.8.0
> Environment: CentOS 7
>Reporter: Jens Fettig
>Priority: Blocker
> Attachments: agentfailure.pcap
>
>
> We used Cloudstack 4.7 in our testing environment. For future uses we updated 
> to Cloudstack 4.8 for some tests. But after the update from 4.7 to 4.8 
> cloudstack has some problems to connect to the hosts again. Here are the 
> error logs.
> {code:title=XEN Server from the managment-server.log:|borderStyle=solid}
> 2016-02-15 11:22:03,476 DEBUG [c.c.a.m.ClusteredAgentManagerImpl] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec) Loading directly 
> connected host 13(xen5)
> 2016-02-15 11:22:06,204 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-1:ctx-aaf933f3) (logid:d157e37d) AutoScaling Monitor is 
> running...
> 2016-02-15 11:22:06,206 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-2:ctx-8e2fc084) (logid:379074cb) VmStatsCollector is 
> running...
> 2016-02-15 11:22:06,217 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-4:ctx-cc1cf960) (logid:b6397e2e) HostStatsCollector is 
> running...
> 2016-02-15 11:22:06,298 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-3:ctx-efda9c07) (logid:5731423e) StorageCollector is 
> running...
> 2016-02-15 11:22:06,302 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-3:ctx-efda9c07) (logid:5731423e) There is no secondary 
> storage VM for secondary storage host cdsdev-secondary
> 2016-02-15 11:22:11,018 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl] 
> (AsyncJobMgr-Heartbeat-1:ctx-bc740746) (logid:cc619bad) Begin cleanup expired 
> async-jobs
> 2016-02-15 11:22:11,024 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl] 
> (AsyncJobMgr-Heartbeat-1:ctx-bc740746) (logid:cc619bad) End cleanup expired 
> async-jobs
> 2016-02-15 11:22:11,055 DEBUG [c.c.h.x.r.XenServerConnectionPool] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec) Unable to create 
> master connection to host(192.168.0.97) , due to The credentials given by the 
> user are incorrect, so access has been denied, and you have not been issued a 
> session handle.
> 2016-02-15 11:22:11,055 DEBUG [c.c.h.Status] (ClusteredAgentManager 
> Timer:ctx-43ef0a0a) (logid:82d7b1ec) Transition:[Resource state = Enabled, 
> Agent event = AgentDisconnected, Host id = 13, name = xen5]
> 2016-02-15 11:22:11,101 WARN  [c.c.a.m.ClusteredAgentManagerImpl] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec)  can not load 
> directly connected host 13(xen5) due to 
> com.cloud.utils.exception.CloudRuntimeException: Unable to create master 
> connection to host(192.168.0.97) , due to The credentials given by the user 
> are incorrect, so access has been denied, and you have not been issued a 
> session handle.
>   at 
> com.cloud.hypervisor.xenserver.resource.XenServerConnectionPool.getConnect(XenServerConnectionPool.java:163)
>   at 
> com.cloud.hypervisor.xenserver.resource.CitrixResourceBase.CheckXenHostInfo(CitrixResourceBase.java:523)
>   at 
> com.cloud.hypervisor.xenserver.resource.CitrixResourceBase.configure(CitrixResourceBase.java:827)
>   at 
> com.cloud.resource.DiscovererBase.reloadResource(DiscovererBase.java:158)
>   at 
> com.cloud.agent.manager.AgentManagerImpl.loadDirectlyConnectedHost(AgentManagerImpl.java:697)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.scanDirectAgentToLoad(ClusteredAgentManagerImpl.java:217)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.runDirectAgentScanTimerTask(ClusteredAgentManagerImpl.java:182)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.access$100(ClusteredAgentManagerImpl.java:96)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl$DirectAgentScanTimerTask.runInContext(ClusteredAgentManagerImpl.java:233)
>   at 
> org.apache.cloudstack.managed.context.ManagedContextTimerTask$1.runInContext(ManagedContextTimerTask.java:30)
>   at 
> org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49)
>   at 
> org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:56)
>   at 
> org.apache.cloudstack.managed.context.impl.DefaultManagedContext.callWithContext(DefaultManagedContext.java:103)
>

[jira] [Commented] (CLOUDSTACK-9300) MySQL HA feature StaticStrategy throws exception

2016-03-04 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9300?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15179974#comment-15179974
 ] 

Simon Weller commented on CLOUDSTACK-9300:
--

Pull requested submitted to master for consideration: 
https://github.com/apache/cloudstack/pull/1428

> MySQL HA feature StaticStrategy throws exception
> 
>
> Key: CLOUDSTACK-9300
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9300
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>Affects Versions: Future, 4.7.0, 4.8.0, 4.7.1
> Environment: Centos 7
>Reporter: Simon Weller
>Priority: Minor
>
> 2016-03-03 12:00:13,204 INFO  [c.c.u.d.T.Transaction] 
> (localhost-startStop-1:null) (logid:) Is Data Base High Availiability 
> enabled? Ans : true
> 2016-03-03 12:00:13,239 INFO  [c.c.u.d.T.Transaction] 
> (localhost-startStop-1:null) (logid:) The slaves configured for Cloud Data 
> base is/are : localhost,localhost
> 2016-03-03 12:00:13,303 ERROR [c.c.u.d.Merovingian2] 
> (localhost-startStop-1:null) (logid:) Unable to get a new db connection
> java.sql.SQLException: Invalid load balancing strategy 
> 'com.cloud.utils.db.StaticStrategy'.
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:927)
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:924)
> at com.mysql.jdbc.Util.loadExtensions(Util.java:602)
> at 
> com.mysql.jdbc.LoadBalancingConnectionProxy.(LoadBalancingConnectionProxy.java:280)
> at 
> com.mysql.jdbc.FailoverConnectionProxy.(FailoverConnectionProxy.java:67)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connectFailover(NonRegisteringDriver.java:433)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:346)
> at java.sql.DriverManager.getConnection(DriverManager.java:571)
> at java.sql.DriverManager.getConnection(DriverManager.java:215)
> at 
> org.apache.commons.dbcp.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:75)
> at 
> org.apache.commons.dbcp.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:582)
> at 
> org.apache.commons.pool.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:1188)
> at 
> org.apache.commons.dbcp.PoolingDataSource.getConnection(PoolingDataSource.java:106)
> at 
> com.cloud.utils.db.TransactionLegacy.getStandaloneConnectionWithException(TransactionLegacy.java:202)
> at com.cloud.utils.db.Merovingian2.(Merovingian2.java:68)
> at 
> com.cloud.utils.db.Merovingian2.createLockMaster(Merovingian2.java:88)
> at 
> com.cloud.server.LockMasterListener.(LockMasterListener.java:33)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
> Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> at 
> org.springframework.beans.BeanUtils.instantiateClass(BeanUtils.java:148)
> at 
> org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:121)
> at 
> org.springframework.beans.factory.support.ConstructorResolver.autowireConstructor(ConstructorResolver.java:277)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.autowireConstructor(AbstractAutowireCapableBeanFactory.java:1077)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:981)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:487)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:458)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory$1.getObject(AbstractBeanFactory.java:293)
> at 
> org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:223)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:290)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:191)
> at 
> org.springframework.beans.factory.support.DefaultListableBeanFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:636)
> at 
> 

[jira] [Commented] (CLOUDSTACK-9300) MySQL HA feature StaticStrategy throws exception

2016-03-04 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9300?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15179853#comment-15179853
 ] 

Simon Weller commented on CLOUDSTACK-9300:
--

Pull request available here: https://github.com/myENA/cloudstack/pull/1

We'll build rpms and test.

> MySQL HA feature StaticStrategy throws exception
> 
>
> Key: CLOUDSTACK-9300
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9300
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>Affects Versions: Future, 4.7.0, 4.8.0, 4.7.1
> Environment: Centos 7
>Reporter: Simon Weller
>Priority: Minor
>
> 2016-03-03 12:00:13,204 INFO  [c.c.u.d.T.Transaction] 
> (localhost-startStop-1:null) (logid:) Is Data Base High Availiability 
> enabled? Ans : true
> 2016-03-03 12:00:13,239 INFO  [c.c.u.d.T.Transaction] 
> (localhost-startStop-1:null) (logid:) The slaves configured for Cloud Data 
> base is/are : localhost,localhost
> 2016-03-03 12:00:13,303 ERROR [c.c.u.d.Merovingian2] 
> (localhost-startStop-1:null) (logid:) Unable to get a new db connection
> java.sql.SQLException: Invalid load balancing strategy 
> 'com.cloud.utils.db.StaticStrategy'.
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:927)
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:924)
> at com.mysql.jdbc.Util.loadExtensions(Util.java:602)
> at 
> com.mysql.jdbc.LoadBalancingConnectionProxy.(LoadBalancingConnectionProxy.java:280)
> at 
> com.mysql.jdbc.FailoverConnectionProxy.(FailoverConnectionProxy.java:67)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connectFailover(NonRegisteringDriver.java:433)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:346)
> at java.sql.DriverManager.getConnection(DriverManager.java:571)
> at java.sql.DriverManager.getConnection(DriverManager.java:215)
> at 
> org.apache.commons.dbcp.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:75)
> at 
> org.apache.commons.dbcp.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:582)
> at 
> org.apache.commons.pool.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:1188)
> at 
> org.apache.commons.dbcp.PoolingDataSource.getConnection(PoolingDataSource.java:106)
> at 
> com.cloud.utils.db.TransactionLegacy.getStandaloneConnectionWithException(TransactionLegacy.java:202)
> at com.cloud.utils.db.Merovingian2.(Merovingian2.java:68)
> at 
> com.cloud.utils.db.Merovingian2.createLockMaster(Merovingian2.java:88)
> at 
> com.cloud.server.LockMasterListener.(LockMasterListener.java:33)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
> Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> at 
> org.springframework.beans.BeanUtils.instantiateClass(BeanUtils.java:148)
> at 
> org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:121)
> at 
> org.springframework.beans.factory.support.ConstructorResolver.autowireConstructor(ConstructorResolver.java:277)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.autowireConstructor(AbstractAutowireCapableBeanFactory.java:1077)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:981)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:487)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:458)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory$1.getObject(AbstractBeanFactory.java:293)
> at 
> org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:223)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:290)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:191)
> at 
> org.springframework.beans.factory.support.DefaultListableBeanFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:636)
> at 
> 

[jira] [Commented] (CLOUDSTACK-9285) Cloudstack 4.8 can't connect to XEN and KVM hosts

2016-03-03 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15178852#comment-15178852
 ] 

Simon Weller commented on CLOUDSTACK-9285:
--

We have a fix for this.

We'll post a pull request tomorrow.

> Cloudstack 4.8 can't connect to XEN and KVM hosts
> -
>
> Key: CLOUDSTACK-9285
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9285
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: KVM, Xen
>Affects Versions: 4.8.0
> Environment: CentOS 7
>Reporter: Jens Fettig
>Priority: Blocker
> Attachments: agentfailure.pcap
>
>
> We used Cloudstack 4.7 in our testing environment. For future uses we updated 
> to Cloudstack 4.8 for some tests. But after the update from 4.7 to 4.8 
> cloudstack has some problems to connect to the hosts again. Here are the 
> error logs.
> {code:title=XEN Server from the managment-server.log:|borderStyle=solid}
> 2016-02-15 11:22:03,476 DEBUG [c.c.a.m.ClusteredAgentManagerImpl] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec) Loading directly 
> connected host 13(xen5)
> 2016-02-15 11:22:06,204 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-1:ctx-aaf933f3) (logid:d157e37d) AutoScaling Monitor is 
> running...
> 2016-02-15 11:22:06,206 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-2:ctx-8e2fc084) (logid:379074cb) VmStatsCollector is 
> running...
> 2016-02-15 11:22:06,217 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-4:ctx-cc1cf960) (logid:b6397e2e) HostStatsCollector is 
> running...
> 2016-02-15 11:22:06,298 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-3:ctx-efda9c07) (logid:5731423e) StorageCollector is 
> running...
> 2016-02-15 11:22:06,302 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-3:ctx-efda9c07) (logid:5731423e) There is no secondary 
> storage VM for secondary storage host cdsdev-secondary
> 2016-02-15 11:22:11,018 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl] 
> (AsyncJobMgr-Heartbeat-1:ctx-bc740746) (logid:cc619bad) Begin cleanup expired 
> async-jobs
> 2016-02-15 11:22:11,024 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl] 
> (AsyncJobMgr-Heartbeat-1:ctx-bc740746) (logid:cc619bad) End cleanup expired 
> async-jobs
> 2016-02-15 11:22:11,055 DEBUG [c.c.h.x.r.XenServerConnectionPool] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec) Unable to create 
> master connection to host(192.168.0.97) , due to The credentials given by the 
> user are incorrect, so access has been denied, and you have not been issued a 
> session handle.
> 2016-02-15 11:22:11,055 DEBUG [c.c.h.Status] (ClusteredAgentManager 
> Timer:ctx-43ef0a0a) (logid:82d7b1ec) Transition:[Resource state = Enabled, 
> Agent event = AgentDisconnected, Host id = 13, name = xen5]
> 2016-02-15 11:22:11,101 WARN  [c.c.a.m.ClusteredAgentManagerImpl] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec)  can not load 
> directly connected host 13(xen5) due to 
> com.cloud.utils.exception.CloudRuntimeException: Unable to create master 
> connection to host(192.168.0.97) , due to The credentials given by the user 
> are incorrect, so access has been denied, and you have not been issued a 
> session handle.
>   at 
> com.cloud.hypervisor.xenserver.resource.XenServerConnectionPool.getConnect(XenServerConnectionPool.java:163)
>   at 
> com.cloud.hypervisor.xenserver.resource.CitrixResourceBase.CheckXenHostInfo(CitrixResourceBase.java:523)
>   at 
> com.cloud.hypervisor.xenserver.resource.CitrixResourceBase.configure(CitrixResourceBase.java:827)
>   at 
> com.cloud.resource.DiscovererBase.reloadResource(DiscovererBase.java:158)
>   at 
> com.cloud.agent.manager.AgentManagerImpl.loadDirectlyConnectedHost(AgentManagerImpl.java:697)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.scanDirectAgentToLoad(ClusteredAgentManagerImpl.java:217)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.runDirectAgentScanTimerTask(ClusteredAgentManagerImpl.java:182)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.access$100(ClusteredAgentManagerImpl.java:96)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl$DirectAgentScanTimerTask.runInContext(ClusteredAgentManagerImpl.java:233)
>   at 
> org.apache.cloudstack.managed.context.ManagedContextTimerTask$1.runInContext(ManagedContextTimerTask.java:30)
>   at 
> org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49)
>   at 
> org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:56)
>   at 
> org.apache.cloudstack.managed.context.impl.DefaultManagedContext.callWithContext(DefaultManagedContext.java:103)
>   at 
> 

[jira] [Commented] (CLOUDSTACK-9285) Cloudstack 4.8 can't connect to XEN and KVM hosts

2016-03-03 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15178806#comment-15178806
 ] 

Simon Weller commented on CLOUDSTACK-9285:
--

Offending code:

 _connection = new NioClient("Agent", _shell.getHost(), _shell.getPort(), 
_shell.getWorkers(), this);
do {
s_logger.info("Reconnecting...");
try {
_connection.start();
} catch (final NioConnectionException e) {
throw new CloudRuntimeException("Unable to start the 
connection!", e); // <-- seems odd. This will probably kill the agent, but 
leave Tomcat running
}
_shell.getBackoffAlgorithm().waitBeforeRetry();
} while (!_connection.isStartup());
s_logger.info("Connected to the server");
}


> Cloudstack 4.8 can't connect to XEN and KVM hosts
> -
>
> Key: CLOUDSTACK-9285
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9285
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: KVM, Xen
>Affects Versions: 4.8.0
> Environment: CentOS 7
>Reporter: Jens Fettig
>Priority: Blocker
> Attachments: agentfailure.pcap
>
>
> We used Cloudstack 4.7 in our testing environment. For future uses we updated 
> to Cloudstack 4.8 for some tests. But after the update from 4.7 to 4.8 
> cloudstack has some problems to connect to the hosts again. Here are the 
> error logs.
> {code:title=XEN Server from the managment-server.log:|borderStyle=solid}
> 2016-02-15 11:22:03,476 DEBUG [c.c.a.m.ClusteredAgentManagerImpl] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec) Loading directly 
> connected host 13(xen5)
> 2016-02-15 11:22:06,204 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-1:ctx-aaf933f3) (logid:d157e37d) AutoScaling Monitor is 
> running...
> 2016-02-15 11:22:06,206 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-2:ctx-8e2fc084) (logid:379074cb) VmStatsCollector is 
> running...
> 2016-02-15 11:22:06,217 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-4:ctx-cc1cf960) (logid:b6397e2e) HostStatsCollector is 
> running...
> 2016-02-15 11:22:06,298 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-3:ctx-efda9c07) (logid:5731423e) StorageCollector is 
> running...
> 2016-02-15 11:22:06,302 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-3:ctx-efda9c07) (logid:5731423e) There is no secondary 
> storage VM for secondary storage host cdsdev-secondary
> 2016-02-15 11:22:11,018 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl] 
> (AsyncJobMgr-Heartbeat-1:ctx-bc740746) (logid:cc619bad) Begin cleanup expired 
> async-jobs
> 2016-02-15 11:22:11,024 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl] 
> (AsyncJobMgr-Heartbeat-1:ctx-bc740746) (logid:cc619bad) End cleanup expired 
> async-jobs
> 2016-02-15 11:22:11,055 DEBUG [c.c.h.x.r.XenServerConnectionPool] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec) Unable to create 
> master connection to host(192.168.0.97) , due to The credentials given by the 
> user are incorrect, so access has been denied, and you have not been issued a 
> session handle.
> 2016-02-15 11:22:11,055 DEBUG [c.c.h.Status] (ClusteredAgentManager 
> Timer:ctx-43ef0a0a) (logid:82d7b1ec) Transition:[Resource state = Enabled, 
> Agent event = AgentDisconnected, Host id = 13, name = xen5]
> 2016-02-15 11:22:11,101 WARN  [c.c.a.m.ClusteredAgentManagerImpl] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec)  can not load 
> directly connected host 13(xen5) due to 
> com.cloud.utils.exception.CloudRuntimeException: Unable to create master 
> connection to host(192.168.0.97) , due to The credentials given by the user 
> are incorrect, so access has been denied, and you have not been issued a 
> session handle.
>   at 
> com.cloud.hypervisor.xenserver.resource.XenServerConnectionPool.getConnect(XenServerConnectionPool.java:163)
>   at 
> com.cloud.hypervisor.xenserver.resource.CitrixResourceBase.CheckXenHostInfo(CitrixResourceBase.java:523)
>   at 
> com.cloud.hypervisor.xenserver.resource.CitrixResourceBase.configure(CitrixResourceBase.java:827)
>   at 
> com.cloud.resource.DiscovererBase.reloadResource(DiscovererBase.java:158)
>   at 
> com.cloud.agent.manager.AgentManagerImpl.loadDirectlyConnectedHost(AgentManagerImpl.java:697)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.scanDirectAgentToLoad(ClusteredAgentManagerImpl.java:217)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.runDirectAgentScanTimerTask(ClusteredAgentManagerImpl.java:182)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.access$100(ClusteredAgentManagerImpl.java:96)
>   at 
> 

[jira] [Commented] (CLOUDSTACK-9285) Cloudstack 4.8 can't connect to XEN and KVM hosts

2016-03-03 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15178788#comment-15178788
 ] 

Simon Weller commented on CLOUDSTACK-9285:
--

The agent logic seems to be built around expecting a connection refused, where 
here (in our case) a RST is being sent back.

> Cloudstack 4.8 can't connect to XEN and KVM hosts
> -
>
> Key: CLOUDSTACK-9285
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9285
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: KVM, Xen
>Affects Versions: 4.8.0
> Environment: CentOS 7
>Reporter: Jens Fettig
>Priority: Blocker
> Attachments: agentfailure.pcap
>
>
> We used Cloudstack 4.7 in our testing environment. For future uses we updated 
> to Cloudstack 4.8 for some tests. But after the update from 4.7 to 4.8 
> cloudstack has some problems to connect to the hosts again. Here are the 
> error logs.
> {code:title=XEN Server from the managment-server.log:|borderStyle=solid}
> 2016-02-15 11:22:03,476 DEBUG [c.c.a.m.ClusteredAgentManagerImpl] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec) Loading directly 
> connected host 13(xen5)
> 2016-02-15 11:22:06,204 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-1:ctx-aaf933f3) (logid:d157e37d) AutoScaling Monitor is 
> running...
> 2016-02-15 11:22:06,206 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-2:ctx-8e2fc084) (logid:379074cb) VmStatsCollector is 
> running...
> 2016-02-15 11:22:06,217 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-4:ctx-cc1cf960) (logid:b6397e2e) HostStatsCollector is 
> running...
> 2016-02-15 11:22:06,298 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-3:ctx-efda9c07) (logid:5731423e) StorageCollector is 
> running...
> 2016-02-15 11:22:06,302 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-3:ctx-efda9c07) (logid:5731423e) There is no secondary 
> storage VM for secondary storage host cdsdev-secondary
> 2016-02-15 11:22:11,018 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl] 
> (AsyncJobMgr-Heartbeat-1:ctx-bc740746) (logid:cc619bad) Begin cleanup expired 
> async-jobs
> 2016-02-15 11:22:11,024 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl] 
> (AsyncJobMgr-Heartbeat-1:ctx-bc740746) (logid:cc619bad) End cleanup expired 
> async-jobs
> 2016-02-15 11:22:11,055 DEBUG [c.c.h.x.r.XenServerConnectionPool] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec) Unable to create 
> master connection to host(192.168.0.97) , due to The credentials given by the 
> user are incorrect, so access has been denied, and you have not been issued a 
> session handle.
> 2016-02-15 11:22:11,055 DEBUG [c.c.h.Status] (ClusteredAgentManager 
> Timer:ctx-43ef0a0a) (logid:82d7b1ec) Transition:[Resource state = Enabled, 
> Agent event = AgentDisconnected, Host id = 13, name = xen5]
> 2016-02-15 11:22:11,101 WARN  [c.c.a.m.ClusteredAgentManagerImpl] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec)  can not load 
> directly connected host 13(xen5) due to 
> com.cloud.utils.exception.CloudRuntimeException: Unable to create master 
> connection to host(192.168.0.97) , due to The credentials given by the user 
> are incorrect, so access has been denied, and you have not been issued a 
> session handle.
>   at 
> com.cloud.hypervisor.xenserver.resource.XenServerConnectionPool.getConnect(XenServerConnectionPool.java:163)
>   at 
> com.cloud.hypervisor.xenserver.resource.CitrixResourceBase.CheckXenHostInfo(CitrixResourceBase.java:523)
>   at 
> com.cloud.hypervisor.xenserver.resource.CitrixResourceBase.configure(CitrixResourceBase.java:827)
>   at 
> com.cloud.resource.DiscovererBase.reloadResource(DiscovererBase.java:158)
>   at 
> com.cloud.agent.manager.AgentManagerImpl.loadDirectlyConnectedHost(AgentManagerImpl.java:697)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.scanDirectAgentToLoad(ClusteredAgentManagerImpl.java:217)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.runDirectAgentScanTimerTask(ClusteredAgentManagerImpl.java:182)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.access$100(ClusteredAgentManagerImpl.java:96)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl$DirectAgentScanTimerTask.runInContext(ClusteredAgentManagerImpl.java:233)
>   at 
> org.apache.cloudstack.managed.context.ManagedContextTimerTask$1.runInContext(ManagedContextTimerTask.java:30)
>   at 
> org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49)
>   at 
> org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:56)
>   at 
> 

[jira] [Updated] (CLOUDSTACK-9285) Cloudstack 4.8 can't connect to XEN and KVM hosts

2016-03-03 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9285?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller updated CLOUDSTACK-9285:
-
Attachment: agentfailure.pcap

Here's a pcap of the issue. haproxy is sending a RST

> Cloudstack 4.8 can't connect to XEN and KVM hosts
> -
>
> Key: CLOUDSTACK-9285
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9285
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: KVM, Xen
>Affects Versions: 4.8.0
> Environment: CentOS 7
>Reporter: Jens Fettig
>Priority: Blocker
> Attachments: agentfailure.pcap
>
>
> We used Cloudstack 4.7 in our testing environment. For future uses we updated 
> to Cloudstack 4.8 for some tests. But after the update from 4.7 to 4.8 
> cloudstack has some problems to connect to the hosts again. Here are the 
> error logs.
> {code:title=XEN Server from the managment-server.log:|borderStyle=solid}
> 2016-02-15 11:22:03,476 DEBUG [c.c.a.m.ClusteredAgentManagerImpl] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec) Loading directly 
> connected host 13(xen5)
> 2016-02-15 11:22:06,204 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-1:ctx-aaf933f3) (logid:d157e37d) AutoScaling Monitor is 
> running...
> 2016-02-15 11:22:06,206 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-2:ctx-8e2fc084) (logid:379074cb) VmStatsCollector is 
> running...
> 2016-02-15 11:22:06,217 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-4:ctx-cc1cf960) (logid:b6397e2e) HostStatsCollector is 
> running...
> 2016-02-15 11:22:06,298 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-3:ctx-efda9c07) (logid:5731423e) StorageCollector is 
> running...
> 2016-02-15 11:22:06,302 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-3:ctx-efda9c07) (logid:5731423e) There is no secondary 
> storage VM for secondary storage host cdsdev-secondary
> 2016-02-15 11:22:11,018 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl] 
> (AsyncJobMgr-Heartbeat-1:ctx-bc740746) (logid:cc619bad) Begin cleanup expired 
> async-jobs
> 2016-02-15 11:22:11,024 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl] 
> (AsyncJobMgr-Heartbeat-1:ctx-bc740746) (logid:cc619bad) End cleanup expired 
> async-jobs
> 2016-02-15 11:22:11,055 DEBUG [c.c.h.x.r.XenServerConnectionPool] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec) Unable to create 
> master connection to host(192.168.0.97) , due to The credentials given by the 
> user are incorrect, so access has been denied, and you have not been issued a 
> session handle.
> 2016-02-15 11:22:11,055 DEBUG [c.c.h.Status] (ClusteredAgentManager 
> Timer:ctx-43ef0a0a) (logid:82d7b1ec) Transition:[Resource state = Enabled, 
> Agent event = AgentDisconnected, Host id = 13, name = xen5]
> 2016-02-15 11:22:11,101 WARN  [c.c.a.m.ClusteredAgentManagerImpl] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec)  can not load 
> directly connected host 13(xen5) due to 
> com.cloud.utils.exception.CloudRuntimeException: Unable to create master 
> connection to host(192.168.0.97) , due to The credentials given by the user 
> are incorrect, so access has been denied, and you have not been issued a 
> session handle.
>   at 
> com.cloud.hypervisor.xenserver.resource.XenServerConnectionPool.getConnect(XenServerConnectionPool.java:163)
>   at 
> com.cloud.hypervisor.xenserver.resource.CitrixResourceBase.CheckXenHostInfo(CitrixResourceBase.java:523)
>   at 
> com.cloud.hypervisor.xenserver.resource.CitrixResourceBase.configure(CitrixResourceBase.java:827)
>   at 
> com.cloud.resource.DiscovererBase.reloadResource(DiscovererBase.java:158)
>   at 
> com.cloud.agent.manager.AgentManagerImpl.loadDirectlyConnectedHost(AgentManagerImpl.java:697)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.scanDirectAgentToLoad(ClusteredAgentManagerImpl.java:217)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.runDirectAgentScanTimerTask(ClusteredAgentManagerImpl.java:182)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.access$100(ClusteredAgentManagerImpl.java:96)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl$DirectAgentScanTimerTask.runInContext(ClusteredAgentManagerImpl.java:233)
>   at 
> org.apache.cloudstack.managed.context.ManagedContextTimerTask$1.runInContext(ManagedContextTimerTask.java:30)
>   at 
> org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49)
>   at 
> org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:56)
>   at 
> org.apache.cloudstack.managed.context.impl.DefaultManagedContext.callWithContext(DefaultManagedContext.java:103)
>   at 
> 

[jira] [Commented] (CLOUDSTACK-9285) Cloudstack 4.8 can't connect to XEN and KVM hosts

2016-03-03 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15178769#comment-15178769
 ] 

Simon Weller commented on CLOUDSTACK-9285:
--

Are you using a load balancer between your management server and the host agent?
We are seeing this with haproxy between our agents and management server. If we 
down all the management servers, the agents connection is cut to the load 
balancer and the agent throws a  "Could not find exception: 
com.cloud.utils.exception.NioConnectionException in error code list for 
exceptions."

We're digging into this right now.



> Cloudstack 4.8 can't connect to XEN and KVM hosts
> -
>
> Key: CLOUDSTACK-9285
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9285
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: KVM, Xen
>Affects Versions: 4.8.0
> Environment: CentOS 7
>Reporter: Jens Fettig
>Priority: Blocker
>
> We used Cloudstack 4.7 in our testing environment. For future uses we updated 
> to Cloudstack 4.8 for some tests. But after the update from 4.7 to 4.8 
> cloudstack has some problems to connect to the hosts again. Here are the 
> error logs.
> {code:title=XEN Server from the managment-server.log:|borderStyle=solid}
> 2016-02-15 11:22:03,476 DEBUG [c.c.a.m.ClusteredAgentManagerImpl] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec) Loading directly 
> connected host 13(xen5)
> 2016-02-15 11:22:06,204 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-1:ctx-aaf933f3) (logid:d157e37d) AutoScaling Monitor is 
> running...
> 2016-02-15 11:22:06,206 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-2:ctx-8e2fc084) (logid:379074cb) VmStatsCollector is 
> running...
> 2016-02-15 11:22:06,217 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-4:ctx-cc1cf960) (logid:b6397e2e) HostStatsCollector is 
> running...
> 2016-02-15 11:22:06,298 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-3:ctx-efda9c07) (logid:5731423e) StorageCollector is 
> running...
> 2016-02-15 11:22:06,302 DEBUG [c.c.s.StatsCollector] 
> (StatsCollector-3:ctx-efda9c07) (logid:5731423e) There is no secondary 
> storage VM for secondary storage host cdsdev-secondary
> 2016-02-15 11:22:11,018 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl] 
> (AsyncJobMgr-Heartbeat-1:ctx-bc740746) (logid:cc619bad) Begin cleanup expired 
> async-jobs
> 2016-02-15 11:22:11,024 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl] 
> (AsyncJobMgr-Heartbeat-1:ctx-bc740746) (logid:cc619bad) End cleanup expired 
> async-jobs
> 2016-02-15 11:22:11,055 DEBUG [c.c.h.x.r.XenServerConnectionPool] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec) Unable to create 
> master connection to host(192.168.0.97) , due to The credentials given by the 
> user are incorrect, so access has been denied, and you have not been issued a 
> session handle.
> 2016-02-15 11:22:11,055 DEBUG [c.c.h.Status] (ClusteredAgentManager 
> Timer:ctx-43ef0a0a) (logid:82d7b1ec) Transition:[Resource state = Enabled, 
> Agent event = AgentDisconnected, Host id = 13, name = xen5]
> 2016-02-15 11:22:11,101 WARN  [c.c.a.m.ClusteredAgentManagerImpl] 
> (ClusteredAgentManager Timer:ctx-43ef0a0a) (logid:82d7b1ec)  can not load 
> directly connected host 13(xen5) due to 
> com.cloud.utils.exception.CloudRuntimeException: Unable to create master 
> connection to host(192.168.0.97) , due to The credentials given by the user 
> are incorrect, so access has been denied, and you have not been issued a 
> session handle.
>   at 
> com.cloud.hypervisor.xenserver.resource.XenServerConnectionPool.getConnect(XenServerConnectionPool.java:163)
>   at 
> com.cloud.hypervisor.xenserver.resource.CitrixResourceBase.CheckXenHostInfo(CitrixResourceBase.java:523)
>   at 
> com.cloud.hypervisor.xenserver.resource.CitrixResourceBase.configure(CitrixResourceBase.java:827)
>   at 
> com.cloud.resource.DiscovererBase.reloadResource(DiscovererBase.java:158)
>   at 
> com.cloud.agent.manager.AgentManagerImpl.loadDirectlyConnectedHost(AgentManagerImpl.java:697)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.scanDirectAgentToLoad(ClusteredAgentManagerImpl.java:217)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.runDirectAgentScanTimerTask(ClusteredAgentManagerImpl.java:182)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl.access$100(ClusteredAgentManagerImpl.java:96)
>   at 
> com.cloud.agent.manager.ClusteredAgentManagerImpl$DirectAgentScanTimerTask.runInContext(ClusteredAgentManagerImpl.java:233)
>   at 
> org.apache.cloudstack.managed.context.ManagedContextTimerTask$1.runInContext(ManagedContextTimerTask.java:30)
>   at 
> 

[jira] [Commented] (CLOUDSTACK-9300) MySQL HA feature StaticStrategy throws exception

2016-03-03 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9300?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15178748#comment-15178748
 ] 

Simon Weller commented on CLOUDSTACK-9300:
--

This appears to be a 2 part issue.

Got tomcat to dump the class loader info into journald and got this...  
WARNING: Problem with directory [/usr/share/cloudstack-mysql-ha/lib/*jar], 
exists: [false], isDirectory: [false], canRead

I changed ​*jar to *​.jar in the catalina.properities common.loader and that 
issue was no longer seen in the class loader logs..
After further debugging, I found that having mysql-connector-java.jar 
referenced in the global class path in /etc/sysconfig/cloudstack-management 
prevents the inherited StaticStrategy class from loading.
If you remove that and change to 
CLASSPATH=/etc/cloudstack/management:/usr/share/cloudstack-common:/usr/share/cloudstack-management/setup
 StaticStrategy loads
It was already referenced in the catalina.properities common.loader
I've tested that fix now on master and on 4.7.1. 

We'll complete testing after building new master rpms and then we'll submit a 
pull request on this.

> MySQL HA feature StaticStrategy throws exception
> 
>
> Key: CLOUDSTACK-9300
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9300
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>Affects Versions: Future, 4.7.0, 4.8.0, 4.7.1
> Environment: Centos 7
>Reporter: Simon Weller
>Priority: Minor
>
> 2016-03-03 12:00:13,204 INFO  [c.c.u.d.T.Transaction] 
> (localhost-startStop-1:null) (logid:) Is Data Base High Availiability 
> enabled? Ans : true
> 2016-03-03 12:00:13,239 INFO  [c.c.u.d.T.Transaction] 
> (localhost-startStop-1:null) (logid:) The slaves configured for Cloud Data 
> base is/are : localhost,localhost
> 2016-03-03 12:00:13,303 ERROR [c.c.u.d.Merovingian2] 
> (localhost-startStop-1:null) (logid:) Unable to get a new db connection
> java.sql.SQLException: Invalid load balancing strategy 
> 'com.cloud.utils.db.StaticStrategy'.
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:927)
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:924)
> at com.mysql.jdbc.Util.loadExtensions(Util.java:602)
> at 
> com.mysql.jdbc.LoadBalancingConnectionProxy.(LoadBalancingConnectionProxy.java:280)
> at 
> com.mysql.jdbc.FailoverConnectionProxy.(FailoverConnectionProxy.java:67)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connectFailover(NonRegisteringDriver.java:433)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:346)
> at java.sql.DriverManager.getConnection(DriverManager.java:571)
> at java.sql.DriverManager.getConnection(DriverManager.java:215)
> at 
> org.apache.commons.dbcp.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:75)
> at 
> org.apache.commons.dbcp.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:582)
> at 
> org.apache.commons.pool.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:1188)
> at 
> org.apache.commons.dbcp.PoolingDataSource.getConnection(PoolingDataSource.java:106)
> at 
> com.cloud.utils.db.TransactionLegacy.getStandaloneConnectionWithException(TransactionLegacy.java:202)
> at com.cloud.utils.db.Merovingian2.(Merovingian2.java:68)
> at 
> com.cloud.utils.db.Merovingian2.createLockMaster(Merovingian2.java:88)
> at 
> com.cloud.server.LockMasterListener.(LockMasterListener.java:33)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
> Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> at 
> org.springframework.beans.BeanUtils.instantiateClass(BeanUtils.java:148)
> at 
> org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:121)
> at 
> org.springframework.beans.factory.support.ConstructorResolver.autowireConstructor(ConstructorResolver.java:277)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.autowireConstructor(AbstractAutowireCapableBeanFactory.java:1077)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:981)
> at 
> 

[jira] [Created] (CLOUDSTACK-9300) MySQL HA feature StaticStrategy throws exception

2016-03-03 Thread Simon Weller (JIRA)
Simon Weller created CLOUDSTACK-9300:


 Summary: MySQL HA feature StaticStrategy throws exception
 Key: CLOUDSTACK-9300
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9300
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public (Anyone can view this level - this is the default.)
Affects Versions: Future, 4.7.0, 4.8.0, 4.7.1
 Environment: Centos 7
Reporter: Simon Weller
Priority: Minor


2016-03-03 12:00:13,204 INFO  [c.c.u.d.T.Transaction] 
(localhost-startStop-1:null) (logid:) Is Data Base High Availiability enabled? 
Ans : true
2016-03-03 12:00:13,239 INFO  [c.c.u.d.T.Transaction] 
(localhost-startStop-1:null) (logid:) The slaves configured for Cloud Data base 
is/are : localhost,localhost
2016-03-03 12:00:13,303 ERROR [c.c.u.d.Merovingian2] 
(localhost-startStop-1:null) (logid:) Unable to get a new db connection
java.sql.SQLException: Invalid load balancing strategy 
'com.cloud.utils.db.StaticStrategy'.
at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:927)
at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:924)
at com.mysql.jdbc.Util.loadExtensions(Util.java:602)
at 
com.mysql.jdbc.LoadBalancingConnectionProxy.(LoadBalancingConnectionProxy.java:280)
at 
com.mysql.jdbc.FailoverConnectionProxy.(FailoverConnectionProxy.java:67)
at 
com.mysql.jdbc.NonRegisteringDriver.connectFailover(NonRegisteringDriver.java:433)
at 
com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:346)
at java.sql.DriverManager.getConnection(DriverManager.java:571)
at java.sql.DriverManager.getConnection(DriverManager.java:215)
at 
org.apache.commons.dbcp.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:75)
at 
org.apache.commons.dbcp.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:582)
at 
org.apache.commons.pool.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:1188)
at 
org.apache.commons.dbcp.PoolingDataSource.getConnection(PoolingDataSource.java:106)
at 
com.cloud.utils.db.TransactionLegacy.getStandaloneConnectionWithException(TransactionLegacy.java:202)
at com.cloud.utils.db.Merovingian2.(Merovingian2.java:68)
at 
com.cloud.utils.db.Merovingian2.createLockMaster(Merovingian2.java:88)
at 
com.cloud.server.LockMasterListener.(LockMasterListener.java:33)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at 
org.springframework.beans.BeanUtils.instantiateClass(BeanUtils.java:148)
at 
org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:121)
at 
org.springframework.beans.factory.support.ConstructorResolver.autowireConstructor(ConstructorResolver.java:277)
at 
org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.autowireConstructor(AbstractAutowireCapableBeanFactory.java:1077)
at 
org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:981)
at 
org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:487)
at 
org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:458)
at 
org.springframework.beans.factory.support.AbstractBeanFactory$1.getObject(AbstractBeanFactory.java:293)
at 
org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:223)
at 
org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:290)
at 
org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:191)
at 
org.springframework.beans.factory.support.DefaultListableBeanFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:636)
at 
org.springframework.context.support.AbstractApplicationContext.finishBeanFactoryInitialization(AbstractApplicationContext.java:934)
at 
org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:479)
at 
org.apache.cloudstack.spring.module.model.impl.DefaultModuleDefinitionSet.loadContext(DefaultModuleDefinitionSet.java:145)
at 

[jira] [Commented] (CLOUDSTACK-9209) Database HA not working on a fresh install on Ubuntu 12.04

2016-02-19 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9209?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15154586#comment-15154586
 ] 

Simon Weller commented on CLOUDSTACK-9209:
--

Have you got the MySQL ha jar working on Centos 7, with 4.7.x?

> Database HA not working on a fresh install on Ubuntu 12.04
> --
>
> Key: CLOUDSTACK-9209
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9209
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: Install and Setup
>Affects Versions: 4.7.0
> Environment: Two Ubuntu 12.04 Management Servers, Mysql Master Master 
> Replication, Official community repository: 
> http://cloudstack.apt-get.eu/ubuntu/dists/precise/4.7/
>Reporter: Thomas
>Priority: Blocker
>
> The cloudstack-management service will not come up because of this error:
> 2016-01-05 16:15:00,899 INFO  [c.c.u.d.T.Transaction] (main:null) (logid:) Is 
> Data Base High Availiability enabled? Ans : true
> 2016-01-05 16:15:00,924 INFO  [c.c.u.d.T.Transaction] (main:null) (logid:) 
> The slaves configured for Cloud Data base is/are : 192.168.0.90
> 2016-01-05 16:15:00,974 ERROR [c.c.u.d.Merovingian2] (main:null) (logid:) 
> Unable to get a new db connection
> java.sql.SQLException: Invalid load balancing strategy 
> 'com.cloud.utils.db.StaticStrategy'.
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:934)
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:931)
> at com.mysql.jdbc.Util.loadExtensions(Util.java:588)
> at 
> com.mysql.jdbc.LoadBalancingConnectionProxy.(LoadBalancingConnectionProxy.java:279)
> at 
> com.mysql.jdbc.FailoverConnectionProxy.(FailoverConnectionProxy.java:66)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connectFailover(NonRegisteringDriver.java:382)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:301)
> at java.sql.DriverManager.getConnection(DriverManager.java:571)
> at java.sql.DriverManager.getConnection(DriverManager.java:215)
> at 
> org.apache.commons.dbcp.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:75)
> at 
> org.apache.commons.dbcp.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:582)
> at 
> org.apache.commons.pool.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:1188)
> at 
> org.apache.commons.dbcp.PoolingDataSource.getConnection(PoolingDataSource.java:106)
> at 
> com.cloud.utils.db.TransactionLegacy.getStandaloneConnectionWithException(TransactionLegacy.java:202)
> at com.cloud.utils.db.Merovingian2.(Merovingian2.java:68)
> at 
> com.cloud.utils.db.Merovingian2.createLockMaster(Merovingian2.java:88)
> at 
> com.cloud.server.LockMasterListener.(LockMasterListener.java:33)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
> Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> at 
> org.springframework.beans.BeanUtils.instantiateClass(BeanUtils.java:148)
> at 
> org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:121)
> at 
> org.springframework.beans.factory.support.ConstructorResolver.autowireConstructor(ConstructorResolver.java:277)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.autowireConstructor(AbstractAutowireCapableBeanFactory.java:1077)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:981)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:487)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:458)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory$1.getObject(AbstractBeanFactory.java:293)
> at 
> org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:223)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:290)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:191)
> at 
> 

[jira] [Commented] (CLOUDSTACK-9209) Database HA not working on a fresh install on Ubuntu 12.04

2016-02-18 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-9209?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152490#comment-15152490
 ] 

Simon Weller commented on CLOUDSTACK-9209:
--

Thomas,

Do you have the log4j ubuntu package installed? 

- Si

> Database HA not working on a fresh install on Ubuntu 12.04
> --
>
> Key: CLOUDSTACK-9209
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-9209
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
>  Components: Install and Setup
>Affects Versions: 4.7.0
> Environment: Two Ubuntu 12.04 Management Servers, Mysql Master Master 
> Replication, Official community repository: 
> http://cloudstack.apt-get.eu/ubuntu/dists/precise/4.7/
>Reporter: Thomas
>Priority: Blocker
>
> The cloudstack-management service will not come up because of this error:
> 2016-01-05 16:15:00,899 INFO  [c.c.u.d.T.Transaction] (main:null) (logid:) Is 
> Data Base High Availiability enabled? Ans : true
> 2016-01-05 16:15:00,924 INFO  [c.c.u.d.T.Transaction] (main:null) (logid:) 
> The slaves configured for Cloud Data base is/are : 192.168.0.90
> 2016-01-05 16:15:00,974 ERROR [c.c.u.d.Merovingian2] (main:null) (logid:) 
> Unable to get a new db connection
> java.sql.SQLException: Invalid load balancing strategy 
> 'com.cloud.utils.db.StaticStrategy'.
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:934)
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:931)
> at com.mysql.jdbc.Util.loadExtensions(Util.java:588)
> at 
> com.mysql.jdbc.LoadBalancingConnectionProxy.(LoadBalancingConnectionProxy.java:279)
> at 
> com.mysql.jdbc.FailoverConnectionProxy.(FailoverConnectionProxy.java:66)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connectFailover(NonRegisteringDriver.java:382)
> at 
> com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:301)
> at java.sql.DriverManager.getConnection(DriverManager.java:571)
> at java.sql.DriverManager.getConnection(DriverManager.java:215)
> at 
> org.apache.commons.dbcp.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:75)
> at 
> org.apache.commons.dbcp.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:582)
> at 
> org.apache.commons.pool.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:1188)
> at 
> org.apache.commons.dbcp.PoolingDataSource.getConnection(PoolingDataSource.java:106)
> at 
> com.cloud.utils.db.TransactionLegacy.getStandaloneConnectionWithException(TransactionLegacy.java:202)
> at com.cloud.utils.db.Merovingian2.(Merovingian2.java:68)
> at 
> com.cloud.utils.db.Merovingian2.createLockMaster(Merovingian2.java:88)
> at 
> com.cloud.server.LockMasterListener.(LockMasterListener.java:33)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
> Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> at 
> org.springframework.beans.BeanUtils.instantiateClass(BeanUtils.java:148)
> at 
> org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:121)
> at 
> org.springframework.beans.factory.support.ConstructorResolver.autowireConstructor(ConstructorResolver.java:277)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.autowireConstructor(AbstractAutowireCapableBeanFactory.java:1077)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:981)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:487)
> at 
> org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:458)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory$1.getObject(AbstractBeanFactory.java:293)
> at 
> org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:223)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:290)
> at 
> org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:191)
> at 
> 

[jira] [Commented] (CLOUDSTACK-8943) KVM HA is broken, let's fix it

2015-10-26 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-8943?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14974765#comment-14974765
 ] 

Simon Weller commented on CLOUDSTACK-8943:
--

So we've been looking around the code today and noticed that ipmi has already 
been implemented in the bare metal server plugin.

scripts/util/ipmi.py is a wrapper for /usr/bin/ipmitool and is called out by:

plugins/hypervisors/baremetal/src/com/cloud/baremetal/networkservice/BareMetalResourceBase.java
plugins/hypervisors/baremetal/src/com/cloud/baremetal/manager/BareMetalDiscoverer.java


> KVM HA is broken, let's fix it
> --
>
> Key: CLOUDSTACK-8943
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-8943
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
> Environment: Linux distros with KVM/libvirt
>Reporter: Nux
>
> Currently KVM HA works by monitoring an NFS based heartbeat file and it can 
> often fail whenever this network share becomes slower, causing the 
> hypervisors to reboot.
> This can be particularly annoying when you have different kinds of primary 
> storages in place which are working fine (people running CEPH etc).
> Having to wait for the affected HV which triggered this to come back and 
> declare it's not running VMs is a bad idea; this HV could require hours or 
> days of maintenance!
> This is embarrassing. How can we fix it? Ideas, suggestions? How are other 
> hypervisors doing it?
> Let's discuss, test, implement. :)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CLOUDSTACK-8943) KVM HA is broken, let's fix it

2015-10-13 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-8943?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14954834#comment-14954834
 ] 

Simon Weller commented on CLOUDSTACK-8943:
--

In my opinion, you have to be careful with quorum style functionality on the 
host itself. It can make the hosts overly complicated, especially when you have 
a lot of hosts within a cluster. 
If you've built your storage correctly, it should be highly unlikely that you 
lose your storage (unless you're running NFS with no HA functionality). And if 
you do, that becomes a cluster problem assuming you've dedicated storage just 
to that cluster. 
We've been running quorums using Redhat Cluster Suite for years on top of CLVM 
SAN backed storage and it has proven to be very difficult to manage as our 
infrastructure has grown. You really want to keep your hosts as simple as 
possible.
Now, if the CS agent was leveraged to provide a deeper view into what was going 
on with the host (and the attached storage), it may be you could gather enough 
information to make a determination of what should be done with a given host 
that appeared to be misbehaving. You could then fence it remotely using IPMI 
and that's what I was indicating above. That would simulate the principals of a 
quorum without having to add a complicated clustering layer onto the hosts.


> KVM HA is broken, let's fix it
> --
>
> Key: CLOUDSTACK-8943
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-8943
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
> Environment: Linux distros with KVM/libvirt
>Reporter: Nux
>
> Currently KVM HA works by monitoring an NFS based heartbeat file and it can 
> often fail whenever this network share becomes slower, causing the 
> hypervisors to reboot.
> This can be particularly annoying when you have different kinds of primary 
> storages in place which are working fine (people running CEPH etc).
> Having to wait for the affected HV which triggered this to come back and 
> declare it's not running VMs is a bad idea; this HV could require hours or 
> days of maintenance!
> This is embarrassing. How can we fix it? Ideas, suggestions? How are other 
> hypervisors doing it?
> Let's discuss, test, implement. :)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CLOUDSTACK-8943) KVM HA is broken, let's fix it

2015-10-12 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-8943?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14953017#comment-14953017
 ] 

Simon Weller commented on CLOUDSTACK-8943:
--

Perhaps one of the easiest ways to deal with this would be to introduce IPMI 
functionality into Cloudstack, so a KVM host could be fenced via an out-of-band 
IPMI interface. Upon successful fencing, CS MGMT could mark the host as 
disabled. I know deleting a host is enough to force CS MGMT to attempt to 
restart affected VMs on other hosts, but I'm not sure whether disabling a host 
will at this point in time.

There are other considerations that will need to be made as well, especially 
around storage locking (e.g. CEPH).

> KVM HA is broken, let's fix it
> --
>
> Key: CLOUDSTACK-8943
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-8943
> Project: CloudStack
>  Issue Type: Bug
>  Security Level: Public(Anyone can view this level - this is the 
> default.) 
> Environment: Linux distros with KVM/libvirt
>Reporter: Nux
>
> Currently KVM HA works by monitoring an NFS based heartbeat file and it can 
> often fail whenever this network share becomes slower, causing the 
> hypervisors to reboot.
> This can be particularly annoying when you have different kinds of primary 
> storages in place which are working fine (people running CEPH etc).
> Having to wait for the affected HV which triggered this to come back and 
> declare it's not running VMs is a bad idea; this HV could require hours or 
> days of maintenance!
> This is embarrassing. How can we fix it? Ideas, suggestions? How are other 
> hypervisors doing it?
> Let's discuss, test, implement. :)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-09-22 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14143170#comment-14143170
 ] 

Simon Weller commented on CLOUDSTACK-6460:
--

I added some debugging into getStoragePool, and createPhysicalDisk.

Agent logs in debug:

2014-09-22 07:03:50,164{GMT} DEBUG [cloud.agent.Agent] 
(agentRequest-Handler-4:) Request:Seq 3-192151576:  { Cmd , MgmtId: 
144343444026, via: 3, Ver: v1, Flags: 100011, 
[{org.apache.cloudstack.storage.command.CreateObjectCommand:{data:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:a8dbbcb0-5f27-470a-a482-e00ac0e68a75,volumeType:DATADISK,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:{uuid:076d1cd7-9c80-4302-8e13-ea8187a9a96b,id:201,poolType:CLVM,host:localhost,path:/csstore01,port:0,url:CLVM://localhost//csstore01/?ROLE=PrimarySTOREUUID=076d1cd7-9c80-4302-8e13-ea8187a9a96b}},name:test10_2,size:5368709120,volumeId:17461,accountId:6,id:17461,hypervisorType:KVM}},wait:0}}]
 }
2014-09-22 07:03:50,165{GMT} DEBUG [cloud.agent.Agent] 
(agentRequest-Handler-4:) Processing command: 
org.apache.cloudstack.storage.command.CreateObjectCommand
2014-09-22 07:03:50,172{GMT} DEBUG [kvm.storage.LibvirtStorageAdaptor] 
(agentRequest-Handler-4:) getStoragePool: logical
2014-09-22 07:03:52,044{GMT} DEBUG [kvm.storage.LibvirtStorageAdaptor] 
(agentRequest-Handler-4:) createPhysicalDisk format is raw
2014-09-22 07:03:52,045{GMT} DEBUG [kvm.storage.LibvirtStorageAdaptor] 
(agentRequest-Handler-4:) volume
namea8dbbcb0-5f27-470a-a482-e00ac0e68a75/name
capacity5368709120/capacity
target
format type='raw'/
permissionsmode0744/mode/permissions/target
/volume

Mgmt logs:

2014-09-22 07:03:50,133 DEBUG [c.c.s.StorageManagerImpl] 
(Job-Executor-7:ctx-e61f9a16 ctx-e997a276) Checking pool: 201 for volume 
allocation [Vol[17461|vm=null|DATADISK]], maxSize : 1099507433472, 
totalAllocatedSize : 83823165440, askingSize : 5368709120, allocated disable 
threshold: 0.85
2014-09-22 07:03:50,133 DEBUG [o.a.c.s.a.ClusterScopeStoragePoolAllocator] 
(Job-Executor-7:ctx-e61f9a16 ctx-e997a276) ClusterScopeStoragePoolAllocator 
returning 1 suitable storage pools
2014-09-22 07:03:50,133 DEBUG [o.a.c.e.o.VolumeOrchestrator] 
(Job-Executor-7:ctx-e61f9a16 ctx-e997a276) Trying to create 
org.apache.cloudstack.storage.volume.VolumeObject@76e38e5e on 
org.apache.cloudstack.storage.datastore.PrimaryDataStoreImpl@5597fa4c
2014-09-22 07:03:50,147 DEBUG 
[o.a.c.s.d.d.CloudStackPrimaryDataStoreDriverImpl] (Job-Executor-7:ctx-e61f9a16 
ctx-e997a276) Creating volume: 
org.apache.cloudstack.storage.volume.VolumeObject@53789cfc
2014-09-22 07:03:50,151 DEBUG [c.c.a.t.Request] (Job-Executor-7:ctx-e61f9a16 
ctx-e997a276) Seq 3-192151576: Sending  { Cmd , MgmtId: 144343444026, via: 
3(csh01-lab.zone01.nsvltn.cloud.ena.net), Ver: v1, Flags: 100011, 
[{org.apache.cloudstack.storage.command.CreateObjectCommand:{data:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:a8dbbcb0-5f27-470a-a482-e00ac0e68a75,volumeType:DATADISK,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:{uuid:076d1cd7-9c80-4302-8e13-ea8187a9a96b,id:201,poolType:CLVM,host:localhost,path:/csstore01,port:0,url:CLVM://localhost//csstore01/?ROLE=PrimarySTOREUUID=076d1cd7-9c80-4302-8e13-ea8187a9a96b}},name:test10_2,size:5368709120,volumeId:17461,accountId:6,id:17461,hypervisorType:KVM}},wait:0}}]
 }
2014-09-22 07:03:52,871 DEBUG [c.c.a.m.AgentManagerImpl] 
(AgentManager-Handler-7:null) SeqA 9-1551798: Processing Seq 9-1551798:  { Cmd 
, MgmtId: -1, via: 9, Ver: v1, Flags: 11, 
[{com.cloud.agent.api.ConsoleProxyLoadReportCommand:{_proxyVmId:22358,_loadInfo:{\n
  \connections\: []\n},wait:0}}] }
2014-09-22 07:03:52,873 DEBUG [c.c.a.m.AgentManagerImpl] 
(AgentManager-Handler-7:null) SeqA 9-1551798: Sending Seq 9-1551798:  { Ans: , 
MgmtId: 144343444026, via: 9, Ver: v1, Flags: 100010, 
[{com.cloud.agent.api.AgentControlAnswer:{result:true,wait:0}}] }
2014-09-22 07:03:53,145 DEBUG [c.c.a.ApiServlet] (catalina-exec-4:ctx-27d5d351) 
===START===  172.27.2.190 -- GET  
command=queryAsyncJobResultjobId=3b9a0e48-cca9-49bd-8ba5-f3ee34fecd0aresponse=jsonsessionkey=7fKPZgeg1u%2Bw7dPiPjqnCTze56k%3Dprojectid=729be1b4-429f-410c-b4fb-a9f5bfa3ec50_=1411387433075
2014-09-22 07:03:53,154 DEBUG [c.c.a.ApiServlet] (catalina-exec-4:ctx-27d5d351 
ctx-01bf906b) ===END===  172.27.2.190 -- GET  
command=queryAsyncJobResultjobId=3b9a0e48-cca9-49bd-8ba5-f3ee34fecd0aresponse=jsonsessionkey=7fKPZgeg1u%2Bw7dPiPjqnCTze56k%3Dprojectid=729be1b4-429f-410c-b4fb-a9f5bfa3ec50_=1411387433075
2014-09-22 07:03:53,613 DEBUG [c.c.a.t.Request] (AgentManager-Handler-1:null) 
Seq 3-192151576: Processing:  { Ans: , MgmtId: 144343444026, via: 3, Ver: v1, 
Flags: 10, 

[jira] [Commented] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-09-18 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14138882#comment-14138882
 ] 

Simon Weller commented on CLOUDSTACK-6460:
--

Some more debugging of the database issue:


org.apache.cloudstack.storage.command.AttachCommand:
{disk:
  {data:
{org.apache.cloudstack.storage.to.VolumeObjectTO:
  
{uuid:558b0140-8d98-429c-9de3-44dc7ec96f51,volumeType:DATADISK,dataStore:
{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:
  
{uuid:076d1cd7-9c80-4302-8e13-ea8187a9a96b,id:201,poolType:CLVM,host:localhost,path:/csstore01,port:0,url:CLVM://localhost//csstore01/?ROLE=PrimarySTOREUUID=076d1cd7-9c80-4302-8e13-ea8187a9a96b
  }

},name:test7_2,size:5368709120,path:558b0140-8d98-429c-9de3-44dc7ec96f51,volumeId:17456,accountId:6,format:QCOW2,id:17456,hypervisorType:KVM
   } 
 }
,diskSeq:5,path:558b0140-8d98-429c-9de3-44dc7ec96f51,type:DATADISK,_details:
  
{managed:false,storagePort:0,storageHost:localhost,volumeSize:5368709120
  }
 },vmName:i-6-22370-VM,inSeq:false,wait:0}}] 
}

 Migration of CLVM volumes to another primary storage fail
 -

 Key: CLOUDSTACK-6460
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6460
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: KVM, Volumes
Affects Versions: 4.2.0, 4.2.1, 4.3.0, 4.4.0
 Environment: KVM clusters with fiber channel SAN storage, CLVM volumes
Reporter: Salvatore Sciacco
 Attachments: CLOUDSTACK-6460-darrentang.patch, 
 CLOUDSTACK-6460-darrentang_2.patch, cloudstack-6460.patch, 
 cloudstack-6460_44.patch


 ACS version: 4.2.1 
 Hypervisors: KVM 
 Storage pool type: CLVM
 Since we upgraded from 4.1 to 4.2.1 moving volumes to a different primary 
 storage pool fail. I've enabled debug on the agents side and I think there is 
 a problem with the format  type conversion
 Volume on database had format QCOW2
 these are the parameters for the first step (CLVM - NFS):
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO: 
 uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:uuid:655d6965-b3f3-4118-a970-d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 _url:nfs://192.168.11.6/home/a1iwstack,_role:Image},name:ROOT-4450,size:5368709120,path:volumes/402/5937,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 {quote}
 Those commads are translated into the agent:
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: qemu-img 
 info /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Execution is 
 successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: 
 /bin/bash -c cp -f /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/5c713376-c418-478c-8a31-89c4181cb48e.qcow2
  
 {quote}
 With the result that the output file isn't a qcow2 file but a raw partition, 
 which in turn make the next step fail.
 (NFS - CLVM)
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 info 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Execution is successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 convert -f qcow2 -O raw 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Exit value is 1
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) qemu-img: Could not 
 open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-2:) Failed to 
 convert 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  to /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 the error was: 
 qemu-img: Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 

[jira] [Commented] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-09-16 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14136210#comment-14136210
 ] 

Simon Weller commented on CLOUDSTACK-6460:
--

I've been digging into this a lot today, and I think I've narrowed the problem 
down to the attachVolume functionality.

It appears that during the attacheVolume operation, the volume format is being 
set to QCOW2 in the database.

Steps to reproduce this:

1. Create a new volume either via GUI or API.
Database shows format field to be NULL.
2. Attach volume to VM.
At this point the format field is set to QCOW2, even though it's really RAW.

How's the database table data:

After Volume Creation:

mysql select * from volumes where id='17451';
+---++---+-+--+-+---+-+--+++--++++-+-+---+--+-++-+-+--+-+-+---++--+---+++++--+--+---+
| id| account_id | domain_id | pool_id | last_pool_id | instance_id | 
device_id | name| uuid | size   | 
folder | path | pod_id | data_center_id | iscsi_name | host_ip | volume_type | 
pool_type | disk_offering_id | template_id | first_snapshot_backup_uuid | 
recreatable | created | attached | updated | removed | 
state | chain_info | update_count | disk_type | vm_snapshot_chain_size | 
iso_id | display_volume | format | min_iops | max_iops | hv_ss_reserve |
+---++---+-+--+-+---+-+--+++--++++-+-+---+--+-++-+-+--+-+-+---++--+---+++++--+--+---+
| 17451 |  6 | 1 |NULL | NULL |NULL |  
NULL | testnew | 9fc165e4-d383-4367-ab0a-9a55e5566110 | 5368709120 | NULL   | 
NULL |   NULL |  2 | NULL   | NULL| DATADISK| NULL  
|3 |NULL | NULL   |   0 | 
2014-09-16 20:14:25 | NULL | 2014-09-16 20:14:25 | NULL| Allocated | 
NULL   |0 | NULL  |   NULL |   NULL |   
   1 | NULL   | NULL | NULL |  NULL |
+---++---+-+--+-+---+-+--+++--++++-+-+---+--+-++-+-+--+-+-+---++--+---+++++--+--+---+
1 row in set (0.00 sec)


mysql select * from storage_pool;
+-+---+--+---+--++++---++--+---++-+-+-++---+-++-+---+
| id  | name  | uuid | pool_type | port | 
data_center_id | pod_id | cluster_id | used_bytes| capacity_bytes | 
host_address | user_info | path   | created | removed | 
update_time | status | storage_provider_name | scope   | hypervisor | managed | 
capacity_iops |
+-+---+--+---+--++++---++--+---++-+-+-++---+-++-+---+
| 201 | clvm1 | 076d1cd7-9c80-4302-8e13-ea8187a9a96b | CLVM  |0 |   
   2 |  2 |  2 | 1022336434176 |  1099507433472 | localhost 
   | NULL  | /csstore01 | 2014-01-15 22:02:18 | NULL| NULL| Up  
   | DefaultPrimary| CLUSTER | NULL   |   0 |  NULL |
| 202 | csstore02 | 8f624e46-33fe-473c-9d05-62aebc151d1b | CLVM  |0 |   
   2 |  2 |  2 | 1019077459968 |  1099507433472 | localhost 
   | NULL  | 

[jira] [Updated] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-09-15 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller updated CLOUDSTACK-6460:
-
Attachment: CLOUDSTACK-6460-darrentang_2.patch

New patch for darrentang.

 Migration of CLVM volumes to another primary storage fail
 -

 Key: CLOUDSTACK-6460
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6460
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: KVM, Volumes
Affects Versions: 4.2.0, 4.2.1, 4.3.0, 4.4.0
 Environment: KVM clusters with fiber channel SAN storage, CLVM volumes
Reporter: Salvatore Sciacco
 Attachments: CLOUDSTACK-6460-darrentang.patch, 
 CLOUDSTACK-6460-darrentang_2.patch, cloudstack-6460.patch, 
 cloudstack-6460_44.patch


 ACS version: 4.2.1 
 Hypervisors: KVM 
 Storage pool type: CLVM
 Since we upgraded from 4.1 to 4.2.1 moving volumes to a different primary 
 storage pool fail. I've enabled debug on the agents side and I think there is 
 a problem with the format  type conversion
 Volume on database had format QCOW2
 these are the parameters for the first step (CLVM - NFS):
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO: 
 uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:uuid:655d6965-b3f3-4118-a970-d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 _url:nfs://192.168.11.6/home/a1iwstack,_role:Image},name:ROOT-4450,size:5368709120,path:volumes/402/5937,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 {quote}
 Those commads are translated into the agent:
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: qemu-img 
 info /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Execution is 
 successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: 
 /bin/bash -c cp -f /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/5c713376-c418-478c-8a31-89c4181cb48e.qcow2
  
 {quote}
 With the result that the output file isn't a qcow2 file but a raw partition, 
 which in turn make the next step fail.
 (NFS - CLVM)
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 info 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Execution is successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 convert -f qcow2 -O raw 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Exit value is 1
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) qemu-img: Could not 
 open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-2:) Failed to 
 convert 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  to /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 the error was: 
 qemu-img: Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 {quote}
 If I change on the database the format of the volume to RAW the effect is 
 even worse as data is lost in the process!
 These are the parameter for the first step (CLVM = NFS)
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:{uuid:655d6965-b3f3-4118-a970d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450
 ,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-4024450VM,accountId:402,format:RAW,id:5937,hypervisorType:KVM
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 

[jira] [Commented] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-09-15 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14134040#comment-14134040
 ] 

Simon Weller commented on CLOUDSTACK-6460:
--

darrentang,

Please use the new patch I just uploaded (CLOUDSTACK-6460-darrentang_2.patch). 

qemu-img is no longer being used, as the code is now detecting its raw to raw, 
so it's using cp instead.

The issue here seems to be that for some reason .raw is being appended onto the 
path, and rather than using qemu-img to determine the file type, the code makes 
as assumption based on the incorrect file extension that it's a raw file, when 
it fact it was a QCOW2 file. This patch should get you going, although really a 
hack. The entire code based around this functionality needs to be re-factored, 
as it's using lots of (bad) assumptions, and numerous ugly conditional 
statements that have been jacked into the code via different developers for new 
features.

 Migration of CLVM volumes to another primary storage fail
 -

 Key: CLOUDSTACK-6460
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6460
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: KVM, Volumes
Affects Versions: 4.2.0, 4.2.1, 4.3.0, 4.4.0
 Environment: KVM clusters with fiber channel SAN storage, CLVM volumes
Reporter: Salvatore Sciacco
 Attachments: CLOUDSTACK-6460-darrentang.patch, 
 CLOUDSTACK-6460-darrentang_2.patch, cloudstack-6460.patch, 
 cloudstack-6460_44.patch


 ACS version: 4.2.1 
 Hypervisors: KVM 
 Storage pool type: CLVM
 Since we upgraded from 4.1 to 4.2.1 moving volumes to a different primary 
 storage pool fail. I've enabled debug on the agents side and I think there is 
 a problem with the format  type conversion
 Volume on database had format QCOW2
 these are the parameters for the first step (CLVM - NFS):
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO: 
 uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:uuid:655d6965-b3f3-4118-a970-d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 _url:nfs://192.168.11.6/home/a1iwstack,_role:Image},name:ROOT-4450,size:5368709120,path:volumes/402/5937,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 {quote}
 Those commads are translated into the agent:
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: qemu-img 
 info /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Execution is 
 successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: 
 /bin/bash -c cp -f /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/5c713376-c418-478c-8a31-89c4181cb48e.qcow2
  
 {quote}
 With the result that the output file isn't a qcow2 file but a raw partition, 
 which in turn make the next step fail.
 (NFS - CLVM)
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 info 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Execution is successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 convert -f qcow2 -O raw 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Exit value is 1
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) qemu-img: Could not 
 open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-2:) Failed to 
 convert 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  to /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 the error was: 
 qemu-img: Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 {quote}
 If I change on the database the format of the volume to RAW 

[jira] [Commented] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-09-12 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14131507#comment-14131507
 ] 

Simon Weller commented on CLOUDSTACK-6460:
--

darrentang,

We're running 4.3 with this patch. Can you be a bit more specific when you say 
it doesn't work?

Are you having problems applying the patch, or does a successful patch and 
rebuild not fix the problem for you?

- Si

 Migration of CLVM volumes to another primary storage fail
 -

 Key: CLOUDSTACK-6460
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6460
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: KVM, Volumes
Affects Versions: 4.2.0, 4.2.1, 4.3.0, 4.4.0
 Environment: KVM clusters with fiber channel SAN storage, CLVM volumes
Reporter: Salvatore Sciacco
 Attachments: cloudstack-6460.patch, cloudstack-6460_44.patch


 ACS version: 4.2.1 
 Hypervisors: KVM 
 Storage pool type: CLVM
 Since we upgraded from 4.1 to 4.2.1 moving volumes to a different primary 
 storage pool fail. I've enabled debug on the agents side and I think there is 
 a problem with the format  type conversion
 Volume on database had format QCOW2
 these are the parameters for the first step (CLVM - NFS):
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO: 
 uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:uuid:655d6965-b3f3-4118-a970-d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 _url:nfs://192.168.11.6/home/a1iwstack,_role:Image},name:ROOT-4450,size:5368709120,path:volumes/402/5937,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 {quote}
 Those commads are translated into the agent:
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: qemu-img 
 info /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Execution is 
 successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: 
 /bin/bash -c cp -f /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/5c713376-c418-478c-8a31-89c4181cb48e.qcow2
  
 {quote}
 With the result that the output file isn't a qcow2 file but a raw partition, 
 which in turn make the next step fail.
 (NFS - CLVM)
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 info 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Execution is successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 convert -f qcow2 -O raw 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Exit value is 1
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) qemu-img: Could not 
 open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-2:) Failed to 
 convert 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  to /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 the error was: 
 qemu-img: Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 {quote}
 If I change on the database the format of the volume to RAW the effect is 
 even worse as data is lost in the process!
 These are the parameter for the first step (CLVM = NFS)
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:{uuid:655d6965-b3f3-4118-a970d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450
 ,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-4024450VM,accountId:402,format:RAW,id:5937,hypervisorType:KVM
 

[jira] [Commented] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-09-12 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14131620#comment-14131620
 ] 

Simon Weller commented on CLOUDSTACK-6460:
--

I'll take a look and see if I can figure out what's going on.

 Migration of CLVM volumes to another primary storage fail
 -

 Key: CLOUDSTACK-6460
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6460
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: KVM, Volumes
Affects Versions: 4.2.0, 4.2.1, 4.3.0, 4.4.0
 Environment: KVM clusters with fiber channel SAN storage, CLVM volumes
Reporter: Salvatore Sciacco
 Attachments: cloudstack-6460.patch, cloudstack-6460_44.patch


 ACS version: 4.2.1 
 Hypervisors: KVM 
 Storage pool type: CLVM
 Since we upgraded from 4.1 to 4.2.1 moving volumes to a different primary 
 storage pool fail. I've enabled debug on the agents side and I think there is 
 a problem with the format  type conversion
 Volume on database had format QCOW2
 these are the parameters for the first step (CLVM - NFS):
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO: 
 uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:uuid:655d6965-b3f3-4118-a970-d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 _url:nfs://192.168.11.6/home/a1iwstack,_role:Image},name:ROOT-4450,size:5368709120,path:volumes/402/5937,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 {quote}
 Those commads are translated into the agent:
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: qemu-img 
 info /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Execution is 
 successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: 
 /bin/bash -c cp -f /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/5c713376-c418-478c-8a31-89c4181cb48e.qcow2
  
 {quote}
 With the result that the output file isn't a qcow2 file but a raw partition, 
 which in turn make the next step fail.
 (NFS - CLVM)
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 info 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Execution is successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 convert -f qcow2 -O raw 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Exit value is 1
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) qemu-img: Could not 
 open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-2:) Failed to 
 convert 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  to /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 the error was: 
 qemu-img: Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 {quote}
 If I change on the database the format of the volume to RAW the effect is 
 even worse as data is lost in the process!
 These are the parameter for the first step (CLVM = NFS)
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:{uuid:655d6965-b3f3-4118-a970d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450
 ,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-4024450VM,accountId:402,format:RAW,id:5937,hypervisorType:KVM
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 

[jira] [Updated] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-09-12 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller updated CLOUDSTACK-6460:
-
Attachment: CLOUDSTACK-6460-darrentang.patch

Test patch for darrentang.

 Migration of CLVM volumes to another primary storage fail
 -

 Key: CLOUDSTACK-6460
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6460
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: KVM, Volumes
Affects Versions: 4.2.0, 4.2.1, 4.3.0, 4.4.0
 Environment: KVM clusters with fiber channel SAN storage, CLVM volumes
Reporter: Salvatore Sciacco
 Attachments: CLOUDSTACK-6460-darrentang.patch, cloudstack-6460.patch, 
 cloudstack-6460_44.patch


 ACS version: 4.2.1 
 Hypervisors: KVM 
 Storage pool type: CLVM
 Since we upgraded from 4.1 to 4.2.1 moving volumes to a different primary 
 storage pool fail. I've enabled debug on the agents side and I think there is 
 a problem with the format  type conversion
 Volume on database had format QCOW2
 these are the parameters for the first step (CLVM - NFS):
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO: 
 uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:uuid:655d6965-b3f3-4118-a970-d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 _url:nfs://192.168.11.6/home/a1iwstack,_role:Image},name:ROOT-4450,size:5368709120,path:volumes/402/5937,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 {quote}
 Those commads are translated into the agent:
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: qemu-img 
 info /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Execution is 
 successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: 
 /bin/bash -c cp -f /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/5c713376-c418-478c-8a31-89c4181cb48e.qcow2
  
 {quote}
 With the result that the output file isn't a qcow2 file but a raw partition, 
 which in turn make the next step fail.
 (NFS - CLVM)
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 info 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Execution is successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 convert -f qcow2 -O raw 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Exit value is 1
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) qemu-img: Could not 
 open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-2:) Failed to 
 convert 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  to /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 the error was: 
 qemu-img: Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 {quote}
 If I change on the database the format of the volume to RAW the effect is 
 even worse as data is lost in the process!
 These are the parameter for the first step (CLVM = NFS)
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:{uuid:655d6965-b3f3-4118-a970d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450
 ,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-4024450VM,accountId:402,format:RAW,id:5937,hypervisorType:KVM
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 

[jira] [Commented] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-09-12 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14132110#comment-14132110
 ] 

Simon Weller commented on CLOUDSTACK-6460:
--

darrentang,

I've uploaded a patch for you to test. All I've tested thus far is that is 
compiles, but I haven't done a feature test yet with your issue. I'll try and 
get to that this weekend, but I didn't want to hold you up if you have time to 
test yourself.

The patch is against the 4.3 source tarball.

- Si

 Migration of CLVM volumes to another primary storage fail
 -

 Key: CLOUDSTACK-6460
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6460
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: KVM, Volumes
Affects Versions: 4.2.0, 4.2.1, 4.3.0, 4.4.0
 Environment: KVM clusters with fiber channel SAN storage, CLVM volumes
Reporter: Salvatore Sciacco
 Attachments: CLOUDSTACK-6460-darrentang.patch, cloudstack-6460.patch, 
 cloudstack-6460_44.patch


 ACS version: 4.2.1 
 Hypervisors: KVM 
 Storage pool type: CLVM
 Since we upgraded from 4.1 to 4.2.1 moving volumes to a different primary 
 storage pool fail. I've enabled debug on the agents side and I think there is 
 a problem with the format  type conversion
 Volume on database had format QCOW2
 these are the parameters for the first step (CLVM - NFS):
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO: 
 uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:uuid:655d6965-b3f3-4118-a970-d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 _url:nfs://192.168.11.6/home/a1iwstack,_role:Image},name:ROOT-4450,size:5368709120,path:volumes/402/5937,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 {quote}
 Those commads are translated into the agent:
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: qemu-img 
 info /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Execution is 
 successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: 
 /bin/bash -c cp -f /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/5c713376-c418-478c-8a31-89c4181cb48e.qcow2
  
 {quote}
 With the result that the output file isn't a qcow2 file but a raw partition, 
 which in turn make the next step fail.
 (NFS - CLVM)
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 info 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Execution is successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 convert -f qcow2 -O raw 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Exit value is 1
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) qemu-img: Could not 
 open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-2:) Failed to 
 convert 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  to /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 the error was: 
 qemu-img: Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 {quote}
 If I change on the database the format of the volume to RAW the effect is 
 even worse as data is lost in the process!
 These are the parameter for the first step (CLVM = NFS)
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:{uuid:655d6965-b3f3-4118-a970d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450
 

[jira] [Commented] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-06-24 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14042381#comment-14042381
 ] 

Simon Weller commented on CLOUDSTACK-6460:
--

I've submitted a patch for review against master for this issue.

https://reviews.apache.org/r/22939/

- Si

 Migration of CLVM volumes to another primary storage fail
 -

 Key: CLOUDSTACK-6460
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6460
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: KVM, Volumes
Affects Versions: 4.2.0, 4.2.1, 4.3.0, 4.4.0
 Environment: KVM clusters with fiber channel SAN storage, CLVM volumes
Reporter: Salvatore Sciacco
 Attachments: cloudstack-6460.patch, cloudstack-6460_44.patch


 ACS version: 4.2.1 
 Hypervisors: KVM 
 Storage pool type: CLVM
 Since we upgraded from 4.1 to 4.2.1 moving volumes to a different primary 
 storage pool fail. I've enabled debug on the agents side and I think there is 
 a problem with the format  type conversion
 Volume on database had format QCOW2
 these are the parameters for the first step (CLVM - NFS):
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO: 
 uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:uuid:655d6965-b3f3-4118-a970-d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 _url:nfs://192.168.11.6/home/a1iwstack,_role:Image},name:ROOT-4450,size:5368709120,path:volumes/402/5937,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 {quote}
 Those commads are translated into the agent:
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: qemu-img 
 info /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Execution is 
 successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: 
 /bin/bash -c cp -f /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/5c713376-c418-478c-8a31-89c4181cb48e.qcow2
  
 {quote}
 With the result that the output file isn't a qcow2 file but a raw partition, 
 which in turn make the next step fail.
 (NFS - CLVM)
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 info 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Execution is successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 convert -f qcow2 -O raw 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Exit value is 1
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) qemu-img: Could not 
 open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-2:) Failed to 
 convert 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  to /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 the error was: 
 qemu-img: Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 {quote}
 If I change on the database the format of the volume to RAW the effect is 
 even worse as data is lost in the process!
 These are the parameter for the first step (CLVM = NFS)
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:{uuid:655d6965-b3f3-4118-a970d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450
 ,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-4024450VM,accountId:402,format:RAW,id:5937,hypervisorType:KVM
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 

[jira] [Updated] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-06-23 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller updated CLOUDSTACK-6460:
-

Attachment: cloudstack-6460_44.patch

Patch for 4.4

 Migration of CLVM volumes to another primary storage fail
 -

 Key: CLOUDSTACK-6460
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6460
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: KVM, Volumes
Affects Versions: 4.2.0, 4.2.1, 4.3.0, 4.4.0
 Environment: KVM clusters with fiber channel SAN storage, CLVM volumes
Reporter: Salvatore Sciacco
 Attachments: cloudstack-6460.patch, cloudstack-6460_44.patch


 ACS version: 4.2.1 
 Hypervisors: KVM 
 Storage pool type: CLVM
 Since we upgraded from 4.1 to 4.2.1 moving volumes to a different primary 
 storage pool fail. I've enabled debug on the agents side and I think there is 
 a problem with the format  type conversion
 Volume on database had format QCOW2
 these are the parameters for the first step (CLVM - NFS):
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO: 
 uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:uuid:655d6965-b3f3-4118-a970-d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 _url:nfs://192.168.11.6/home/a1iwstack,_role:Image},name:ROOT-4450,size:5368709120,path:volumes/402/5937,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 {quote}
 Those commads are translated into the agent:
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: qemu-img 
 info /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Execution is 
 successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: 
 /bin/bash -c cp -f /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/5c713376-c418-478c-8a31-89c4181cb48e.qcow2
  
 {quote}
 With the result that the output file isn't a qcow2 file but a raw partition, 
 which in turn make the next step fail.
 (NFS - CLVM)
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 info 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Execution is successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 convert -f qcow2 -O raw 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Exit value is 1
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) qemu-img: Could not 
 open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-2:) Failed to 
 convert 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  to /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 the error was: 
 qemu-img: Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 {quote}
 If I change on the database the format of the volume to RAW the effect is 
 even worse as data is lost in the process!
 These are the parameter for the first step (CLVM = NFS)
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:{uuid:655d6965-b3f3-4118-a970d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450
 ,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-4024450VM,accountId:402,format:RAW,id:5937,hypervisorType:KVM
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 

[jira] [Commented] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-06-21 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14039847#comment-14039847
 ] 

Simon Weller commented on CLOUDSTACK-6460:
--

I wrote a patch for this yesterday that is identical to yours. I've been 
testing it, and it does seem to address the problem. I agree that source will 
always be raw, so unless we're missing something very obvious here, this should 
take care of the problem. 



 Migration of CLVM volumes to another primary storage fail
 -

 Key: CLOUDSTACK-6460
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6460
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: KVM, Volumes
Affects Versions: 4.2.0, 4.2.1, 4.3.0, 4.4.0
 Environment: KVM clusters with fiber channel SAN storage, CLVM volumes
Reporter: Salvatore Sciacco

 ACS version: 4.2.1 
 Hypervisors: KVM 
 Storage pool type: CLVM
 Since we upgraded from 4.1 to 4.2.1 moving volumes to a different primary 
 storage pool fail. I've enabled debug on the agents side and I think there is 
 a problem with the format  type conversion
 Volume on database had format QCOW2
 these are the parameters for the first step (CLVM - NFS):
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO: 
 uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:uuid:655d6965-b3f3-4118-a970-d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 _url:nfs://192.168.11.6/home/a1iwstack,_role:Image},name:ROOT-4450,size:5368709120,path:volumes/402/5937,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 {quote}
 Those commads are translated into the agent:
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: qemu-img 
 info /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Execution is 
 successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: 
 /bin/bash -c cp -f /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/5c713376-c418-478c-8a31-89c4181cb48e.qcow2
  
 {quote}
 With the result that the output file isn't a qcow2 file but a raw partition, 
 which in turn make the next step fail.
 (NFS - CLVM)
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 info 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Execution is successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 convert -f qcow2 -O raw 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Exit value is 1
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) qemu-img: Could not 
 open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-2:) Failed to 
 convert 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  to /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 the error was: 
 qemu-img: Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 {quote}
 If I change on the database the format of the volume to RAW the effect is 
 even worse as data is lost in the process!
 These are the parameter for the first step (CLVM = NFS)
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:{uuid:655d6965-b3f3-4118-a970d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450
 ,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-4024450VM,accountId:402,format:RAW,id:5937,hypervisorType:KVM
 

[jira] [Updated] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-06-21 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller updated CLOUDSTACK-6460:
-

Attachment: cloudstack-6460.patch

CS 4.3 patch attached

 Migration of CLVM volumes to another primary storage fail
 -

 Key: CLOUDSTACK-6460
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6460
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: KVM, Volumes
Affects Versions: 4.2.0, 4.2.1, 4.3.0, 4.4.0
 Environment: KVM clusters with fiber channel SAN storage, CLVM volumes
Reporter: Salvatore Sciacco
 Attachments: cloudstack-6460.patch


 ACS version: 4.2.1 
 Hypervisors: KVM 
 Storage pool type: CLVM
 Since we upgraded from 4.1 to 4.2.1 moving volumes to a different primary 
 storage pool fail. I've enabled debug on the agents side and I think there is 
 a problem with the format  type conversion
 Volume on database had format QCOW2
 these are the parameters for the first step (CLVM - NFS):
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO: 
 uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:uuid:655d6965-b3f3-4118-a970-d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 _url:nfs://192.168.11.6/home/a1iwstack,_role:Image},name:ROOT-4450,size:5368709120,path:volumes/402/5937,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 {quote}
 Those commads are translated into the agent:
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: qemu-img 
 info /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Execution is 
 successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: 
 /bin/bash -c cp -f /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/5c713376-c418-478c-8a31-89c4181cb48e.qcow2
  
 {quote}
 With the result that the output file isn't a qcow2 file but a raw partition, 
 which in turn make the next step fail.
 (NFS - CLVM)
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 info 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Execution is successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 convert -f qcow2 -O raw 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Exit value is 1
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) qemu-img: Could not 
 open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-2:) Failed to 
 convert 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  to /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 the error was: 
 qemu-img: Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 {quote}
 If I change on the database the format of the volume to RAW the effect is 
 even worse as data is lost in the process!
 These are the parameter for the first step (CLVM = NFS)
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:{uuid:655d6965-b3f3-4118-a970d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450
 ,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-4024450VM,accountId:402,format:RAW,id:5937,hypervisorType:KVM
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 

[jira] [Commented] (CLOUDSTACK-6460) Migration of CLVM volumes to another primary storage fail

2014-06-20 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14038795#comment-14038795
 ] 

Simon Weller commented on CLOUDSTACK-6460:
--

This issue also affects 4.3 and 4.4 from what I can tell during my testing. 
Affected versions needs to be changed so this gets some attention.

 Migration of CLVM volumes to another primary storage fail
 -

 Key: CLOUDSTACK-6460
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6460
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: KVM, Volumes
Affects Versions: 4.2.0
 Environment: KVM clusters with fiber channel SAN storage, CLVM volumes
Reporter: Salvatore Sciacco

 ACS version: 4.2.1 
 Hypervisors: KVM 
 Storage pool type: CLVM
 Since we upgraded from 4.1 to 4.2.1 moving volumes to a different primary 
 storage pool fail. I've enabled debug on the agents side and I think there is 
 a problem with the format  type conversion
 Volume on database had format QCOW2
 these are the parameters for the first step (CLVM - NFS):
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO: 
 uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:uuid:655d6965-b3f3-4118-a970-d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 _url:nfs://192.168.11.6/home/a1iwstack,_role:Image},name:ROOT-4450,size:5368709120,path:volumes/402/5937,volumeId:5937,vmName:i-402-4450-VM,accountId:402,format:QCOW2,id:5937,hypervisorType:KVM}
 {quote}
 Those commads are translated into the agent:
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: qemu-img 
 info /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Execution is 
 successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-1:null) Executing: 
 /bin/bash -c cp -f /dev/FC10KY1/39a25daf-23a1-4b65-99ac-fb98469ac197 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/5c713376-c418-478c-8a31-89c4181cb48e.qcow2
  
 {quote}
 With the result that the output file isn't a qcow2 file but a raw partition, 
 which in turn make the next step fail.
 (NFS - CLVM)
 {quote}
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 info 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Execution is successful.
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Executing: qemu-img 
 convert -f qcow2 -O raw 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) Exit value is 1
 DEBUG [utils.script.Script] (agentRequest-Handler-2:) qemu-img: Could not 
 open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-2:) Failed to 
 convert 
 /mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2
  to /dev/FCSTORAGE/da162325-467b-4e78-af07-4bad85470d66 the error was: 
 qemu-img: Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'qemu-img:
  Could not open 
 '/mnt/b8311c72-fe75-3832-98fc-975445028a12/b9303d8d-cd51-4b6c-a244-43c405df4238.qcow2'
 {quote}
 If I change on the database the format of the volume to RAW the effect is 
 even worse as data is lost in the process!
 These are the parameter for the first step (CLVM = NFS)
 {quote}
 srcTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:{uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{org.apache.cloudstack.storage.to.PrimaryDataStoreTO:{uuid:655d6965-b3f3-4118-a970d50cf6afc365,id:211,poolType:CLVM,host:localhost,path:/FC10KY1,port:0,name:ROOT-4450
 ,size:5368709120,path:39a25daf-23a1-4b65-99ac-fb98469ac197,volumeId:5937,vmName:i-4024450VM,accountId:402,format:RAW,id:5937,hypervisorType:KVM
 destTO:{org.apache.cloudstack.storage.to.VolumeObjectTO:uuid:cda46430-52d7-4bf0-b0c2-adfc78dd011c,volumeType:ROOT,dataStore:{com.cloud.agent.api.to.NfsTO:
  
 

[jira] [Commented] (CLOUDSTACK-6464) [KVM:basic zone- upgrade to 4.3],after any vm restart,all the nics are plugged to default bridge even though trafiic labels are being used

2014-06-04 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6464?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14018502#comment-14018502
 ] 

Simon Weller commented on CLOUDSTACK-6464:
--

After reading the comments from Marcus on the vlan table, I took a look at my 
vlan table in my lab, and noticed the vlan_id just displayed the vlan (22). So 
I changed that to vlan://22, restarted the management server, then restarted a 
project virtual router pair. The network assignments appeared to be normal on 
router startup. I then acquired a new ip to the network assigned it to a LB 
rule and added a VM. The alias was correctly assigned to eth2 (my public 
interface). I then removed the ip, and the ip was correctly removed from eth2.
I'll do a bit more testing after I get back from vacation, but at the current 
time this certainly appears to fix the issue I was experiencing.

- Si

 [KVM:basic zone- upgrade to  4.3],after   any vm restart,all the nics  are 
 plugged to default bridge even though trafiic labels are being used
 --

 Key: CLOUDSTACK-6464
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6464
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: Management Server
Affects Versions: 4.3.0
Reporter: sadhu suresh
Priority: Critical
 Fix For: 4.4.0


  Steps:
 1. create a KVM basic zone with 2 nics on host (pre 4.3 build)
 2.use cloudbr0 for management and cloudbr1 for guest by specifying the 
 traffic labels in the physical networks.
 3.deploy few vms
 4.upgrade to felton GA build as per the Upgrade instructions.
 actual result:
 Upgrade successful but all the vnets that were attached to cloudbr1 before 
 upgrade are attached to cloudbr0.
 Due to this network connectivity is lost.
 Expected result:
 Even after upgrade ,all the vnets should be attached to the same bridge as 
 before upgrade.
 ex:
 before Upgrade : this vms(i-5-616-VM) nic was attached to cloudbr1 and after 
 upgrade and VM stop/start.
 the network rules are getting programmed in cloudbr0 .check below output
 ,984 DEBUG [kvm.resource.LibvirtComputingResource] 
 (agentRequest-Handler-2:null) Executing: 
 /usr/share/cloudstack-common/scripts/vm/network/security_group.py 
 default_network_rules --vmname i-5-616-VM --vmid 616 --vmip 10.x.x245 --vmmac 
 06:14:48:00:00:7f --vif vnet15 --brname cloudbr0 --nicsecips 0:
 dumpxml output for i-5-616-VM after upgrade( after VM restart)
 *
 virsh # dumpxml 38
 domain type='kvm' id='38'
 namei-5-616-VM/name
 uuid87557942-1393-49b3-a73e-ae24c40541d1/uuid
 descriptionOther CentOS (64-bit)/description
 memory unit='KiB'2097152/memory
 currentMemory unit='KiB'2097152/currentMemory
 vcpu placement='static'1/vcpu
 cputune
 shares1000/shares
 /cputune
 os
 type arch='x86_64' machine='rhel6.2.0'hvm/type
 boot dev='cdrom'/
 boot dev='hd'/
 /os
 features
 acpi/
 apic/
 pae/
 /features
 cpu
 /cpu
 clock offset='utc'/
 on_poweroffdestroy/on_poweroff
 on_rebootrestart/on_reboot
 on_crashdestroy/on_crash
 devices
 emulator/usr/libexec/qemu-kvm/emulator
 disk type='file' device='disk'
 driver name='qemu' type='qcow2' cache='none'/
 source 
 file='/mnt/041e5d8e-d9c1-346d-aea9-cd9c7b80a211/75544e9d-a4c9-4a94-943e-b20827676a27'/
 target dev='hda' bus='ide'/
 alias name='ide0-0-0'/
 address type='drive' controller='0' bus='0' target='0' unit='0'/
 /disk
 disk type='file' device='cdrom'
 driver name='qemu' type='raw' cache='none'/
 target dev='hdc' bus='ide'/
 readonly/
 alias name='ide0-1-0'/
 address type='drive' controller='0' bus='1' target='0' unit='0'/
 /disk
 controller type='usb' index='0'
 alias name='usb0'/
 address type='pci' domain='0x' bus='0x00' slot='0x01' function='0x2'/
 /controller
 controller type='ide' index='0'
 alias name='ide0'/
 address type='pci' domain='0x' bus='0x00' slot='0x01' function='0x1'/
 /controller
 interface type='bridge'
 mac address='06:14:48:00:00:7f'/
 source bridge='cloudbr0'/
 target dev='vnet15'/
 model type='e1000'/
 bandwidth
 inbound average='25600' peak='25600'/
 outbound average='25600' peak='25600'/
 /bandwidth
 alias name='net0'/
 address type='pci' domain='0x' bus='0x00' slot='0x03' function='0x0'/
 /interface
 serial type='pty'
 source path='/dev/pts/12'/
 target port='0'/
 alias name='serial0'/
 /serial
 console type='pty' tty='/dev/pts/12'
 source path='/dev/pts/12'/
 target type='serial' port='0'/
 alias name='serial0'/
 /console
 input type='tablet' bus='usb'
 alias name='input0'/
 /input
 input type='mouse' bus='ps2'/
 graphics type='vnc' port='5912' autoport='yes' listen='10.x.x.3'
 listen type='address' 

[jira] [Commented] (CLOUDSTACK-6464) [KVM:basic zone- upgrade to 4.3],after any vm restart,all the nics are plugged to default bridge even though trafiic labels are being used

2014-05-16 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6464?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13993653#comment-13993653
 ] 

Simon Weller commented on CLOUDSTACK-6464:
--

We are experiencing the exact same issue as Serg noted above using vlan 
isolation (traffic labels) in KVM with advanced networking on 6.3.
What's also interesting is that releasing an ip appears to add an additional 
interface to the VR, as if it doesn't think a usable interface currently 
exists, so it hot-plugs a duplicate via libvirtd. It looks like the multiple 
interfaces are causing ARP problems.

Please also look at https://issues.apache.org/jira/browse/CLOUDSTACK-5282 as 
that references a similar issue, but claims that it was related to 
https://issues.apache.org/jira/browse/CLOUDSTACK-5280 and resolved. It may be 
in our case that this is the same issue, but with traffic labels in use being 
the difference.

Note that our platform was upgraded from 4.1.x. 

I'm more than happy to supply any required logs, and I have a lab platform to 
test on.

- Si

 [KVM:basic zone- upgrade to  4.3],after   any vm restart,all the nics  are 
 plugged to default bridge even though trafiic labels are being used
 --

 Key: CLOUDSTACK-6464
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6464
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: Management Server
Affects Versions: 4.3.0
Reporter: sadhu suresh
Priority: Critical
 Fix For: 4.3.1


  Steps:
 1. create a KVM basic zone with 2 nics on host (pre 4.3 build)
 2.use cloudbr0 for management and cloudbr1 for guest by specifying the 
 traffic labels in the physical networks.
 3.deploy few vms
 4.upgrade to felton GA build as per the Upgrade instructions.
 actual result:
 Upgrade successful but all the vnets that were attached to cloudbr1 before 
 upgrade are attached to cloudbr0.
 Due to this network connectivity is lost.
 Expected result:
 Even after upgrade ,all the vnets should be attached to the same bridge as 
 before upgrade.
 ex:
 before Upgrade : this vms(i-5-616-VM) nic was attached to cloudbr1 and after 
 upgrade and VM stop/start.
 the network rules are getting programmed in cloudbr0 .check below output
 ,984 DEBUG [kvm.resource.LibvirtComputingResource] 
 (agentRequest-Handler-2:null) Executing: 
 /usr/share/cloudstack-common/scripts/vm/network/security_group.py 
 default_network_rules --vmname i-5-616-VM --vmid 616 --vmip 10.x.x245 --vmmac 
 06:14:48:00:00:7f --vif vnet15 --brname cloudbr0 --nicsecips 0:
 dumpxml output for i-5-616-VM after upgrade( after VM restart)
 *
 virsh # dumpxml 38
 domain type='kvm' id='38'
 namei-5-616-VM/name
 uuid87557942-1393-49b3-a73e-ae24c40541d1/uuid
 descriptionOther CentOS (64-bit)/description
 memory unit='KiB'2097152/memory
 currentMemory unit='KiB'2097152/currentMemory
 vcpu placement='static'1/vcpu
 cputune
 shares1000/shares
 /cputune
 os
 type arch='x86_64' machine='rhel6.2.0'hvm/type
 boot dev='cdrom'/
 boot dev='hd'/
 /os
 features
 acpi/
 apic/
 pae/
 /features
 cpu
 /cpu
 clock offset='utc'/
 on_poweroffdestroy/on_poweroff
 on_rebootrestart/on_reboot
 on_crashdestroy/on_crash
 devices
 emulator/usr/libexec/qemu-kvm/emulator
 disk type='file' device='disk'
 driver name='qemu' type='qcow2' cache='none'/
 source 
 file='/mnt/041e5d8e-d9c1-346d-aea9-cd9c7b80a211/75544e9d-a4c9-4a94-943e-b20827676a27'/
 target dev='hda' bus='ide'/
 alias name='ide0-0-0'/
 address type='drive' controller='0' bus='0' target='0' unit='0'/
 /disk
 disk type='file' device='cdrom'
 driver name='qemu' type='raw' cache='none'/
 target dev='hdc' bus='ide'/
 readonly/
 alias name='ide0-1-0'/
 address type='drive' controller='0' bus='1' target='0' unit='0'/
 /disk
 controller type='usb' index='0'
 alias name='usb0'/
 address type='pci' domain='0x' bus='0x00' slot='0x01' function='0x2'/
 /controller
 controller type='ide' index='0'
 alias name='ide0'/
 address type='pci' domain='0x' bus='0x00' slot='0x01' function='0x1'/
 /controller
 interface type='bridge'
 mac address='06:14:48:00:00:7f'/
 source bridge='cloudbr0'/
 target dev='vnet15'/
 model type='e1000'/
 bandwidth
 inbound average='25600' peak='25600'/
 outbound average='25600' peak='25600'/
 /bandwidth
 alias name='net0'/
 address type='pci' domain='0x' bus='0x00' slot='0x03' function='0x0'/
 /interface
 serial type='pty'
 source path='/dev/pts/12'/
 target port='0'/
 alias name='serial0'/
 /serial
 console type='pty' tty='/dev/pts/12'
 source path='/dev/pts/12'/
 target type='serial' port='0'/
 alias name='serial0'/
 /console
 input 

[jira] [Commented] (CLOUDSTACK-6065) No HA for shutdown VM

2014-02-11 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6065?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13897849#comment-13897849
 ] 

Simon Weller commented on CLOUDSTACK-6065:
--

Animesh mentioned in the CLOUDSTACK-5731 notes that the new vmsync code isn't 
in the 4.3 release. The NPE looks virtually identical though. Can someone 
clarify this? 
This is a blocker in my opinion.

 No HA for shutdown VM
 -

 Key: CLOUDSTACK-6065
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6065
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: Management Server, UI
Affects Versions: 4.3.0
 Environment: CentOS 6.5 Mgmt  HVs, KVM
Reporter: Nux
Assignee: Kelven Yang
Priority: Critical
  Labels: ha, kvm

 I create a service offering with HA, launch a VM with it. 
 I stop the VM from its console or SSH by issuing `poweroff` and the VM is not 
 started back on.
 This is what I see in the management log:
 http://fpaste.org/75667/67633139/raw/
 The UI keeps showing this VM as Running.
 However, if I force a reboot of its hypervisor ACS does do the right thing 
 and starts the VM on another HV.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Comment Edited] (CLOUDSTACK-6065) No HA for shutdown VM

2014-02-11 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-6065?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13897849#comment-13897849
 ] 

Simon Weller edited comment on CLOUDSTACK-6065 at 2/11/14 1:44 PM:
---

Animesh mentioned in the CLOUDSTACK-5713 notes that the new vmsync code isn't 
in the 4.3 release. The NPE looks virtually identical though. Can someone 
clarify this? 
This is a blocker in my opinion.


was (Author: sweller):
Animesh mentioned in the CLOUDSTACK-5731 notes that the new vmsync code isn't 
in the 4.3 release. The NPE looks virtually identical though. Can someone 
clarify this? 
This is a blocker in my opinion.

 No HA for shutdown VM
 -

 Key: CLOUDSTACK-6065
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-6065
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: Management Server, UI
Affects Versions: 4.3.0
 Environment: CentOS 6.5 Mgmt  HVs, KVM
Reporter: Nux
Assignee: Kelven Yang
Priority: Critical
  Labels: ha, kvm

 I create a service offering with HA, launch a VM with it. 
 I stop the VM from its console or SSH by issuing `poweroff` and the VM is not 
 started back on.
 This is what I see in the management log:
 http://fpaste.org/75667/67633139/raw/
 The UI keeps showing this VM as Running.
 However, if I force a reboot of its hypervisor ACS does do the right thing 
 and starts the VM on another HV.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Closed] (CLOUDSTACK-3716) NPE triggered in DownloadListener

2013-09-18 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-3716?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller closed CLOUDSTACK-3716.



 NPE triggered in DownloadListener
 -

 Key: CLOUDSTACK-3716
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-3716
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: Management Server
Affects Versions: 4.1.0, 4.2.0
 Environment: RHEL 6.3 platform.
Reporter: Simon Weller
Assignee: Min Chen
Priority: Blocker
 Fix For: 4.2.0

 Attachments: DownloadMonitorImpl.4.1.patch


 NPE thrown in DownLoadListener process:
 2013-07-22 11:46:51,670 DEBUG [agent.transport.Request] 
 (AgentConnectTaskPool-2323:null) Seq 30-1100873733: Received:  { Ans: , 
 MgmtId: 159090354809909, via: 30, Ver: v1, Flags: 110, { ListTemplateAnswer } 
 }
 2013-07-22 11:46:51,670 DEBUG [agent.manager.AgentManagerImpl] 
 (AgentConnectTaskPool-2323:null) Details from executing class 
 com.cloud.agent.api.storage.ListTemplateCommand: success
 2013-07-22 11:46:51,682 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found routing-3 already in the 
 template host table
 2013-07-22 11:46:51,685 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found centos55-x86_64 already 
 in the template host table
 2013-07-22 11:46:51,688 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync did not find 
 centos56-x86_64-xen ready on server 11, will request download to start/resume 
 shortly
 2013-07-22 11:46:51,689 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync did not find centos53-x64 
 ready on server 11, will request download to start/resume shortly
 2013-07-22 11:46:51,693 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 225-19-353032af-fd13-33d5-843b-1f7b1c321182 already in the template host table
 2013-07-22 11:46:51,696 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 226-19-08172144-a8e2-3377-9d76-9a6231de5060 already in the template host table
 2013-07-22 11:46:51,699 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 7686af3e-34f7-4ae3-ac69-0d089bcbd55e already in the template host table
 2013-07-22 11:46:51,702 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 5e0726a4-73db-45d6-be2a-003c6ce8ff6e already in the template host table
 2013-07-22 11:46:51,705 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 a647c887-59ef-4b81-aa54-1d15bba5272a already in the template host table
 2013-07-22 11:46:51,708 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 272-9-699a6075-54fb-3505-aa5b-68191614286e already in the template host table
 2013-07-22 11:46:51,711 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 273-19-ffa343f8-6933-384c-846a-cf150a970791 already in the template host table
 2013-07-22 11:46:51,714 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 a2b9fcd7-2f19-48dd-affa-a60c574a75b1 already in the template host table
 2013-07-22 11:46:51,716 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 5545c5fb-1b45-4dbc-8b1d-3cdac2c6394e already in the template host table
 2013-07-22 11:46:51,729 DEBUG [agent.transport.Request] 
 (AgentConnectTaskPool-2323:null) Seq 30-1100873734: Sending  { Cmd , MgmtId: 
 159090354809909, via: 30, Ver: v1, Flags: 100111, 
 [{storage.ListVolumeCommand:{secUrl:nfs://172.27.68.201/secondarystorage,wait:0}}]
  }
 2013-07-22 11:46:51,832 DEBUG [agent.transport.Request] 
 (AgentManager-Handler-9:null) Seq 30-1100873734: Processing:  { Ans: , 
 MgmtId: 159090354809909, via: 30, Ver: v1, Flags: 110, 
 [{storage.ListVolumeAnswer:{secUrl:nfs://172.27.68.201/secondarystorage,templateInfos:{215:{templateName:deviis01_1,installPath:volumes/215//82c2d426-3361-347c-93f4-c355d49fc7e2.qcow2,size:21487961088,physicalSize:8349614080,id:215,isPublic:true,isCorrupted:false}},result:true,details:success,wait:0}}]
  }
 2013-07-22 11:46:51,832 DEBUG [agent.manager.AgentAttache] 
 (AgentManager-Handler-9:null) Seq 30-1100873734: No more commands found
 2013-07-22 11:46:51,832 DEBUG [agent.transport.Request] 
 (AgentConnectTaskPool-2323:null) Seq 30-1100873734: Received:  { Ans: , 
 MgmtId: 159090354809909, via: 30, Ver: v1, Flags: 110, { ListVolumeAnswer } 

[jira] [Updated] (CLOUDSTACK-3716) NPE triggered in DownloadListener

2013-07-23 Thread Simon Weller (JIRA)

 [ 
https://issues.apache.org/jira/browse/CLOUDSTACK-3716?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Weller updated CLOUDSTACK-3716:
-

Attachment: DownloadMonitorImpl.4.1.patch

Back ported patch for 4.1

 NPE triggered in DownloadListener
 -

 Key: CLOUDSTACK-3716
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-3716
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: Management Server
Affects Versions: 4.1.0, 4.2.0
 Environment: RHEL 6.3 platform.
Reporter: Simon Weller
Assignee: Min Chen
Priority: Blocker
 Fix For: 4.2.0

 Attachments: DownloadMonitorImpl.4.1.patch


 NPE thrown in DownLoadListener process:
 2013-07-22 11:46:51,670 DEBUG [agent.transport.Request] 
 (AgentConnectTaskPool-2323:null) Seq 30-1100873733: Received:  { Ans: , 
 MgmtId: 159090354809909, via: 30, Ver: v1, Flags: 110, { ListTemplateAnswer } 
 }
 2013-07-22 11:46:51,670 DEBUG [agent.manager.AgentManagerImpl] 
 (AgentConnectTaskPool-2323:null) Details from executing class 
 com.cloud.agent.api.storage.ListTemplateCommand: success
 2013-07-22 11:46:51,682 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found routing-3 already in the 
 template host table
 2013-07-22 11:46:51,685 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found centos55-x86_64 already 
 in the template host table
 2013-07-22 11:46:51,688 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync did not find 
 centos56-x86_64-xen ready on server 11, will request download to start/resume 
 shortly
 2013-07-22 11:46:51,689 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync did not find centos53-x64 
 ready on server 11, will request download to start/resume shortly
 2013-07-22 11:46:51,693 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 225-19-353032af-fd13-33d5-843b-1f7b1c321182 already in the template host table
 2013-07-22 11:46:51,696 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 226-19-08172144-a8e2-3377-9d76-9a6231de5060 already in the template host table
 2013-07-22 11:46:51,699 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 7686af3e-34f7-4ae3-ac69-0d089bcbd55e already in the template host table
 2013-07-22 11:46:51,702 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 5e0726a4-73db-45d6-be2a-003c6ce8ff6e already in the template host table
 2013-07-22 11:46:51,705 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 a647c887-59ef-4b81-aa54-1d15bba5272a already in the template host table
 2013-07-22 11:46:51,708 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 272-9-699a6075-54fb-3505-aa5b-68191614286e already in the template host table
 2013-07-22 11:46:51,711 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 273-19-ffa343f8-6933-384c-846a-cf150a970791 already in the template host table
 2013-07-22 11:46:51,714 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 a2b9fcd7-2f19-48dd-affa-a60c574a75b1 already in the template host table
 2013-07-22 11:46:51,716 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 5545c5fb-1b45-4dbc-8b1d-3cdac2c6394e already in the template host table
 2013-07-22 11:46:51,729 DEBUG [agent.transport.Request] 
 (AgentConnectTaskPool-2323:null) Seq 30-1100873734: Sending  { Cmd , MgmtId: 
 159090354809909, via: 30, Ver: v1, Flags: 100111, 
 [{storage.ListVolumeCommand:{secUrl:nfs://172.27.68.201/secondarystorage,wait:0}}]
  }
 2013-07-22 11:46:51,832 DEBUG [agent.transport.Request] 
 (AgentManager-Handler-9:null) Seq 30-1100873734: Processing:  { Ans: , 
 MgmtId: 159090354809909, via: 30, Ver: v1, Flags: 110, 
 [{storage.ListVolumeAnswer:{secUrl:nfs://172.27.68.201/secondarystorage,templateInfos:{215:{templateName:deviis01_1,installPath:volumes/215//82c2d426-3361-347c-93f4-c355d49fc7e2.qcow2,size:21487961088,physicalSize:8349614080,id:215,isPublic:true,isCorrupted:false}},result:true,details:success,wait:0}}]
  }
 2013-07-22 11:46:51,832 DEBUG [agent.manager.AgentAttache] 
 (AgentManager-Handler-9:null) Seq 30-1100873734: No more commands found
 2013-07-22 11:46:51,832 DEBUG [agent.transport.Request] 
 (AgentConnectTaskPool-2323:null) Seq 30-1100873734: Received:  

[jira] [Commented] (CLOUDSTACK-3716) NPE triggered in DownloadListener

2013-07-23 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-3716?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13716398#comment-13716398
 ] 

Simon Weller commented on CLOUDSTACK-3716:
--

Min,

I've added a patch for review for 4.1 to this issue. 

 NPE triggered in DownloadListener
 -

 Key: CLOUDSTACK-3716
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-3716
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: Management Server
Affects Versions: 4.1.0, 4.2.0
 Environment: RHEL 6.3 platform.
Reporter: Simon Weller
Assignee: Min Chen
Priority: Blocker
 Fix For: 4.2.0

 Attachments: DownloadMonitorImpl.4.1.patch


 NPE thrown in DownLoadListener process:
 2013-07-22 11:46:51,670 DEBUG [agent.transport.Request] 
 (AgentConnectTaskPool-2323:null) Seq 30-1100873733: Received:  { Ans: , 
 MgmtId: 159090354809909, via: 30, Ver: v1, Flags: 110, { ListTemplateAnswer } 
 }
 2013-07-22 11:46:51,670 DEBUG [agent.manager.AgentManagerImpl] 
 (AgentConnectTaskPool-2323:null) Details from executing class 
 com.cloud.agent.api.storage.ListTemplateCommand: success
 2013-07-22 11:46:51,682 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found routing-3 already in the 
 template host table
 2013-07-22 11:46:51,685 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found centos55-x86_64 already 
 in the template host table
 2013-07-22 11:46:51,688 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync did not find 
 centos56-x86_64-xen ready on server 11, will request download to start/resume 
 shortly
 2013-07-22 11:46:51,689 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync did not find centos53-x64 
 ready on server 11, will request download to start/resume shortly
 2013-07-22 11:46:51,693 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 225-19-353032af-fd13-33d5-843b-1f7b1c321182 already in the template host table
 2013-07-22 11:46:51,696 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 226-19-08172144-a8e2-3377-9d76-9a6231de5060 already in the template host table
 2013-07-22 11:46:51,699 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 7686af3e-34f7-4ae3-ac69-0d089bcbd55e already in the template host table
 2013-07-22 11:46:51,702 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 5e0726a4-73db-45d6-be2a-003c6ce8ff6e already in the template host table
 2013-07-22 11:46:51,705 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 a647c887-59ef-4b81-aa54-1d15bba5272a already in the template host table
 2013-07-22 11:46:51,708 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 272-9-699a6075-54fb-3505-aa5b-68191614286e already in the template host table
 2013-07-22 11:46:51,711 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 273-19-ffa343f8-6933-384c-846a-cf150a970791 already in the template host table
 2013-07-22 11:46:51,714 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 a2b9fcd7-2f19-48dd-affa-a60c574a75b1 already in the template host table
 2013-07-22 11:46:51,716 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 5545c5fb-1b45-4dbc-8b1d-3cdac2c6394e already in the template host table
 2013-07-22 11:46:51,729 DEBUG [agent.transport.Request] 
 (AgentConnectTaskPool-2323:null) Seq 30-1100873734: Sending  { Cmd , MgmtId: 
 159090354809909, via: 30, Ver: v1, Flags: 100111, 
 [{storage.ListVolumeCommand:{secUrl:nfs://172.27.68.201/secondarystorage,wait:0}}]
  }
 2013-07-22 11:46:51,832 DEBUG [agent.transport.Request] 
 (AgentManager-Handler-9:null) Seq 30-1100873734: Processing:  { Ans: , 
 MgmtId: 159090354809909, via: 30, Ver: v1, Flags: 110, 
 [{storage.ListVolumeAnswer:{secUrl:nfs://172.27.68.201/secondarystorage,templateInfos:{215:{templateName:deviis01_1,installPath:volumes/215//82c2d426-3361-347c-93f4-c355d49fc7e2.qcow2,size:21487961088,physicalSize:8349614080,id:215,isPublic:true,isCorrupted:false}},result:true,details:success,wait:0}}]
  }
 2013-07-22 11:46:51,832 DEBUG [agent.manager.AgentAttache] 
 (AgentManager-Handler-9:null) Seq 30-1100873734: No more commands found
 2013-07-22 11:46:51,832 DEBUG [agent.transport.Request] 
 

[jira] [Commented] (CLOUDSTACK-3716) NPE triggered in DownloadListener

2013-07-22 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-3716?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715854#comment-13715854
 ] 

Simon Weller commented on CLOUDSTACK-3716:
--

I've been working on trying to reproduce this in our lab environment, and have 
so far failed. I'm going to continue looking at it this evening.

As you can probably see in the debug, we're running KVM. This NPE also existed 
in 4.0.x.

 NPE triggered in DownloadListener
 -

 Key: CLOUDSTACK-3716
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-3716
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: Management Server
Affects Versions: 4.1.0, 4.2.0
 Environment: RHEL 6.3 platform.
Reporter: Simon Weller
Assignee: Min Chen
Priority: Blocker

 NPE thrown in DownLoadListener process:
 2013-07-22 11:46:51,670 DEBUG [agent.transport.Request] 
 (AgentConnectTaskPool-2323:null) Seq 30-1100873733: Received:  { Ans: , 
 MgmtId: 159090354809909, via: 30, Ver: v1, Flags: 110, { ListTemplateAnswer } 
 }
 2013-07-22 11:46:51,670 DEBUG [agent.manager.AgentManagerImpl] 
 (AgentConnectTaskPool-2323:null) Details from executing class 
 com.cloud.agent.api.storage.ListTemplateCommand: success
 2013-07-22 11:46:51,682 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found routing-3 already in the 
 template host table
 2013-07-22 11:46:51,685 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found centos55-x86_64 already 
 in the template host table
 2013-07-22 11:46:51,688 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync did not find 
 centos56-x86_64-xen ready on server 11, will request download to start/resume 
 shortly
 2013-07-22 11:46:51,689 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync did not find centos53-x64 
 ready on server 11, will request download to start/resume shortly
 2013-07-22 11:46:51,693 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 225-19-353032af-fd13-33d5-843b-1f7b1c321182 already in the template host table
 2013-07-22 11:46:51,696 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 226-19-08172144-a8e2-3377-9d76-9a6231de5060 already in the template host table
 2013-07-22 11:46:51,699 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 7686af3e-34f7-4ae3-ac69-0d089bcbd55e already in the template host table
 2013-07-22 11:46:51,702 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 5e0726a4-73db-45d6-be2a-003c6ce8ff6e already in the template host table
 2013-07-22 11:46:51,705 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 a647c887-59ef-4b81-aa54-1d15bba5272a already in the template host table
 2013-07-22 11:46:51,708 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 272-9-699a6075-54fb-3505-aa5b-68191614286e already in the template host table
 2013-07-22 11:46:51,711 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 273-19-ffa343f8-6933-384c-846a-cf150a970791 already in the template host table
 2013-07-22 11:46:51,714 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 a2b9fcd7-2f19-48dd-affa-a60c574a75b1 already in the template host table
 2013-07-22 11:46:51,716 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 5545c5fb-1b45-4dbc-8b1d-3cdac2c6394e already in the template host table
 2013-07-22 11:46:51,729 DEBUG [agent.transport.Request] 
 (AgentConnectTaskPool-2323:null) Seq 30-1100873734: Sending  { Cmd , MgmtId: 
 159090354809909, via: 30, Ver: v1, Flags: 100111, 
 [{storage.ListVolumeCommand:{secUrl:nfs://172.27.68.201/secondarystorage,wait:0}}]
  }
 2013-07-22 11:46:51,832 DEBUG [agent.transport.Request] 
 (AgentManager-Handler-9:null) Seq 30-1100873734: Processing:  { Ans: , 
 MgmtId: 159090354809909, via: 30, Ver: v1, Flags: 110, 
 [{storage.ListVolumeAnswer:{secUrl:nfs://172.27.68.201/secondarystorage,templateInfos:{215:{templateName:deviis01_1,installPath:volumes/215//82c2d426-3361-347c-93f4-c355d49fc7e2.qcow2,size:21487961088,physicalSize:8349614080,id:215,isPublic:true,isCorrupted:false}},result:true,details:success,wait:0}}]
  }
 2013-07-22 11:46:51,832 DEBUG [agent.manager.AgentAttache] 
 (AgentManager-Handler-9:null) Seq 30-1100873734: No more 

[jira] [Commented] (CLOUDSTACK-3716) NPE triggered in DownloadListener

2013-07-22 Thread Simon Weller (JIRA)

[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-3716?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715954#comment-13715954
 ] 

Simon Weller commented on CLOUDSTACK-3716:
--

Min,

Thanks for the extremely fast work on this bug. Are you able to back port this 
to the 4.1.1 branch as well?

Thanks.

 NPE triggered in DownloadListener
 -

 Key: CLOUDSTACK-3716
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-3716
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public(Anyone can view this level - this is the 
 default.) 
  Components: Management Server
Affects Versions: 4.1.0, 4.2.0
 Environment: RHEL 6.3 platform.
Reporter: Simon Weller
Assignee: Min Chen
Priority: Blocker
 Fix For: 4.2.0


 NPE thrown in DownLoadListener process:
 2013-07-22 11:46:51,670 DEBUG [agent.transport.Request] 
 (AgentConnectTaskPool-2323:null) Seq 30-1100873733: Received:  { Ans: , 
 MgmtId: 159090354809909, via: 30, Ver: v1, Flags: 110, { ListTemplateAnswer } 
 }
 2013-07-22 11:46:51,670 DEBUG [agent.manager.AgentManagerImpl] 
 (AgentConnectTaskPool-2323:null) Details from executing class 
 com.cloud.agent.api.storage.ListTemplateCommand: success
 2013-07-22 11:46:51,682 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found routing-3 already in the 
 template host table
 2013-07-22 11:46:51,685 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found centos55-x86_64 already 
 in the template host table
 2013-07-22 11:46:51,688 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync did not find 
 centos56-x86_64-xen ready on server 11, will request download to start/resume 
 shortly
 2013-07-22 11:46:51,689 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync did not find centos53-x64 
 ready on server 11, will request download to start/resume shortly
 2013-07-22 11:46:51,693 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 225-19-353032af-fd13-33d5-843b-1f7b1c321182 already in the template host table
 2013-07-22 11:46:51,696 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 226-19-08172144-a8e2-3377-9d76-9a6231de5060 already in the template host table
 2013-07-22 11:46:51,699 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 7686af3e-34f7-4ae3-ac69-0d089bcbd55e already in the template host table
 2013-07-22 11:46:51,702 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 5e0726a4-73db-45d6-be2a-003c6ce8ff6e already in the template host table
 2013-07-22 11:46:51,705 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 a647c887-59ef-4b81-aa54-1d15bba5272a already in the template host table
 2013-07-22 11:46:51,708 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 272-9-699a6075-54fb-3505-aa5b-68191614286e already in the template host table
 2013-07-22 11:46:51,711 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 273-19-ffa343f8-6933-384c-846a-cf150a970791 already in the template host table
 2013-07-22 11:46:51,714 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 a2b9fcd7-2f19-48dd-affa-a60c574a75b1 already in the template host table
 2013-07-22 11:46:51,716 INFO  [storage.download.DownloadMonitorImpl] 
 (AgentConnectTaskPool-2323:null) Template Sync found 
 5545c5fb-1b45-4dbc-8b1d-3cdac2c6394e already in the template host table
 2013-07-22 11:46:51,729 DEBUG [agent.transport.Request] 
 (AgentConnectTaskPool-2323:null) Seq 30-1100873734: Sending  { Cmd , MgmtId: 
 159090354809909, via: 30, Ver: v1, Flags: 100111, 
 [{storage.ListVolumeCommand:{secUrl:nfs://172.27.68.201/secondarystorage,wait:0}}]
  }
 2013-07-22 11:46:51,832 DEBUG [agent.transport.Request] 
 (AgentManager-Handler-9:null) Seq 30-1100873734: Processing:  { Ans: , 
 MgmtId: 159090354809909, via: 30, Ver: v1, Flags: 110, 
 [{storage.ListVolumeAnswer:{secUrl:nfs://172.27.68.201/secondarystorage,templateInfos:{215:{templateName:deviis01_1,installPath:volumes/215//82c2d426-3361-347c-93f4-c355d49fc7e2.qcow2,size:21487961088,physicalSize:8349614080,id:215,isPublic:true,isCorrupted:false}},result:true,details:success,wait:0}}]
  }
 2013-07-22 11:46:51,832 DEBUG [agent.manager.AgentAttache] 
 (AgentManager-Handler-9:null) Seq 30-1100873734: No more commands found
 2013-07-22 11:46:51,832 DEBUG [agent.transport.Request] 
 

[jira] [Created] (CLOUDSTACK-3716) NPE triggered in DownloadListener

2013-07-22 Thread Simon Weller (JIRA)
Simon Weller created CLOUDSTACK-3716:


 Summary: NPE triggered in DownloadListener
 Key: CLOUDSTACK-3716
 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-3716
 Project: CloudStack
  Issue Type: Bug
  Security Level: Public (Anyone can view this level - this is the default.)
  Components: Management Server
Affects Versions: 4.1.0
 Environment: RHEL 6.3 platform.
Reporter: Simon Weller


NPE thrown in DownLoadListener process:

2013-07-22 11:46:51,670 DEBUG [agent.transport.Request] 
(AgentConnectTaskPool-2323:null) Seq 30-1100873733: Received:  { Ans: , MgmtId: 
159090354809909, via: 30, Ver: v1, Flags: 110, { ListTemplateAnswer } }
2013-07-22 11:46:51,670 DEBUG [agent.manager.AgentManagerImpl] 
(AgentConnectTaskPool-2323:null) Details from executing class 
com.cloud.agent.api.storage.ListTemplateCommand: success
2013-07-22 11:46:51,682 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync found routing-3 already in the 
template host table
2013-07-22 11:46:51,685 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync found centos55-x86_64 already in 
the template host table
2013-07-22 11:46:51,688 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync did not find centos56-x86_64-xen 
ready on server 11, will request download to start/resume shortly
2013-07-22 11:46:51,689 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync did not find centos53-x64 ready 
on server 11, will request download to start/resume shortly
2013-07-22 11:46:51,693 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync found 
225-19-353032af-fd13-33d5-843b-1f7b1c321182 already in the template host table
2013-07-22 11:46:51,696 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync found 
226-19-08172144-a8e2-3377-9d76-9a6231de5060 already in the template host table
2013-07-22 11:46:51,699 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync found 
7686af3e-34f7-4ae3-ac69-0d089bcbd55e already in the template host table
2013-07-22 11:46:51,702 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync found 
5e0726a4-73db-45d6-be2a-003c6ce8ff6e already in the template host table
2013-07-22 11:46:51,705 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync found 
a647c887-59ef-4b81-aa54-1d15bba5272a already in the template host table
2013-07-22 11:46:51,708 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync found 
272-9-699a6075-54fb-3505-aa5b-68191614286e already in the template host table
2013-07-22 11:46:51,711 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync found 
273-19-ffa343f8-6933-384c-846a-cf150a970791 already in the template host table
2013-07-22 11:46:51,714 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync found 
a2b9fcd7-2f19-48dd-affa-a60c574a75b1 already in the template host table
2013-07-22 11:46:51,716 INFO  [storage.download.DownloadMonitorImpl] 
(AgentConnectTaskPool-2323:null) Template Sync found 
5545c5fb-1b45-4dbc-8b1d-3cdac2c6394e already in the template host table
2013-07-22 11:46:51,729 DEBUG [agent.transport.Request] 
(AgentConnectTaskPool-2323:null) Seq 30-1100873734: Sending  { Cmd , MgmtId: 
159090354809909, via: 30, Ver: v1, Flags: 100111, 
[{storage.ListVolumeCommand:{secUrl:nfs://172.27.68.201/secondarystorage,wait:0}}]
 }
2013-07-22 11:46:51,832 DEBUG [agent.transport.Request] 
(AgentManager-Handler-9:null) Seq 30-1100873734: Processing:  { Ans: , MgmtId: 
159090354809909, via: 30, Ver: v1, Flags: 110, 
[{storage.ListVolumeAnswer:{secUrl:nfs://172.27.68.201/secondarystorage,templateInfos:{215:{templateName:deviis01_1,installPath:volumes/215//82c2d426-3361-347c-93f4-c355d49fc7e2.qcow2,size:21487961088,physicalSize:8349614080,id:215,isPublic:true,isCorrupted:false}},result:true,details:success,wait:0}}]
 }
2013-07-22 11:46:51,832 DEBUG [agent.manager.AgentAttache] 
(AgentManager-Handler-9:null) Seq 30-1100873734: No more commands found
2013-07-22 11:46:51,832 DEBUG [agent.transport.Request] 
(AgentConnectTaskPool-2323:null) Seq 30-1100873734: Received:  { Ans: , MgmtId: 
159090354809909, via: 30, Ver: v1, Flags: 110, { ListVolumeAnswer } }
2013-07-22 11:46:51,832 DEBUG [agent.manager.AgentManagerImpl] 
(AgentConnectTaskPool-2323:null) Details from executing class 
com.cloud.agent.api.storage.ListVolumeCommand: success
2013-07-22 11:46:51,836 ERROR [agent.manager.AgentManagerImpl] 
(AgentConnectTaskPool-2323:null) Monitor DownloadListener says there is an 
error in the connect process for 30 due