Hi Prasanna, I managed to fix the problem, thanks for your advice to turn the agent log level to debug:
https://cwiki.apache.org/confluence/display/CLOUDSTACK/KVM+agent+debug >From the log, I found out that the agent on the KVM host tried to NFS mount directly to 103.25.200.19:/mnt/vol1/sec-storage/template/tmpl/1/3, which was not allowed by the NFS server due to its default configuration to only allow to mount to /mnt/vol1/sec-storage (the root of the NFS share). After I changed the NFS server configuration to allow mount to all sub-directories, re-export the NFS and voila, the system was able to download the template and now both the system VMs (CPVM and SSVM) are running! Many thanks for your help! :) Cheers. On Fri, Jul 12, 2013 at 3:31 PM, Indra Pramana <in...@sg.or.id> wrote: > Hi Prasanna, > > Good day to you, and thank you for your e-mail. > > Yes, the cloudstack-agent service is running on both the KVM hosts. There > is no "cloud" user being created though, when I installed the agent. I > installed the agent as root. > > root@hv-kvm-01:/home/indra# service cloudstack-agent status > * cloud-agent is running > > root@hv-kvm-01:/home/indra# su - cloud > Unknown id: cloud > > Please advise how can I resolve this problem, shall I create the Unix > "cloud" user manually? Basically I follow this instruction to prepare the > KVM host and install the CloudStack agent: > > > http://cloudstack.apache.org/docs/en-US/Apache_CloudStack/4.1.0/html/Installation_Guide/hypervisor-kvm-install-flow.html > > with this instruction from Wido on how to prepare libvirt with Ceph RBD > storage pool support: > > > http://blog.widodh.nl/2013/06/a-quick-note-on-running-cloudstack-with-rbd-on-ubuntu-12-04/ > > I also have checked /var/log/cloud/agent/agent.log and I don't see any > error messages, except this error message which will show up every time I > restart the agent: > > 2013-07-12 15:22:47,454 ERROR [cloud.resource.ServerResourceBase] > (main:null) Nics are not configured! > 2013-07-12 15:22:47,459 INFO [cloud.resource.ServerResourceBase] > (main:null) Designating private to be nic eth0.5 > > More logs can be found here: http://pastebin.com/yeNmCt7S > > I have configured the network bridges on the NIC interface as per this > instruction: > > > http://cloudstack.apache.org/docs/en-US/Apache_CloudStack/4.1.0/html/Installation_Guide/hypervisor-kvm-install-flow.html#hypervisor-host-install-network > > On the zone, I used advanced network configuration with just one physical > network for management, public and guest/private. I didn't include storage, > which by default the traffic will use the management VLAN network. > > Please advise if there's anything else I might have been missing. > > Looking forward to your reply, thank you. > > Cheers. > > > > On Fri, Jul 12, 2013 at 2:56 PM, Prasanna Santhanam <t...@apache.org>wrote: > >> Indeed, cloudstack will go through the allocation to startup the >> system VMs too. So that process is failing to recognize the volume >> (.qcow2) present on your NFS storage. >> >> Can you check if your cloudstack agent service is running with the KVM >> host? And it should've created the user cloud. $id cloud to check if >> there is a user. >> >> Did you see what's happening in the agent logs? These are under >> /var/log/cloud/ on your host when the systemVMs are coming up. If the >> logs are not showing any useful information you can turn on debug >> level for more verbosity. >> >> See here: https://cwiki.apache.org/confluence/x/FgPMAQ >> >> On Fri, Jul 12, 2013 at 02:40:35PM +0800, Indra Pramana wrote: >> > Hi Prasanna, >> > >> > Good day to you, and thank you for your e-mail. >> > >> > Yes, when I export the NFS, I set the permission so that normal user >> will >> > be able to have read/write access to the files (no_root_squash). >> > >> > I have tested and I can have read/write access from my KVM hosts using >> > normal user. BTW, there's no "cloud" user on the hosts, I believe it's >> not >> > created during cloudstack-agent installation? >> > >> > In any case, do you think the template issue and the storage pool >> > allocation issue might be related, or are they two different problems >> > altogether? >> > >> > Looking forward to your reply, thank you. >> > >> > Cheers. >> > >> > >> > >> > On Fri, Jul 12, 2013 at 2:26 PM, Prasanna Santhanam <t...@apache.org> >> wrote: >> > >> > > Can you access the file as user root? Or user cloud? The cloudstack >> > > agent on your KVM host runs as user cloud and the NFS permissions >> > > might be disallowing the volume (.qcow2) from being accessed. >> > > >> > > On Fri, Jul 12, 2013 at 02:16:41PM +0800, Indra Pramana wrote: >> > > > Hi Prasanna, >> > > > >> > > > Good day to you, and thank you for your e-mail. >> > > > >> > > > Yes, the file exists. I can access the file from the management >> server >> > > and >> > > > the two hypervisors hosts if I mount manually. >> > > > >> > > > [root@cs-nas-01 /mnt/vol1/sec-storage/template/tmpl/1/3]# ls >> > > > -la >> > > > total >> > > > 1418787 >> > > > drwxr-xr-x 2 root wheel 4 Jul 11 20:21 >> > > > . >> > > > drwxr-xr-x 3 root wheel 3 Jul 11 20:17 >> > > > .. >> > > > -rw-r--r-- 1 root wheel 725811200 Jul 11 20:21 >> > > > 425b9e5a-fbc7-4637-a33a-fe9d0ed4fa98.qcow2 >> > > > -rw-r--r-- 1 root wheel 295 Jul 11 20:21 >> > > > template.properties >> > > > [root@cs-nas-01 /mnt/vol1/sec-storage/template/tmpl/1/3]# >> > > > pwd >> > > > /mnt/vol1/sec-storage/template/tmpl/1/3 >> > > > >> > > > >> > > > Any advise? >> > > > >> > > > Looking forward to your reply, thank you. >> > > > >> > > > Cheers. >> > > > >> > > > >> > > > >> > > > On Fri, Jul 12, 2013 at 2:07 PM, Prasanna Santhanam <t...@apache.org >> > >> > > wrote: >> > > > >> > > > > Can you check whether there is a file at: >> > > > > nfs://103.25.200.19/mnt/vol1/sec-storage/template/tmpl/1/3/ >> > > > > >> > > > > On Fri, Jul 12, 2013 at 01:59:34PM +0800, Indra Pramana wrote: >> > > > > > Hi Prasanna, >> > > > > > >> > > > > > Thanks for your e-mail. >> > > > > > >> > > > > > I have tried restarting the management server, and the problem >> still >> > > > > > persists. I even tried to re-do the installation and >> configuration >> > > again >> > > > > > from scratch last night, but the problem still there. >> > > > > > >> > > > > > I also noted that on the beginning of the logs, I found some >> error >> > > > > messages >> > > > > > saying that the template cannot be downloaded to the pool. See >> this >> > > logs: >> > > > > > >> > > > > > http://pastebin.com/BY1AVJ08 >> > > > > > >> > > > > > It says it failed because cannot get volume from the pool. >> Could it >> > > be >> > > > > > related, i.e. the absence of the template caused the system VMs >> > > cannot be >> > > > > > created and started? >> > > > > > >> > > > > > I have ensured that I downloaded the system VM template using >> > > > > > cloud-install-sys-tmplt and verified that the template is >> already >> > > there >> > > > > in >> > > > > > the secondary storage server. >> > > > > > >> > > > > > Any advice is appreciated. >> > > > > > >> > > > > > Looking forward to your reply, thank you. >> > > > > > >> > > > > > Cheers. >> > > > > > >> > > > > > >> > > > > > >> > > > > > On Fri, Jul 12, 2013 at 1:21 PM, Prasanna Santhanam < >> t...@apache.org> >> > > > > wrote: >> > > > > > >> > > > > > > It looks like a previous attempt to start the systemVMs has >> failed >> > > > > > > putting the nfs storage in the avoid set. Did you try >> restarting >> > > your >> > > > > > > management server? >> > > > > > > >> > > > > > > This line leads me to the above mentioned: >> > > > > > > 2013-07-12 13:10:48,236 DEBUG >> > > > > > > [storage.allocator.AbstractStoragePoolAllocator] >> > > (secstorage-1:null) >> > > > > > > StoragePool is in avoid set, skipping this pool >> > > > > > > >> > > > > > > >> > > > > > > On Fri, Jul 12, 2013 at 01:16:53PM +0800, Indra Pramana wrote: >> > > > > > > > Dear Wido and all, >> > > > > > > > >> > > > > > > > I have managed to get the hosts, primary and secondary >> storage >> > > > > running. >> > > > > > > > >> > > > > > > > - 2 KVM hypervisor hosts >> > > > > > > > - One RBD primary storage >> > > > > > > > - One NFS primary storage (for system VMs, since I >> understand >> > > that >> > > > > system >> > > > > > > > VMs cannot use RBD) >> > > > > > > > - One NFS secondary storage >> > > > > > > > >> > > > > > > > However, now I am having problem with the system VMs: CPVM >> and >> > > SSVM, >> > > > > > > unable >> > > > > > > > to start. >> > > > > > > > >> > > > > > > > Excerpt from management-server.log file is here: >> > > > > > > > http://pastebin.com/ENkpCALY >> > > > > > > > >> > > > > > > > It seems that the VMs were not able to be created because >> unable >> > > to >> > > > > find >> > > > > > > > suitable StoragePools. >> > > > > > > > >> > > > > > > > I understand that system VMs will be using the NFS primary >> > > storage >> > > > > > > instead >> > > > > > > > of RBD, so I have confirmed that I am able to mount the >> primary >> > > > > storage >> > > > > > > via >> > > > > > > > NFS and have read and write access, from both the >> hypervisor and >> > > the >> > > > > > > > management server. >> > > > > > > > >> > > > > > > > Any advise how can I resolve the problem to make both the >> system >> > > VMs >> > > > > > > > created and started? >> > > > > > > > >> > > > > > > > Looking forward to your reply, thank you. >> > > > > > > > >> > > > > > > > Cheers. >> > > > > > > > >> > > > > > > > >> > > > > > > > On Fri, Jul 12, 2013 at 9:43 AM, Indra Pramana < >> in...@sg.or.id> >> > > > > wrote: >> > > > > > > > >> > > > > > > > > Hi Wido, >> > > > > > > > > >> > > > > > > > > Thanks for the advice, I'm now able to add the RBD pool as >> > > primary >> > > > > > > storage. >> > > > > > > > > >> > > > > > > > > Many thanks! :) >> > > > > > > > > >> > > > > > > > > Cheers. >> > > > > > > > > >> > > > > > > > > >> > > > > > > > > On Thursday, July 11, 2013, Wido den Hollander wrote: >> > > > > > > > > >> > > > > > > > >> Hi, >> > > > > > > > >> >> > > > > > > > >> On 07/10/2013 03:42 PM, Chip Childers wrote: >> > > > > > > > >> >> > > > > > > > >>> Cc'ing Wido, our resident Ceph expert. ;-) >> > > > > > > > >>> >> > > > > > > > >>> >> > > > > > > > >> Hehe ;) >> > > > > > > > >> >> > > > > > > > >> On Wed, Jul 10, 2013 at 05:45:25PM +0800, Indra Pramana >> > > wrote: >> > > > > > > > >>> >> > > > > > > > >>>> Dear all, >> > > > > > > > >>>> >> > > > > > > > >>>> I am installing CloudStack 4.1.0 (upgraded from 4.0.2) >> and I >> > > > > also >> > > > > > > have a >> > > > > > > > >>>> Ceph cluster running. However, I am having issues in >> adding >> > > the >> > > > > RBD >> > > > > > > as >> > > > > > > > >>>> primary storage. Tried to follow the instruction here, >> but >> > > > > unable to >> > > > > > > > >>>> make >> > > > > > > > >>>> it work: >> > > > > > > > >>>> >> > > > > > > > >>>> http://ceph.com/docs/master/**rbd/rbd-cloudstack/< >> > > > > > > http://ceph.com/docs/master/rbd/rbd-cloudstack/> >> > > > > > > > >>>> >> > > > > > > > >>>> I have setup a pool on the Ceph cluster. The status of >> the >> > > > > cluster >> > > > > > > is >> > > > > > > > >>>> healthy. Since I am using Ubuntu 12.04.2 LTS (Precise) >> for >> > > the >> > > > > > > > >>>> hypervisors, >> > > > > > > > >>>> I also have compiled libvirt manually to ensure that >> the >> > > version >> > > > > > > 0.9.13 >> > > > > > > > >>>> is >> > > > > > > > >>>> installed (previously it's 0.9.8). >> > > > > > > > >>>> >> > > > > > > > >>>> >> > > > > > > > >> You can also use the Ubuntu Cloud Archive, I still need >> to >> > > get the >> > > > > > > docs >> > > > > > > > >> updated for that. >> > > > > > > > >> >> > > > > > > > >> I described the process in a blogpost: >> > > > > > > http://blog.widodh.nl/2013/06/** >> > > > > > > > >> >> > > a-quick-note-on-running-**cloudstack-with-rbd-on-ubuntu-**12-04/< >> > > > > > > >> > > > > >> > > >> http://blog.widodh.nl/2013/06/a-quick-note-on-running-cloudstack-with-rbd-on-ubuntu-12-04/ >> > > > > > > > >> > > > > > > > >> >> > > > > > > > >> indra@hv-kvm-01:~/rbd$ ceph >> > > > > > > > >>>> ceph> health >> > > > > > > > >>>> HEALTH_OK >> > > > > > > > >>>> >> > > > > > > > >>>> indra@hv-kvm-01:~$ ceph osd lspools >> > > > > > > > >>>> 0 data,1 metadata,2 rbd,3 sc1, >> > > > > > > > >>>> >> > > > > > > > >>>> root@hv-kvm-01:/home/indra# libvirtd --version >> > > > > > > > >>>> libvirtd (libvirt) 0.9.13 >> > > > > > > > >>>> >> > > > > > > > >>>> I tried to add Primary Storage into the Cloudstack zone >> > > which I >> > > > > have >> > > > > > > > >>>> created: >> > > > > > > > >>>> >> > > > > > > > >>>> Add Primary Storage: >> > > > > > > > >>>> >> > > > > > > > >>>> Zone: my zone name >> > > > > > > > >>>> Pod: my pod name >> > > > > > > > >>>> Cluster: my cluster name >> > > > > > > > >>>> Name: ceph-rbd-pri-storage >> > > > > > > > >>>> Protocol: RBD >> > > > > > > > >>>> RADOS Monitor: my first Ceph monitor IP address >> > > > > > > > >>>> RADOS Pool: sc1 (the pool name on Ceph cluster) >> > > > > > > > >>>> RADOS User: client.admin >> > > > > > > > >>>> RADOS Secret: /etc/ceph/ceph.client.admin.**keyring >> (keyring >> > > > > file >> > > > > > > > >>>> location) >> > > > > > > > >>>> >> > > > > > > > >>> >> > > > > > > > >> This is your problem. That shouldn't be the location of >> the >> > > file, >> > > > > but >> > > > > > > it >> > > > > > > > >> should be the secret, which is a base64 encoded string. >> > > > > > > > >> >> > > > > > > > >> $ ceph auth list >> > > > > > > > >> >> > > > > > > > >> That should tell you what the secret is. >> > > > > > > > >> >> > > > > > > > >> Storage Tags: rbd >> > > > > > > > >> >> > > > > > > > >> This is the error message when I tried to add the primary >> > > storage >> > > > > by >> > > > > > > > >> clicking OK: >> > > > > > > > >> >> > > > > > > > >> DB Exception on: >> > > com.mysql.jdbc.**JDBC4PreparedStatement@4b2eb56 >> > > > > **: >> > > > > > > > >> INSERT INTO >> > > > > > > > >> storage_pool (storage_pool.id, storage_pool.name, >> > > > > storage_pool.uuid, >> > > > > > > > >> storage_pool.pool_type, storage_pool.created, >> > > > > > > storage_pool.update_time, >> > > > > > > > >> storage_pool.data_center_id, storage_pool.pod_id, >> > > > > > > > >> storage_pool.available_bytes, >> storage_pool.capacity_bytes, >> > > > > > > > >> storage_pool.status, storage_pool.scope, >> > > > > > > storage_pool.storage_provider_** >> > > > > > > > >> id, >> > > > > > > > >> storage_pool.host_address, storage_pool.path, >> > > storage_pool.port, >> > > > > > > > >> storage_pool.user_info, storage_pool.cluster_id) VALUES >> (217, >> > > > > > > > >> _binary'ceph-rbd-pri-storage', >> > > > > > > > >> _binary'a226c9a1-da78-3f3a-**b5ac-e18b925c9634', 'RBD', >> > > > > '2013-07-10 >> > > > > > > > >> 09:08:28', null, 2, 2, 0, 0, 'Up', null, null, null, >> > > _binary'ceph/ >> > > > > > > > >> ceph.client.admin.keyring@10.**237.11.2/sc1< >> > > > > > > http://ceph.client.admin.keyring@10.237.11.2/sc1>', >> > > > > > > > >> 6789, null, 2) >> > > > > > > > >> >> > > > > > > > >> On the management-server.log file: >> > > > > > > > >> >> > > > > > > > >> 2013-07-10 17:08:28,845 DEBUG [cloud.api.ApiServlet] >> > > > > > > > >> (catalina-exec-2:null) >> > > > > > > > >> ===START=== 192.168.0.100 -- GET >> > > > > > > > >> >> command=createStoragePool&**zoneid=c116950e-e4ae-4f23-**a7e7- >> > > > > > > > >> >> 74a75c4ee638&podId=a748b063-**3a83-4175-a0e9-de39118fe5ce&** >> > > > > > > > >> >> clusterid=1f87eb09-324d-4d49-**83c2-88d84d7a15df&name=ceph-** >> > > > > > > > >> rbd-pri-storage&url=rbd%3A%2F%**2Fclient.admin%3A_etc%2Fc >> > > > > > > > >> eph%2Fceph.client.admin.**keyring%4010.237.11.2%2Fsc1&** >> > > > > > > > >> >> > > > > >> tags=rbd&response=json&**sessionkey=**rDRfWpqeKfQKbKZtHr398ULV%2F8k%** >> > > > > > > > >> 3D&_=1373447307839 >> > > > > > > > >> 2013-07-10 17:08:28,862 DEBUG >> > > [cloud.storage.**StorageManagerImpl] >> > > > > > > > >> (catalina-exec-2:null) createPool Params @ scheme - rbd >> > > > > storageHost - >> > > > > > > null >> > > > > > > > >> hostPath - /ceph/ceph.client >> > > > > > > > >> .admin.keyring@10.237.11.2/sc1 port - -1 >> > > > > > > > >> 2013-07-10 17:08:28,918 DEBUG >> > > [cloud.storage.**StorageManagerImpl] >> > > > > > > > >> (catalina-exec-2:null) In createPool Setting poolId - 217 >> > > uuid - >> > > > > > > > >> a226c9a1-da78-3f3a-b5ac-**e18b925c9634 z >> > > > > > > > >> oneId - 2 podId - 2 poolName - ceph-rbd-pri-storage >> > > > > > > > >> 2013-07-10 17:08:28,921 DEBUG >> [db.Transaction.Transaction] >> > > > > > > > >> (catalina-exec-2:null) Rolling back the transaction: >> Time = 3 >> > > > > Name = >> > > > > > > > >> persist; called by -Transaction.rollbac >> > > > > > > > >> >> k:890-Transaction.removeUpTo:**833-Transaction.close:657-** >> > > > > > > > >> TransactionContextBuilder.**interceptException:63-** >> > > > > > > > >> >> > > > > >> ComponentInstantiationPostProc**essor$InterceptorDispatcher.**interce >> > > > > > > > >> pt:133-StorageManagerImpl.**createPool:1378-** >> > > > > > > > >> >> StorageManagerImpl.createPool:**147-CreateStoragePoolCmd.** >> > > > > > > > >> execute:123-ApiDispatcher.**dispatch:162-ApiServer.** >> > > > > > > > >> queueCommand:505-ApiSe >> > > > > > > > >> rver.handleRequest:355-**ApiServlet.processRequest:302 >> > > > > > > > >> 2013-07-10 17:08:28,923 ERROR [cloud.api.ApiServer] >> > > > > > > (catalina-exec-2:null) >> > > > > > > > >> unhandled exception executing api command: >> createStoragePool >> > > > > > > > >> com.cloud.utils.exception.**CloudRuntimeException: DB >> > > Exception >> > > > > on: >> > > > > > > > >> com.mysql.jdbc.**JDBC4PreparedStatement@4b2eb56**: >> INSERT >> > > INTO >> > > > > > > > >> storage_pool ( >> > > > > > > > >> storage_pool.id, storage_pool >> > > > > > > > >> .name, storage_pool.uuid, storage_pool.pool_type, >> > > > > > > storage_pool.created, >> > > > > > > > >> storage_pool.update_time, storage_pool.data_center_id, >> > > > > > > > >> storage_pool.pod_id, >> > > > > > > > >> storage_pool.availab >> > > > > > > > >> le_bytes, storage_pool.capacity_bytes, >> storage_pool.status, >> > > > > > > > >> storage_pool.scope, storage_pool.storage_provider_**id, >> > > > > > > > >> storage_pool.host_address, storage_pool.path, storage_ >> > > > > > > > >> pool.port, storage_pool.user_info, >> storage_pool.cluster_id) >> > > VALUES >> > > > > > > (217, >> > > > > > > > >> _binary'ceph-rbd-pri-storage', >> > > > > > > > >> _binary'a226c9a1-da78-3f3a-**b5ac-e18b925c9634', 'RBD', >> > > '2013-07-1 >> > > > > > > > >> 0 09:08:28', null, 2, 2, 0, 0, 'Up', null, null, null, >> > > > > _binary'ceph/ >> > > > > > > > >> ceph.client.admin.keyring@10.**237.11.2/sc1< >> > > > > > > http://ceph.client.admin.keyring@10.237.11.2/sc1>', >> > > > > > > > >> 6789, null, 2) >> > > > > > > > >> at >> > > > > > > > >> >> > > > > > > >> > > >> com.cloud.utils.db.**GenericDaoBase.persist(**GenericDaoBase.java:1342) >> > > > > > > > >> at >> > > > > > > > >> com.cloud.storage.dao.**StoragePoolDaoImpl.persist(** >> > > > > > > > >> StoragePoolDaoImpl.java:232) >> > > > > > > > >> at >> > > > > > > > >> >> com.cloud.utils.component.**ComponentInstantiationPostProc** >> > > > > > > > >> >> > > > > > > >> > > > > >> > > >> essor$InterceptorDispatcher.**intercept(**ComponentInstantiationPostProc* >> > > > > > > > >> *es >> > > > > > > > >> >> > > > > > > > >> >> > > > > > > >> > > > > > > -- >> > > > > > > Prasanna., >> > > > > > > >> > > > > > > ------------------------ >> > > > > > > Powered by BigRock.com >> > > > > > > >> > > > > > > >> > > > > >> > > > > -- >> > > > > Prasanna., >> > > > > >> > > > > ------------------------ >> > > > > Powered by BigRock.com >> > > > > >> > > > > >> > > >> > > -- >> > > Prasanna., >> > > >> > > ------------------------ >> > > Powered by BigRock.com >> > > >> > > >> >> -- >> Prasanna., >> >> ------------------------ >> Powered by BigRock.com >> >> >