Re: [one-users] Ceph snapshot
Hi, Am 22.10.14 um 10:42 schrieb Ruben S. Montero: They should be supported for RBD 2 images, you may want to take a look to the actual cloning process in tm/ceph/clone looks like /var/lib/one/remotes/tm/ceph/cpds is used in case of a hot hdd-snapshot. Here my datastore-template -- DATASTORE 100 INFORMATION ID : 100 NAME : ceph-one USER : oneadmin GROUP : oneadmin CLUSTER: - TYPE : IMAGE DS_MAD : ceph TM_MAD : ceph BASE PATH : /var/lib/one//datastores/100 DISK_TYPE : RBD DATASTORE CAPACITY TOTAL: : 9.8T FREE: : 9.7T USED: : 118.8G LIMIT: : - PERMISSIONS OWNER : um- GROUP : u-- OTHER : --- DATASTORE TEMPLATE BASE_PATH=/var/lib/one//datastores/ BRIDGE_LIST=virt17 virt18 CEPH_HOST=10.0.29.0 10.0.29.1 10.0.29.2 CEPH_SECRET=x-x-x-x-x CEPH_USER=libvirt CLONE_TARGET=SELF DISK_TYPE=RBD DS_MAD=ceph LN_TARGET=NONE POOL_NAME=one RBD_FORMAT=2 TM_MAD=ceph TYPE=IMAGE_DS -- Should one use this script to create a hot snapshot? If yes, is it a good idea to reuse the if RBD_FORMAT=2 then clone else copy system of tm/ceph/clone in tm/ceph/cpds or may this lead to other problems? Thanks, Fabian ___ Users mailing list Users@lists.opennebula.org http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
Re: [one-users] Queries regarding Storage Backend for DataStore
Hi Robert, Am 14.01.15 um 15:08 schrieb Robert Foote: wever, the initial deployments of VMs now, are quite bogged down, and can take upwards of an hour or more to deploy a single VM with a 100Gb drive. We believe this is now due to some NFS bogging. We are currently in the process of moving to iSCSI with our OpenNebula + Nexenta deployments, and are expecting a boost in performance as it relates to the initial deployment of VMs. Do *NOT* use iSCSI on Nexenta. We used it in the past and run into different problems. F.e. it caused: a complete freeze/panic of nexenta, reaching connection limit and blocking until reboot of nexenta. Maybe your problem is not related to NFS. Do you use ZFS dedup or did you use ZFS dedup in the past? Did you move your ZIL to SSD? How old are this SSD? Fabian ___ Users mailing list Users@lists.opennebula.org http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
Re: [one-users] Ceph snapshot
Hi, answering myself: Am 16.01.15 um 13:59 schrieb Fabian Zimmermann: Hi, Am 22.10.14 um 10:42 schrieb Ruben S. Montero: They should be supported for RBD 2 images, you may want to take a look to the actual cloning process in tm/ceph/clone looks like /var/lib/one/remotes/tm/ceph/cpds is used in case of a hot hdd-snapshot. yes, because ceph would block the deletion of the source-image until the new image is flatten and all snapshots are removed. Should one use this script to create a hot snapshot? If yes, is it a good idea to reuse the if RBD_FORMAT=2 then clone else copy system of tm/ceph/clone in tm/ceph/cpds or may this lead to other problems? Already asked at ceph-mailinglist if rbd cp is doing an internal snapshot to protect consistency of data. If not a snap, protect, flatten, unprotect, rm would be done with the following patch: -- --- cpds.orig2015-01-16 14:09:10.116016425 +0100 +++ cpds2015-01-16 15:00:26.666220420 +0100 @@ -1,5 +1,7 @@ #!/bin/bash +echo $* /tmp/$$ + # -- # # Copyright 2002-2014, OpenNebula Project (OpenNebula.org), C12G Labs# # # @@ -80,8 +82,27 @@ RBD=$RBD --id ${CEPH_USER} fi +RBD_SNAP=cpds-${VM_ID}-${DISK_ID}-$( date +%s ) + +CP_CMD=$( +cat EOF +set -e + +RBD_FORMAT=\$($RBD info $RBD_DST | sed -n 's/.*format: // p') +if [ \$RBD_FORMAT = 2 ]; then +$RBD snap create $RBD_DST@$RBD_SNAP +$RBD snap protect $RBD_DST@$RBD_SNAP +$RBD clone $RBD_DST@$RBD_SNAP $DST +$RBD flatten $DST +$RBD snap unprotect $RBD_DST@$RBD_SNAP +$RBD snap rm $RBD_DST@$RBD_SNAP +else +$RBD copy $RBD_DST $DST +fi +EOF +) -ssh_exec_and_log $SRC_HOST $RBD copy $RBD_DST $DST \ +ssh_exec_and_log $SRC_HOST $CP_CMD \ Error cloning $RBD_DST to $DST in $SRC_HOST exit 0 -- I will create a feature-request if rbd cp isn't using an internal snapshot. Fabian ___ Users mailing list Users@lists.opennebula.org http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
Re: [one-users] ssh/authorized_keys permission denied
Hi, Am 12.01.15 um 09:56 schrieb Adri Anisa: Hi, I use ssh-copy-id node to ssh without password from frontend to node, but I found this error: bash: line 2: ./ssh/authorized_keys: Permission denied. I also have tried to chmod 600 ~/.ssh/authorized_keys but it seems still error. I assume you mounted /var/lib/one/ from your frontend? Then just copy your pub-key to /var/lib/one/.ssh/authorized_keys in your frontend and it will automatically be used by your nodes. Fabian ___ Users mailing list Users@lists.opennebula.org http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
Re: [one-users] Delete Recreate fails in case of Host-Error
Hi Ruben, Am 12.12.14 11:20, schrieb Ruben S. Montero: In OpenNebula 4.10 if the VM is in UNKNOWN it should go directly to boot (bypassing CLEANUP and PROLOG) , provided you are using shared storage... we are using 4.10.1 and it looks like CLEANUP is executed isn't it? -- Wed Dec 10 16:38:45 2014 [Z0][LCM][I]: New VM state is RUNNING Thu Dec 11 09:16:01 2014 [Z0][LCM][I]: New VM state is UNKNOWN Thu Dec 11 09:23:32 2014 [Z0][LCM][I]: New VM state is CLEANUP. Thu Dec 11 09:23:32 2014 [Z0][DiM][I]: New VM state is PENDING -- As written, i'm just using the ERR-Host-Hook to execute the host-error.rb and deleterecreate the VMs. Fabian ___ Users mailing list Users@lists.opennebula.org http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
[one-users] Delete Recreate fails in case of Host-Error
Hi, I just setup our fencing system and everything is working so far, but if I use the HOST-ERROR-Hook to deleterecreate the VMs this will fail. The CLEANUP is tried on the broken host - of course it will fail - and the followed CLONE will fail, because of already existing snapshots/disks. I just workaround this - see attached patch, which will just skip clone if disk/snap already exists, but what's the correct way to handle this? Thanks a lot, Fabian Zimmermann --- clone.old 2014-12-11 13:13:52.001648056 +0100 +++ clone 2014-12-11 09:32:18.882111020 +0100 @@ -76,14 +76,15 @@ CLONE_CMD=$(cat EOF set -e -RBD_FORMAT=\$($RBD info $SRC_PATH | sed -n 's/.*format: // p') - -if [ \$RBD_FORMAT = 2 ]; then -$RBD snap create $SRC_PATH@$RBD_SNAP -$RBD snap protect $SRC_PATH@$RBD_SNAP -$RBD clone $SRC_PATH@$RBD_SNAP $RBD_DST -else -$RBD copy $SRC_PATH $RBD_DST +if ! $RBD info $RBD_DST; then +RBD_FORMAT=\$($RBD info $SRC_PATH | sed -n 's/.*format: // p') +if [ \$RBD_FORMAT = 2 ]; then +$RBD snap create $SRC_PATH@$RBD_SNAP +$RBD snap protect $SRC_PATH@$RBD_SNAP +$RBD clone $SRC_PATH@$RBD_SNAP $RBD_DST +else +$RBD copy $SRC_PATH $RBD_DST +fi fi EOF ) ___ Users mailing list Users@lists.opennebula.org http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
[one-users] OpenNebula Sheepdog Addon released!
Hi, just want to inform you about the release of https://github.com/OpenNebula/addon-sheepdog I will do my best to keep the addon working. Please give it a try, any feedback and/or help is welcome! Thanks a lot, Fabian Zimmermann ___ Users mailing list Users@lists.opennebula.org http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
[one-users] VM removed from one, but still running
Hi, just wondered, why one of my VMs had a completely broken filesystem. After some research I found an old instance of my template (I'm using persistent storage) running on one of my nodes. It looks like OpenNebula didn't handle the broken auth correctly (caused by NFS-issues) and marked the storage as free even if the VM is still running, isn't it? -- Sat Jun 14 20:20:13 2014 [LCM][I]: New VM state is RUNNING Sat Jun 28 14:11:07 2014 [LCM][I]: New VM state is UNKNOWN Sat Jun 28 14:11:08 2014 [LCM][I]: New VM state is CLEANUP. Sat Jun 28 14:11:08 2014 [VMM][I]: Driver command for 74 cancelled Sat Jun 28 14:11:09 2014 [VMM][I]: Command execution fail: /var/tmp/one/vmm/kvm/cancel 'one-74' 'node1' 74 node1 Sat Jun 28 14:11:09 2014 [VMM][I]: Warning: Permanently added 'node1,192.168.20.21' (ECDSA) to the list of known hosts. Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied, please try again. Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied, please try again. Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied (publickey,password). Sat Jun 28 14:11:09 2014 [VMM][I]: ExitSSHCode: 255 Sat Jun 28 14:11:09 2014 [VMM][E]: Error connecting to node1 Sat Jun 28 14:11:09 2014 [VMM][I]: Failed to execute virtualization driver operation: cancel. Sat Jun 28 14:11:09 2014 [VMM][I]: Command execution fail: /var/tmp/one/vnm/dummy/clean PFZNPjxJRD43NDwvSUQ+PERFUExPWV9JRD5vbmUtNzQ8L0RFUExPWV9JRD48VEVNUExBVEU+PE5JQz48QlJJREdFPjwhW0NEQVRBW2JyMF1dPjwvQlJJREdFPjxJUD48IVtDREFUQVs4OS4yMzguODkuMzZdXT48L0lQPjxJUDZfTElOSz48IVtDREFUQVtmZTgwOjoyMTY6M2VmZjpmZWI5OjU2ZTBdXT48L0lQNl9MSU5LPjxNQUM+PCFbQ0RBVEFbMDA6MTY6M2U6Yjk6NTY6ZTBdXT48L01BQz48TkVUV09SSz48IVtDREFUQVs4OS4yMzguODkuMzItMjldXT48L05FVFdPUks+PE5FVFdPUktfSUQ+PCFbQ0RBVEFbNF1dPjwvTkVUV09SS19JRD48TkVUV09SS19VTkFNRT48IVtDREFUQVtvbmVhZG1pbl1dPjwvTkVUV09SS19VTkFNRT48TklDX0lEPjwhW0NEQVRBWzBdXT48L05JQ19JRD48VkxBTj48IVtDREFUQVtOT11dPjwvVkxBTj48L05JQz48L1RFTVBMQVRFPjxISVNUT1JZX1JFQ09SRFM+PEhJU1RPUlk+PEhPU1ROQU1FPm5vZGUxPC9IT1NUTkFNRT48L0hJU1RPUlk+PC9ISVNUT1JZX1JFQ09SRFM+PC9WTT4= Sat Jun 28 14:11:09 2014 [VMM][I]: Warning: Permanently added 'node1,192.168.20.21' (ECDSA) to the list of known hosts. Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied, please try again. Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied, please try again. Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied (publickey,password). Sat Jun 28 14:11:09 2014 [VMM][I]: ExitSSHCode: 255 Sat Jun 28 14:11:09 2014 [VMM][E]: Error connecting to node1 Sat Jun 28 14:11:09 2014 [VMM][I]: Failed to execute network driver operation: clean. Sat Jun 28 14:11:09 2014 [VMM][I]: Successfully execute transfer manager driver operation: tm_delete. Sat Jun 28 14:11:09 2014 [VMM][I]: Command execution fail: /var/lib/one/remotes/tm/shared/delete node1:/var/lib/one//datastores/0/74 74 0 Sat Jun 28 14:11:09 2014 [VMM][I]: delete: Deleting /var/lib/one/datastores/0/74 Sat Jun 28 14:11:09 2014 [VMM][E]: delete: Command [ -e /var/lib/one/datastores/0/74 ] || exit 0 Sat Jun 28 14:11:09 2014 [VMM][I]: Sat Jun 28 14:11:09 2014 [VMM][I]: times=10 Sat Jun 28 14:11:09 2014 [VMM][I]: function=rm -rf /var/lib/one/datastores/0/74 Sat Jun 28 14:11:09 2014 [VMM][I]: Sat Jun 28 14:11:09 2014 [VMM][I]: count=1 Sat Jun 28 14:11:09 2014 [VMM][I]: Sat Jun 28 14:11:09 2014 [VMM][I]: ret=$($function) Sat Jun 28 14:11:09 2014 [VMM][I]: error=$? Sat Jun 28 14:11:09 2014 [VMM][I]: Sat Jun 28 14:11:09 2014 [VMM][I]: while [ $count -lt $times -a $error != 0 ]; do Sat Jun 28 14:11:09 2014 [VMM][I]: sleep 1 Sat Jun 28 14:11:09 2014 [VMM][I]: count=$(( $count + 1 )) Sat Jun 28 14:11:09 2014 [VMM][I]: ret=$($function) Sat Jun 28 14:11:09 2014 [VMM][I]: error=$? Sat Jun 28 14:11:09 2014 [VMM][I]: done Sat Jun 28 14:11:09 2014 [VMM][I]: Sat Jun 28 14:11:09 2014 [VMM][I]: [ x$error = x0 ] failed: Warning: Permanently added 'node1,192.168.20.21' (ECDSA) to the list of known hosts. Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied, please try again. Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied, please try again. Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied (publickey,password). Sat Jun 28 14:11:09 2014 [VMM][E]: Error deleting /var/lib/one/datastores/0/74 Sat Jun 28 14:11:09 2014 [VMM][I]: ExitCode: 255 Sat Jun 28 14:11:09 2014 [VMM][I]: Failed to execute transfer manager driver operation: tm_delete. Sat Jun 28 14:11:09 2014 [VMM][I]: Host successfully cleaned. Sat Jun 28 14:11:09 2014 [DiM][I]: New VM state is PENDING Sat Jun 28 14:11:33 2014 [DiM][I]: New VM state is ACTIVE. Sat Jun 28 14:11:33 2014 [LCM][I]: New VM state is PROLOG. Sat Jun 28 14:32:57 2014 [LCM][I]: New VM state is CLEANUP. Sat Jun 28 14:32:57 2014 [DiM][I]: New VM state is DONE Sat Jun 28 14:32:57 2014 [TM][W]: Ignored: LOG I 74 Driver command for 74 cancelled Sat Jun 28 14:32:59 2014 [TM][W]: Ignored: TRANSFER SUCCESS 74 - -- Fabian ___ Users mailing list Users@lists.opennebula.org
Re: [one-users] VM running on multiple hosts
Hi, Am 03.07.2014 16:59, schrieb Jaime Melis: As far as I know, the shared_lvm hasn't been updated lately: - https://github.com/OpenNebula/addon-shared-lvm-single-lock - http://community.opennebula.org/shared_lvm also having the issue in OpenNebula 4.6.0 Fabian ___ Users mailing list Users@lists.opennebula.org http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
Re: [one-users] ERROR instantiate VM from cdrom
Hi Am 10.06.14 16:34, schrieb Maria Jular: Then I create a template with two disks but when I run a virtual machine from this template, in console the message always is: Operating system NOT found What happen? You have to disable context, because it's exported as iso-image and qemu tries to boot from this iso instead of the one you think. Fabian ___ Users mailing list Users@lists.opennebula.org http://lists.opennebula.org/listinfo.cgi/users-opennebula.org