Re: [one-users] Ceph snapshot

2015-01-16 Thread Fabian Zimmermann
Hi,

Am 22.10.14 um 10:42 schrieb Ruben S. Montero:
 They should be supported for RBD 2 images, you may want to take a look to
 the actual cloning process in tm/ceph/clone
looks like /var/lib/one/remotes/tm/ceph/cpds is used in case of a hot
hdd-snapshot.

Here my datastore-template
--
DATASTORE 100 INFORMATION
ID : 100
NAME   : ceph-one
USER   : oneadmin
GROUP  : oneadmin
CLUSTER: -
TYPE   : IMAGE
DS_MAD : ceph
TM_MAD : ceph
BASE PATH  : /var/lib/one//datastores/100
DISK_TYPE  : RBD

DATASTORE CAPACITY
TOTAL: : 9.8T
FREE:  : 9.7T
USED:  : 118.8G
LIMIT: : -

PERMISSIONS
OWNER  : um-
GROUP  : u--
OTHER  : ---

DATASTORE TEMPLATE
BASE_PATH=/var/lib/one//datastores/
BRIDGE_LIST=virt17 virt18
CEPH_HOST=10.0.29.0 10.0.29.1 10.0.29.2
CEPH_SECRET=x-x-x-x-x
CEPH_USER=libvirt
CLONE_TARGET=SELF
DISK_TYPE=RBD
DS_MAD=ceph
LN_TARGET=NONE
POOL_NAME=one
RBD_FORMAT=2
TM_MAD=ceph
TYPE=IMAGE_DS
--

Should one use this script to create a hot snapshot?

If yes, is it a good idea to reuse the

if RBD_FORMAT=2
then
   clone
else
  copy

system of tm/ceph/clone in tm/ceph/cpds or may this lead to other problems?

Thanks,

Fabian
___
Users mailing list
Users@lists.opennebula.org
http://lists.opennebula.org/listinfo.cgi/users-opennebula.org


Re: [one-users] Queries regarding Storage Backend for DataStore

2015-01-16 Thread Fabian Zimmermann
Hi Robert,

Am 14.01.15 um 15:08 schrieb Robert Foote:
 wever, the initial deployments of VMs now, are quite bogged down, and can 
 take upwards of an hour or more to deploy a single VM with a 100Gb drive. We 
 believe this is now due to some NFS bogging. We are currently in the process 
 of moving to iSCSI with our OpenNebula + Nexenta deployments, and are 
 expecting a boost in performance as it relates to the initial deployment of 
 VMs.
Do *NOT* use iSCSI on Nexenta.

We used it in the past and run into different problems. F.e. it caused:
a complete freeze/panic of nexenta, reaching connection limit and
blocking until reboot of nexenta.

Maybe your problem is not related to NFS.

Do you use ZFS dedup or did you use ZFS dedup in the past?

Did you move your ZIL to SSD?

How old are this SSD?


Fabian
___
Users mailing list
Users@lists.opennebula.org
http://lists.opennebula.org/listinfo.cgi/users-opennebula.org


Re: [one-users] Ceph snapshot

2015-01-16 Thread Fabian Zimmermann
Hi,

answering myself:

Am 16.01.15 um 13:59 schrieb Fabian Zimmermann:
 Hi,

 Am 22.10.14 um 10:42 schrieb Ruben S. Montero:
 They should be supported for RBD 2 images, you may want to take a look to
 the actual cloning process in tm/ceph/clone
 looks like /var/lib/one/remotes/tm/ceph/cpds is used in case of a hot
 hdd-snapshot.
yes, because ceph would block the deletion of the source-image until the
new image is flatten and all snapshots are removed.
 Should one use this script to create a hot snapshot?

 If yes, is it a good idea to reuse the

 if RBD_FORMAT=2
 then
clone
 else
   copy

 system of tm/ceph/clone in tm/ceph/cpds or may this lead to other problems?
Already asked at ceph-mailinglist if rbd cp is doing an internal
snapshot to protect consistency of data. If not a snap, protect,
flatten, unprotect, rm would be done with the following patch:

--
--- cpds.orig2015-01-16 14:09:10.116016425 +0100
+++ cpds2015-01-16 15:00:26.666220420 +0100
@@ -1,5 +1,7 @@
 #!/bin/bash

+echo $*  /tmp/$$
+
 #
-- #
 # Copyright 2002-2014, OpenNebula Project (OpenNebula.org), C12G
Labs#
 #   
#
@@ -80,8 +82,27 @@
 RBD=$RBD --id ${CEPH_USER}
 fi

+RBD_SNAP=cpds-${VM_ID}-${DISK_ID}-$( date +%s )
+
+CP_CMD=$(
+cat EOF
+set -e
+
+RBD_FORMAT=\$($RBD info $RBD_DST | sed -n 's/.*format: // p')
+if [ \$RBD_FORMAT = 2 ]; then
+$RBD snap create $RBD_DST@$RBD_SNAP
+$RBD snap protect $RBD_DST@$RBD_SNAP
+$RBD clone $RBD_DST@$RBD_SNAP $DST
+$RBD flatten $DST
+$RBD snap unprotect $RBD_DST@$RBD_SNAP
+$RBD snap rm $RBD_DST@$RBD_SNAP
+else
+$RBD copy $RBD_DST $DST
+fi
+EOF
+)

-ssh_exec_and_log $SRC_HOST $RBD copy $RBD_DST $DST \
+ssh_exec_and_log $SRC_HOST $CP_CMD \
  Error cloning $RBD_DST to $DST in $SRC_HOST

 exit 0
--

I will create a feature-request if rbd cp isn't using an internal
snapshot.

Fabian

___
Users mailing list
Users@lists.opennebula.org
http://lists.opennebula.org/listinfo.cgi/users-opennebula.org


Re: [one-users] ssh/authorized_keys permission denied

2015-01-13 Thread Fabian Zimmermann
Hi,

Am 12.01.15 um 09:56 schrieb Adri Anisa:
 Hi,

 I use ssh-copy-id node to ssh without password from frontend to node, but
 I found this error:
 bash: line 2: ./ssh/authorized_keys: Permission denied.

 I also have tried to chmod 600 ~/.ssh/authorized_keys but it seems still
 error.
I assume you mounted /var/lib/one/ from your frontend?

Then just copy your pub-key to /var/lib/one/.ssh/authorized_keys in your
frontend and it will automatically be used by your nodes.

Fabian

___
Users mailing list
Users@lists.opennebula.org
http://lists.opennebula.org/listinfo.cgi/users-opennebula.org


Re: [one-users] Delete Recreate fails in case of Host-Error

2014-12-12 Thread Fabian Zimmermann
Hi Ruben,


Am 12.12.14 11:20, schrieb Ruben S. Montero:
 In OpenNebula 4.10 if the VM is in UNKNOWN it should go directly to boot
 (bypassing CLEANUP and PROLOG) , provided you are using shared storage...
we are using 4.10.1 and it looks like CLEANUP is executed isn't it?

--
Wed Dec 10 16:38:45 2014 [Z0][LCM][I]: New VM state is RUNNING
Thu Dec 11 09:16:01 2014 [Z0][LCM][I]: New VM state is UNKNOWN
Thu Dec 11 09:23:32 2014 [Z0][LCM][I]: New VM state is CLEANUP.
Thu Dec 11 09:23:32 2014 [Z0][DiM][I]: New VM state is PENDING
--

As written, i'm just using the ERR-Host-Hook to execute the
host-error.rb and deleterecreate the VMs.

Fabian

___
Users mailing list
Users@lists.opennebula.org
http://lists.opennebula.org/listinfo.cgi/users-opennebula.org


[one-users] Delete Recreate fails in case of Host-Error

2014-12-11 Thread Fabian Zimmermann
Hi,

I just setup our fencing system and everything is working so far, but if
I use the HOST-ERROR-Hook to deleterecreate the VMs this will fail.

The CLEANUP is tried on the broken host - of course it will fail - and
the followed CLONE will fail, because of already existing snapshots/disks.

I just workaround this - see attached patch, which will just skip clone
if disk/snap already exists, but what's the correct way to handle this?

Thanks a lot,

Fabian Zimmermann
--- clone.old   2014-12-11 13:13:52.001648056 +0100
+++ clone   2014-12-11 09:32:18.882111020 +0100
@@ -76,14 +76,15 @@
 CLONE_CMD=$(cat EOF
 set -e

-RBD_FORMAT=\$($RBD info $SRC_PATH | sed -n 's/.*format: // p')
-
-if [ \$RBD_FORMAT = 2 ]; then
-$RBD snap create $SRC_PATH@$RBD_SNAP
-$RBD snap protect $SRC_PATH@$RBD_SNAP
-$RBD clone $SRC_PATH@$RBD_SNAP $RBD_DST
-else
-$RBD copy $SRC_PATH $RBD_DST
+if ! $RBD info $RBD_DST; then
+RBD_FORMAT=\$($RBD info $SRC_PATH | sed -n 's/.*format: // p')
+if [ \$RBD_FORMAT = 2 ]; then
+$RBD snap create $SRC_PATH@$RBD_SNAP
+$RBD snap protect $SRC_PATH@$RBD_SNAP
+$RBD clone $SRC_PATH@$RBD_SNAP $RBD_DST
+else
+$RBD copy $SRC_PATH $RBD_DST
+fi
 fi
 EOF
 )
___
Users mailing list
Users@lists.opennebula.org
http://lists.opennebula.org/listinfo.cgi/users-opennebula.org


[one-users] OpenNebula Sheepdog Addon released!

2014-11-24 Thread Fabian Zimmermann
Hi,

just want to inform you about the release of

https://github.com/OpenNebula/addon-sheepdog

I will do my best to keep the addon working.

Please give it a try, any feedback and/or help is welcome!

Thanks a lot,

Fabian Zimmermann

___
Users mailing list
Users@lists.opennebula.org
http://lists.opennebula.org/listinfo.cgi/users-opennebula.org


[one-users] VM removed from one, but still running

2014-07-06 Thread Fabian Zimmermann
Hi,

just wondered, why one of my VMs had a completely broken filesystem.
After some research I found an old instance of my template (I'm using
persistent storage) running on one of my nodes. It looks like OpenNebula
didn't handle the broken auth correctly (caused by NFS-issues) and
marked the storage as free even if the VM is still running, isn't it?

--
Sat Jun 14 20:20:13 2014 [LCM][I]: New VM state is RUNNING
Sat Jun 28 14:11:07 2014 [LCM][I]: New VM state is UNKNOWN
Sat Jun 28 14:11:08 2014 [LCM][I]: New VM state is CLEANUP.
Sat Jun 28 14:11:08 2014 [VMM][I]: Driver command for 74 cancelled
Sat Jun 28 14:11:09 2014 [VMM][I]: Command execution fail:
/var/tmp/one/vmm/kvm/cancel 'one-74' 'node1' 74 node1
Sat Jun 28 14:11:09 2014 [VMM][I]: Warning: Permanently added
'node1,192.168.20.21' (ECDSA) to the list of known hosts.
Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied, please try again.
Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied, please try again.
Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied (publickey,password).
Sat Jun 28 14:11:09 2014 [VMM][I]: ExitSSHCode: 255
Sat Jun 28 14:11:09 2014 [VMM][E]: Error connecting to node1
Sat Jun 28 14:11:09 2014 [VMM][I]: Failed to execute virtualization
driver operation: cancel.
Sat Jun 28 14:11:09 2014 [VMM][I]: Command execution fail:
/var/tmp/one/vnm/dummy/clean
PFZNPjxJRD43NDwvSUQ+PERFUExPWV9JRD5vbmUtNzQ8L0RFUExPWV9JRD48VEVNUExBVEU+PE5JQz48QlJJREdFPjwhW0NEQVRBW2JyMF1dPjwvQlJJREdFPjxJUD48IVtDREFUQVs4OS4yMzguODkuMzZdXT48L0lQPjxJUDZfTElOSz48IVtDREFUQVtmZTgwOjoyMTY6M2VmZjpmZWI5OjU2ZTBdXT48L0lQNl9MSU5LPjxNQUM+PCFbQ0RBVEFbMDA6MTY6M2U6Yjk6NTY6ZTBdXT48L01BQz48TkVUV09SSz48IVtDREFUQVs4OS4yMzguODkuMzItMjldXT48L05FVFdPUks+PE5FVFdPUktfSUQ+PCFbQ0RBVEFbNF1dPjwvTkVUV09SS19JRD48TkVUV09SS19VTkFNRT48IVtDREFUQVtvbmVhZG1pbl1dPjwvTkVUV09SS19VTkFNRT48TklDX0lEPjwhW0NEQVRBWzBdXT48L05JQ19JRD48VkxBTj48IVtDREFUQVtOT11dPjwvVkxBTj48L05JQz48L1RFTVBMQVRFPjxISVNUT1JZX1JFQ09SRFM+PEhJU1RPUlk+PEhPU1ROQU1FPm5vZGUxPC9IT1NUTkFNRT48L0hJU1RPUlk+PC9ISVNUT1JZX1JFQ09SRFM+PC9WTT4=
Sat Jun 28 14:11:09 2014 [VMM][I]: Warning: Permanently added
'node1,192.168.20.21' (ECDSA) to the list of known hosts.
Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied, please try again.
Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied, please try again.
Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied (publickey,password).
Sat Jun 28 14:11:09 2014 [VMM][I]: ExitSSHCode: 255
Sat Jun 28 14:11:09 2014 [VMM][E]: Error connecting to node1
Sat Jun 28 14:11:09 2014 [VMM][I]: Failed to execute network driver
operation: clean.
Sat Jun 28 14:11:09 2014 [VMM][I]: Successfully execute transfer manager
driver operation: tm_delete.
Sat Jun 28 14:11:09 2014 [VMM][I]: Command execution fail:
/var/lib/one/remotes/tm/shared/delete
node1:/var/lib/one//datastores/0/74 74 0
Sat Jun 28 14:11:09 2014 [VMM][I]: delete: Deleting
/var/lib/one/datastores/0/74
Sat Jun 28 14:11:09 2014 [VMM][E]: delete: Command [ -e
/var/lib/one/datastores/0/74 ] || exit 0
Sat Jun 28 14:11:09 2014 [VMM][I]:
Sat Jun 28 14:11:09 2014 [VMM][I]: times=10
Sat Jun 28 14:11:09 2014 [VMM][I]: function=rm -rf
/var/lib/one/datastores/0/74
Sat Jun 28 14:11:09 2014 [VMM][I]:
Sat Jun 28 14:11:09 2014 [VMM][I]: count=1
Sat Jun 28 14:11:09 2014 [VMM][I]:
Sat Jun 28 14:11:09 2014 [VMM][I]: ret=$($function)
Sat Jun 28 14:11:09 2014 [VMM][I]: error=$?
Sat Jun 28 14:11:09 2014 [VMM][I]:
Sat Jun 28 14:11:09 2014 [VMM][I]: while [ $count -lt $times -a $error
!= 0 ]; do
Sat Jun 28 14:11:09 2014 [VMM][I]: sleep 1
Sat Jun 28 14:11:09 2014 [VMM][I]: count=$(( $count + 1 ))
Sat Jun 28 14:11:09 2014 [VMM][I]: ret=$($function)
Sat Jun 28 14:11:09 2014 [VMM][I]: error=$?
Sat Jun 28 14:11:09 2014 [VMM][I]: done
Sat Jun 28 14:11:09 2014 [VMM][I]:
Sat Jun 28 14:11:09 2014 [VMM][I]: [ x$error = x0 ] failed:
Warning: Permanently added 'node1,192.168.20.21' (ECDSA) to the list of
known hosts.
Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied, please try again.
Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied, please try again.
Sat Jun 28 14:11:09 2014 [VMM][I]: Permission denied (publickey,password).
Sat Jun 28 14:11:09 2014 [VMM][E]: Error deleting
/var/lib/one/datastores/0/74
Sat Jun 28 14:11:09 2014 [VMM][I]: ExitCode: 255
Sat Jun 28 14:11:09 2014 [VMM][I]: Failed to execute transfer manager
driver operation: tm_delete.
Sat Jun 28 14:11:09 2014 [VMM][I]: Host successfully cleaned.
Sat Jun 28 14:11:09 2014 [DiM][I]: New VM state is PENDING
Sat Jun 28 14:11:33 2014 [DiM][I]: New VM state is ACTIVE.
Sat Jun 28 14:11:33 2014 [LCM][I]: New VM state is PROLOG.
Sat Jun 28 14:32:57 2014 [LCM][I]: New VM state is CLEANUP.
Sat Jun 28 14:32:57 2014 [DiM][I]: New VM state is DONE
Sat Jun 28 14:32:57 2014 [TM][W]: Ignored: LOG I 74 Driver command for
74 cancelled
Sat Jun 28 14:32:59 2014 [TM][W]: Ignored: TRANSFER SUCCESS 74 -

--

Fabian
___
Users mailing list
Users@lists.opennebula.org

Re: [one-users] VM running on multiple hosts

2014-07-06 Thread Fabian Zimmermann
Hi,

Am 03.07.2014 16:59, schrieb Jaime Melis:
 As far as I know, the shared_lvm hasn't been updated lately:
 - https://github.com/OpenNebula/addon-shared-lvm-single-lock
 - http://community.opennebula.org/shared_lvm

also having the issue in OpenNebula 4.6.0

Fabian

___
Users mailing list
Users@lists.opennebula.org
http://lists.opennebula.org/listinfo.cgi/users-opennebula.org


Re: [one-users] ERROR instantiate VM from cdrom

2014-06-11 Thread Fabian Zimmermann
Hi

Am 10.06.14 16:34, schrieb Maria Jular:
 Then I create a template with two disks but when I run a virtual machine from 
 this template, in console the message always is:
 Operating system NOT found

 What happen?
You have to disable context, because it's exported as iso-image and qemu
tries to boot from this iso instead of the one you think.

Fabian

___
Users mailing list
Users@lists.opennebula.org
http://lists.opennebula.org/listinfo.cgi/users-opennebula.org