Hi Dan, Sorry for the delay, kind of hectic this days.
It seems at first sight that you are not including in the VM template the CONTEXT section [1], which will explain why you are not getting the block device attached to your VM. Best regards, -Tino [1] http://opennebula.org/documentation:rel1.4:template#context_section -- Constantino Vázquez Blanco | dsa-research.org/tinova Virtualization Technology Engineer / Researcher OpenNebula Toolkit | opennebula.org On Mon, Jul 19, 2010 at 9:15 PM, Dan Yocum <[email protected]> wrote: > Hi Tino, > > Do you seen anything in here where I've made a mistake? > > Thanks, > Dan > > > > On 07/16/2010 02:58 PM, Dan Yocum wrote: > >> Tino, >> >> Sorry for the delay - we're doing a bunch of upgrades over the last >> couple of days. Here's the output you requested: >> >> $ onevm show 21 >> VIRTUAL MACHINE 21 INFORMATION >> ID : 21 >> NAME : fermicloud012 >> STATE : ACTIVE >> LCM_STATE : RUNNING >> START TIME : 07/08 09:25:49 >> END TIME : - >> DEPLOY ID: : one-21 >> >> VIRTUAL MACHINE TEMPLATE >> CPU=0.2 >> DISK=[ >> READONLY=no, >> SOURCE=/scratch/images/fclvm01.img, >> TARGET=vda ] >> FEATURES=[ >> ACPI=no ] >> GRAPHICS=[ >> LISTEN=127.0.0.1, >> PORT=5900, >> TYPE=vnc ] >> MEMORY=1024 >> NAME=fermicloud012 >> NIC=[ >> BRIDGE=br0, >> IP=131.225.154.145, >> MAC=54:52:83:e1:9a:91, >> NETWORK=Small network, >> VNID=0 ] >> VMID=21 >> >> >> Did I send you the command that is being sent to the workernode from the >> /var/log/libvirtd/qemu/one-21.log file? I can't recall. Here it is: >> >> LC_ALL=C PATH=/sbin:/usr/sbin:/bin:/usr/bin /usr/bin/kvm -S -M rhel5.4.0 >> -m 1024 -smp 1 -name one-21 -uuid 3bab90ea-1dcf-9601-8819-7e0cbdc9eb46 >> -no-kvm-pit-reinjection -monitor pty -pidfile >> /var/run/libvirt/qemu//one-21.pid -no-acpi -boot c -drive >> file=/srv/cloud/one/var//21/images/disk.0,if=virtio,index=0,boot=on -net >> nic,macaddr=54:52:83:e1:9a:91,vlan=0 -net >> tap,fd=16,script=,vlan=0,ifname=vnet0 -serial none -parallel none -usb >> -vnc 127.0.0.1:0 >> char device redirected to /dev/pts/0 >> >> I notice that there is nothing in the -drive options like >> 'if=ide,media=cdrom,index=2'. Do you know why? >> >> Thanks, >> Dan >> >> >> On 07/14/2010 08:45 AM, Tino Vazquez wrote: >> >>> Hi Dan, >>> >>> From the log file provided, it looks like OpenNebula doesn't think it >>> needs to create a context for this VM. >>> >>> Thu Jul 8 09:25:53 2010 [VM][I]: Virtual Machine has no context >>> >>> which is weird. Could you provide also with >>> >>> onevm show 21 >>> >>> ? >>> >>> Regards, >>> >>> -Tino >>> >>> -- >>> Constantino Vázquez Blanco | dsa-research.org/tinova >>> <http://dsa-research.org/tinova> >>> Virtualization Technology Engineer / Researcher >>> OpenNebula Toolkit | opennebula.org <http://opennebula.org> >>> >>> >>> On Mon, Jul 12, 2010 at 11:14 PM, Dan Yocum <[email protected] >>> <mailto:[email protected]>> wrote: >>> >>> Tino, >>> >>> They are attached. >>> >>> Thanks! >>> Dan >>> >>> >>> >>> On 07/10/2010 01:01 PM, Tino Vazquez wrote: >>> >>> Hi Dan, >>> >>> Could you send us the $ONE_LOCATION/var/<vid>/deployment.0 file? >>> >>> Regards, >>> >>> -Tino >>> >>> -- >>> Constantino Vázquez Blanco | dsa-research.org/tinova >>> <http://dsa-research.org/tinova> >>> <http://dsa-research.org/tinova> >>> Virtualization Technology Engineer / Researcher >>> OpenNebula Toolkit | opennebula.org <http://opennebula.org> >>> <http://opennebula.org> >>> >>> >>> >>> On Thu, Jul 8, 2010 at 11:06 PM, Dan Yocum <[email protected] >>> <mailto:[email protected]> >>> <mailto:[email protected] <mailto:[email protected]>>> wrote: >>> >>> Well, I've run across something interesting - I don't see any >>> reference to a cdrom in the /var/log/libvirt/qemu/one-NN.log >>> file >>> for this VM. >>> >>> Why not? I specified an hdc device in the CONTEXT section >>> of the >>> template file... >>> >>> Dan >>> >>> >>> On 07/08/2010 03:29 PM, Dan Yocum wrote: >>> >>> Hi All, >>> >>> I've finally got a VM booting on KVM and Scientific >>> Linux Fermi v55 >>> (RHEL binary compatible), but for some strange reason it >>> doesn't >>> want to >>> mount the hdc device to run the init.sh script to set up the >>> context. >>> >>> Trying to mount it manually, I get the classic "no >>> medium found" >>> error >>> and dmesg says: >>> >>> hdc: drive_cmd: status=0x41 { DriveReady Error } >>> hdc: drive_cmd: error=0x04 { AbortedCommand } >>> >>> I'm sure this is a simple problem but I just can't see >>> what I'm >>> doing >>> wrong. >>> >>> Any ideas? >>> >>> Thanks! >>> Dan >>> >>> >>> >>> -- >>> Dan Yocum >>> Fermilab 630.840.6509 >>> [email protected] <mailto:[email protected]> <mailto:[email protected] >>> <mailto:[email protected]>>, http://fermigrid.fnal.gov >>> >>> "I fly because it releases my mind from the tyranny of petty >>> things." >>> _______________________________________________ >>> Users mailing list >>> [email protected] <mailto:[email protected]> >>> <mailto:[email protected] >>> <mailto:[email protected]>> >>> >>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org >>> >>> >>> >>> -- >>> Dan Yocum >>> Fermilab 630.840.6509 >>> [email protected] <mailto:[email protected]>, http://fermigrid.fnal.gov >>> "I fly because it releases my mind from the tyranny of petty things." >>> >>> >>> >> > -- > Dan Yocum > Fermilab 630.840.6509 > [email protected], http://fermigrid.fnal.gov > "I fly because it releases my mind from the tyranny of petty things." >
_______________________________________________ Users mailing list [email protected] http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
