Hmmm..... Pressing the "send" button way too early.... Here is what I got so far.
(1) Using i686 as processor type, the VM booted, and the VM status was RUNNING, and VNC was accessible. Logging into the VNC console, I saw Kernel panic -- which was right because this image needed x86_64 processor. (2) Using x86_x64 processor, the VM failed to boot, with the following information: Wed Apr 24 23:24:56 2013 [DiM][I]: New VM state is ACTIVE. Wed Apr 24 23:24:56 2013 [LCM][I]: New VM state is PROLOG. Wed Apr 24 23:24:56 2013 [VM][I]: Virtual Machine has no context Wed Apr 24 23:35:11 2013 [TM][I]: clone: Cloning opennebula:/srv/cloud/one/var/datastores/100/c5610951183fa21ac0900ce147a677b7 in /vmfs/volumes/101/138/disk.0 Wed Apr 24 23:35:11 2013 [TM][I]: ExitCode: 0 Wed Apr 24 23:35:11 2013 [LCM][I]: New VM state is BOOT Wed Apr 24 23:35:11 2013 [VMM][I]: Generating deployment file: /srv/cloud/one/var/vms/138/deployment.0 Wed Apr 24 23:35:11 2013 [VMM][I]: ExitCode: 0 Wed Apr 24 23:35:11 2013 [VMM][I]: Successfully execute network driver operation: pre. Wed Apr 24 23:35:22 2013 [VMM][I]: Command execution fail: /srv/cloud/one/var/remotes/vmm/vmware/deploy /srv/cloud/one/var/vms/138/deployment.0 vmware02 138 vmware02 Wed Apr 24 23:35:22 2013 [VMM][D]: deploy: Successfully defined domain one-138. Wed Apr 24 23:35:22 2013 [VMM][E]: deploy: Error executing: virsh -c 'esx://vmware02/?no_verify=1&auto_answer=1' start one-138 err: ExitCode: 1 Wed Apr 24 23:35:22 2013 [VMM][I]: out: Wed Apr 24 23:35:22 2013 [VMM][I]: error: Failed to start domain one-138 Wed Apr 24 23:35:22 2013 [VMM][I]: error: internal error Could not start domain: GenericVmConfigFault - Module DevicePowerOn power on failed. Wed Apr 24 23:35:22 2013 [VMM][I]: Wed Apr 24 23:35:22 2013 [VMM][I]: Wed Apr 24 23:35:22 2013 [VMM][I]: ExitCode: 1 Wed Apr 24 23:35:22 2013 [VMM][I]: Failed to execute virtualization driver operation: deploy. Wed Apr 24 23:35:22 2013 [VMM][E]: Error deploying virtual machine Wed Apr 24 23:35:22 2013 [DiM][I]: New VM state is FAILED Then I looked into vmware.log, found the following information: 2013-04-24T15:35:15.816Z| vmx| Msg_Post: Error 2013-04-24T15:35:15.816Z| vmx| [msg.buslogic.notOn64bit] The BusLogic SCSI adapter is not supported for 64-bit guests. See the documentation for the appropriate type of SCSI adapter to use with 64-bit guests. 2013-04-24T15:35:15.816Z| vmx| [msg.moduletable.powerOnFailed] Module DevicePowerOn power on failed. 2013-04-24T15:35:15.816Z| vmx| ---------------------------------------- 2013-04-24T15:35:15.819Z| vmx| VMX_PowerOn: ModuleTable_PowerOn = 0 2013-04-24T15:35:15.819Z| vmx| MKS REMOTE stopping VNC server at 0.0.0.0:6038 2013-04-24T15:35:15.819Z| vmx| MKS local poweroff 2013-04-24T15:35:15.819Z| vmx| scsi0:0: numIOs = 0 numMergedIOs = 0 numSplitIOs = 0 ( 0.0%) 2013-04-24T15:35:15.819Z| vmx| Closing disk scsi0:0 2013-04-24T15:35:15.823Z| vmx| DISKLIB-VMFS : "/vmfs/volumes/5170f0fb-541ed8d5-abcf-80ee733ae51c/138/disk.0/CentOS-6.2-flat.vmdk" : closed. 2013-04-24T15:35:15.827Z| vmx| WORKER: asyncOps=0 maxActiveOps=0 maxPending=0 maxCompleted=0 2013-04-24T15:35:15.866Z| vmx| Vix: [8359 mainDispatch.c:4084]: VMAutomation_ReportPowerOpFinished: statevar=1, newAppState=1873, success=1 additionalError=0 2013-04-24T15:35:15.866Z| vmx| Vix: [8359 mainDispatch.c:4084]: VMAutomation_ReportPowerOpFinished: statevar=0, newAppState=1870, success=1 additionalError=0 2013-04-24T15:35:15.866Z| vmx| Transitioned vmx/execState/val to poweredOff 2013-04-24T15:35:15.889Z| vmx| Vix: [8359 mainDispatch.c:4084]: VMAutomation_ReportPowerOpFinished: statevar=0, newAppState=1870, success=0 additionalError=0 2013-04-24T15:35:15.889Z| vmx| Vix: [8359 mainDispatch.c:4123]: Error VIX_E_FAIL in VMAutomation_ReportPowerOpFinished(): Unknown error 2013-04-24T15:35:15.889Z| vmx| VMX idle exit 2013-04-24T15:35:15.890Z| vmx| VMIOP: Exit 2013-04-24T15:35:15.905Z| vmx| Vix: [8359 mainDispatch.c:900]: VMAutomation_LateShutdown() 2013-04-24T15:35:15.905Z| vmx| Vix: [8359 mainDispatch.c:850]: VMAutomationCloseListenerSocket. Closing listener socket. 2013-04-24T15:35:15.907Z| vmx| Flushing VMX VMDB connections 2013-04-24T15:35:15.908Z| vmx| VmdbDbRemoveCnx: Removing Cnx from Db for '/db/connection/#1/' 2013-04-24T15:35:15.908Z| vmx| VmdbCnxDisconnect: Disconnect: closed pipe for pub cnx '/db/connection/#1/' (0) 2013-04-24T15:35:15.913Z| vmx| VMX exit (0). 2013-04-24T15:35:15.913Z| vmx| AIOMGR-S : stat o=1 r=3 w=0 i=0 br=49152 bw=0 2013-04-24T15:35:15.913Z| vmx| VMX has left the building: 0. However, in disk.vmdk, we did not used the BusLogic controller. Rather, we did mention LSI Logic: /vmfs/volumes/5170f0fb-541ed8d5-abcf-80ee733ae51c/138/disk.0 # more disk.vmdk # Disk DescriptorFile version=1 encoding="UTF-8" CID=e451587b parentCID=ffffffff isNativeSnapshot="no" createType="vmfs" # Extent description RW 16777216 VMFS "CentOS-6.2-flat.vmdk" # The Disk Data Base #DDB ddb.deletable = "true" ddb.virtualHWVersion = "7" ddb.longContentID = "45ecfa456ae7c14601b32b19e451587b" ddb.uuid = "60 00 C2 96 8f d3 ab c5-36 3a 7a f2 84 f6 52 89" ddb.geometry.cylinders = "1044" ddb.geometry.heads = "255" ddb.geometry.sectors = "63" ddb.thinProvisioned = "1" ddb.adapterType = "lsilogic" Below is what I had in deployment.0: oneadmin@opennebula:~/var/vms/138$ more deployment.0 <domain type='vmware'> <name>one-138</name> <memory>1048576</memory> <os> <type arch='x86_64'>hvm</type> </os> <devices> <disk type='file' device='disk'> <source file='[101] 138/disk.0/disk.vmdk'/> <target dev='sda'/> </disk> <graphics type='vnc' listen='0.0.0.0' port='6038'/> </devices> </domain> Very confused. Chen Xiang > Second try. I looked into disk.vmdk, and found the following: > > # Disk DescriptorFile > version=1 > encoding="UTF-8" > CID=e451587b > parentCID=ffffffff > isNativeSnapshot="no" > createType="vmfs" > > # Extent description > RW 16777216 VMFS "CentOS-6.2.vmdk" > > Obvious I did not have CentOS-6.2.vmdk, but I have CentOS-6.2-flat.vmdk. > So, I manually edited that disk.vmdk, and changed CentOS-6.2.vmdk to > CentOS-6.2-flat.vmdk. Then I instantiated a VM, that worked! > > Now I have CentOS 6.2 running...... > > Chen Xiang > > > >> A lot of new things to learn. After getting ttylinux work with ESXi, I >> headed for CentOS 6.2 (also imported from Marketplace), and it did not >> work as expected. >> >> This time I already have a datastore with DS_MAD="vmware". I imported >> CentOS 6.2 from marketplace, and had the following in that datastore >> directory, which looked alright: >> >> oneadmin@opennebula:~/var/datastores/100/c5610951183fa21ac0900ce147a677b7$ >> ls -l >> total 8388616 >> -rw------- 1 oneadmin cloud 8589934592 Jul 2 2012 CentOS-6.2-flat.vmdk >> -rw------- 1 oneadmin cloud 515 Jul 3 2012 disk.vmdk >> >> Then I instantiated a VM, but it failed to boot (prolog seemed to be >> OK). >> Below is the VM log: >> >> Wed Apr 24 22:40:41 2013 [DiM][I]: New VM state is ACTIVE. >> Wed Apr 24 22:40:41 2013 [LCM][I]: New VM state is PROLOG. >> Wed Apr 24 22:40:41 2013 [VM][I]: Virtual Machine has no context >> Wed Apr 24 22:47:36 2013 [TM][I]: clone: Cloning >> opennebula:/srv/cloud/one/var/datastores/100/c5610951183fa21ac0900ce147a677b7 >> in /vmfs/volumes/101/135/disk.0 >> Wed Apr 24 22:47:36 2013 [TM][I]: ExitCode: 0 >> Wed Apr 24 22:47:37 2013 [LCM][I]: New VM state is BOOT >> Wed Apr 24 22:47:37 2013 [VMM][I]: Generating deployment file: >> /srv/cloud/one/var/vms/135/deployment.1 >> Wed Apr 24 22:47:37 2013 [VMM][I]: ExitCode: 0 >> Wed Apr 24 22:47:37 2013 [VMM][I]: Successfully execute network driver >> operation: pre. >> Wed Apr 24 22:47:47 2013 [VMM][I]: Command execution fail: >> /srv/cloud/one/var/remotes/vmm/vmware/deploy >> /srv/cloud/one/var/vms/135/deployment.1 vmware01 135 vmware01 >> Wed Apr 24 22:47:47 2013 [VMM][D]: deploy: Successfully defined domain >> one-135. >> Wed Apr 24 22:47:47 2013 [VMM][E]: deploy: Error executing: virsh -c >> 'esx://vmware01/?no_verify=1&auto_answer=1' start one-135 err: ExitCode: >> 1 >> Wed Apr 24 22:47:47 2013 [VMM][I]: out: >> Wed Apr 24 22:47:47 2013 [VMM][I]: error: Failed to start domain one-135 >> Wed Apr 24 22:47:47 2013 [VMM][I]: error: internal error Could not start >> domain: FileNotFound - File [101] 135/disk.0/disk.vmdk was not found >> Wed Apr 24 22:47:47 2013 [VMM][I]: >> Wed Apr 24 22:47:47 2013 [VMM][I]: ExitCode: 1 >> Wed Apr 24 22:47:47 2013 [VMM][I]: Failed to execute virtualization >> driver >> operation: deploy. >> Wed Apr 24 22:47:47 2013 [VMM][E]: Error deploying virtual machine >> Wed Apr 24 22:47:47 2013 [DiM][I]: New VM state is FAILED >> >> I double check on the ESXi box, File [101] 135/disk.0/disk.vmdk was >> actually there. >> >> /vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0 # ls -l >> -rw------- 1 oneadmin root 8589934592 Apr 24 14:47 >> CentOS-6.2-flat.vmdk >> -rw------- 1 oneadmin root 515 Apr 24 14:47 disk.vmdk >> -rw-r--r-- 1 root root 0 Apr 24 14:47 >> one-135.vmsd >> -rw-r--r-- 1 root root 875 Apr 24 14:47 >> one-135.vmx >> -rw-r--r-- 1 root root 262 Apr 24 14:47 >> one-135.vmxf >> -rw-r--r-- 1 root root 38580 Apr 24 14:47 vmware.log >> >> So I looked into vmware.log, and found that it was looking for >> CentOS-6.2.vmdk. I did not have this file anywhere. >> >> 2013-04-24T14:47:35.839Z| vmx| DISK: OPEN scsi0:0 >> '/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/disk.vmdk' >> persistent R[] >> 2013-04-24T14:47:35.842Z| vmx| AIOGNRC: Failed to open >> '/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/CentOS-6.2.vmdk' >> : Could not find the file (600000003) (0x2013). >> 2013-04-24T14:47:35.842Z| vmx| DISKLIB-VMFS : >> "/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/CentOS-6.2.vmdk" >> : failed to open (The system cannot find the file specified): >> AIOMgr_Open >> failed. Type 3 >> 2013-04-24T14:47:35.842Z| vmx| DISKLIB-LINK : >> "/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/disk.vmdk" >> : >> failed to open (The system cannot find the file specified). >> 2013-04-24T14:47:35.842Z| vmx| DISKLIB-CHAIN : >> "/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/disk.vmdk" >> : >> failed to open (The system cannot find the file specified). >> 2013-04-24T14:47:35.842Z| vmx| DISKLIB-LIB : Failed to open >> '/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/disk.vmdk' >> with flags 0xa The system cannot find the file specified (25). >> 2013-04-24T14:47:35.842Z| vmx| Msg_Post: Error >> 2013-04-24T14:47:35.842Z| vmx| [msg.disk.fileNotFound] VMware ESX cannot >> find the virtual disk >> "/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/disk.vmdk". >> Verify the path is valid and try again. >> 2013-04-24T14:47:35.842Z| vmx| [msg.disk.noBackEnd] Cannot open the disk >> '/vmfs/volumes/5170ef35-aa800722-6bfa-80ee733ae308/135/disk.0/disk.vmdk' >> or one of the snapshot disks it depends on. >> 2013-04-24T14:47:35.842Z| vmx| [msg.disk.configureDiskError] Reason: The >> system cannot find the file specified. >> >> So, the behavior of the CentOS 6.2 image is different from the ttylinux >> image. What should I do? >> >> >> Chen Xiang >> >> >>> Ok, I see what happened then. Datastores to be used with VMwaqre ESX >>> hosts must be wither DS_MAD="vmware" or DS_MAD="vmfs", otherwise the >>> drivers won't know how to handle VMware vmdk disks. >>> >>> Glad is working now. >>> >>> -T >>> -- >>> Constantino Vázquez Blanco, PhD, MSc >>> Project Engineer >>> OpenNebula - The Open-Source Solution for Data Center Virtualization >>> www.OpenNebula.org | @tinova79 | @OpenNebula >>> >>> >>> On Wed, Apr 24, 2013 at 12:28 PM, <chenxi...@aquala-tech.com> wrote: >>>> The disk.vmdk problem seems to be associated with the DS_MAD setting. >>>> I >>>> created two datastores to test this, the first one with >>>> DS_MAD="vmware" >>>> and the second one with DS_MAD="fs". When I imported the images from >>>> marketplace, in the first datastore I have disk.vmdk, and in the other >>>> one >>>> I have ttylinux.vmdk. >>>> >>>> When I encountered the problem in this thread, I was importing the >>>> ttylinux image into the default datastore, which had DS_MAD="fs". So, >>>> that >>>> caused the problem. >>>> >>>> A lot of things to learn. Thank you so much for your assistance. You >>>> are >>>> so helpful for me. >>>> >>>> Chen Xiang >>>> >>>> >>>>> On Wed, Apr 24, 2013 at 7:39 AM, <chenxi...@aquala-tech.com> wrote: >>>>>> I double checked the VNC issue. I did set the VNC stuff according to >>>>>> the >>>>>> instructions, but that setting got lost across reboots, so I was not >>>>>> able >>>>>> to connect to VNC. >>>>>> >>>>>> I searched on the web, and found a solution. That is to copy the >>>>>> modified >>>>>> version of service.xml to the localdisk of ESXi as a backup, and >>>>>> then >>>>>> modified /etc/rc.local to recover that file from the backup version, >>>>>> and >>>>>> to refresh the firewall settings. (I need to do the same for the >>>>>> oneadmin >>>>>> SSH key, otherwise the front end won't be able to connect to the >>>>>> ESXi >>>>>> box >>>>>> when the box reboots.) Then VNC worked. >>>>> >>>>> Glad you got it working. A similar approach is defined here [1] >>>>> (search for "persistency") >>>>> >>>>> [1] http://opennebula.org/documentation:rel3.8:evmwareg >>>>> >>>>>> >>>>>> Regarding the disk.vmdk. I checked into the oned.log. No I did not >>>>>> find >>>>>> the error message you mentioned. >>>>> >>>>> Can you send us the relevant files in oned.log at the time of >>>>> registering a VMware disk folder _without_ the manual renaming to >>>>> disk.vmdk? >>>>> >>>>> Thanks a lot for your feedback, >>>>> >>>>> -T >>>>> >>>>>> >>>>>> Chen Xiang >>>>>> >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> comments inline, >>>>>>> >>>>>>> On Tue, Apr 23, 2013 at 8:54 AM, <chenxi...@aquala-tech.com> >>>>>>> wrote: >>>>>>>> Now I am able to get VM's running with ESXi nodes. I would like to >>>>>>>> share >>>>>>>> what I did to make things work, and would like you guys to help me >>>>>>>> further. >>>>>>>> >>>>>>>> From Marketplace I imported the ttylinux-vmware image to local >>>>>>>> infrastructure. When the images becomes READY, login to the front >>>>>>>> end >>>>>>>> server and browse to var/datastores/1 and further browse to the >>>>>>>> directory >>>>>>>> holding that particular image. Below was what I had when things >>>>>>>> did >>>>>>>> not >>>>>>>> work: >>>>>>>> >>>>>>>> oneadmin@opennebula:~/var/datastores/1/10fc21f21a452add3838d76d63052457$ >>>>>>>> ls -l >>>>>>>> total 104864 >>>>>>>> -rw------- 1 oneadmin cloud 107374080 Jul 3 2012 >>>>>>>> ttylinux-flat.vmdk >>>>>>>> -rw------- 1 oneadmin cloud 509 Jul 3 2012 ttylinux.vmdk >>>>>>>> >>>>>>>> Base on the error message I got from VM log, I decided to copy >>>>>>>> ttylinux.vmdk to disk.vmdk. So I did that and then I had the >>>>>>>> following: >>>>>>>> >>>>>>>> oneadmin@opennebula:~/var/datastores/1/10fc21f21a452add3838d76d63052457$ >>>>>>>> cp ttylinux.vmdk disk.vmdk >>>>>>>> oneadmin@opennebula:~/var/datastores/1/10fc21f21a452add3838d76d63052457$ >>>>>>>> ls -l >>>>>>>> total 104868 >>>>>>>> -rw------- 1 oneadmin cloud 509 Apr 23 14:33 disk.vmdk >>>>>>>> -rw------- 1 oneadmin cloud 107374080 Jul 3 2012 >>>>>>>> ttylinux-flat.vmdk >>>>>>>> -rw------- 1 oneadmin cloud 509 Jul 3 2012 ttylinux.vmdk >>>>>>>> >>>>>>>> Now I went back to the same template, and instantiated a VM >>>>>>>> instance, >>>>>>>> now >>>>>>>> it worked. >>>>>>> >>>>>>> The vmware/cp script should automatically rename the file. Can you >>>>>>> see >>>>>>> any line in /var/log/one/oned.log similar to: >>>>>>> >>>>>>> -- >>>>>>> Error renaming disk file $BASE_DISK_FILE to disk.vmdk >>>>>>> -- >>>>>>> >>>>>>> >>>>>>>> >>>>>>>> Now I have a new problem. This front-end has been tested for KVM >>>>>>>> with >>>>>>>> hundreds of VM instances before I tried ESXi. Now I am trying >>>>>>>> VM-ID >>>>>>>> above >>>>>>>> 100. Now the VM instances were running, but I was not able to >>>>>>>> connect >>>>>>>> to >>>>>>>> the VM console via VNC from SunStone. What should I do? (VM >>>>>>>> Console >>>>>>>> in >>>>>>>> vSphere Client still works.) >>>>>>> >>>>>>> >>>>>>> Have you configured the ESX host to allow VNC connections? See >>>>>>> http://opennebula.org/documentation:rel3.8:evmwareg#vnc >>>>>>> >>>>>>> Regards, >>>>>>> >>>>>>> -Tino >>>>>>> >>>>>>>> >>>>>>>> Thanks a lot for your assistance. >>>>>>>> >>>>>>>> Chen Xiang >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>> With the proper NFS settings, I was able to define the domain, >>>>>>>>> but >>>>>>>>> fail >>>>>>>>> to >>>>>>>>> create the VM. The error message says "File [0] >>>>>>>>> 120/disk.0/disk.vmdk >>>>>>>>> was >>>>>>>>> not found". >>>>>>>>> >>>>>>>>> I login to the ESXi box, and verified that I do have >>>>>>>>> /vmfs/volumes/0 >>>>>>>>> pointing to the right NFS mount. And inside /vmfs/volumes/0 I do >>>>>>>>> have >>>>>>>>> 120/disk.0/disk.vmdk . >>>>>>>>> >>>>>>>>> Any ideas? >>>>>>>>> >>>>>>>>> Below is a copy of the VM log. >>>>>>>>> >>>>>>>>> Tue Apr 23 13:12:58 2013 [DiM][I]: New VM state is ACTIVE. >>>>>>>>> Tue Apr 23 13:12:58 2013 [LCM][I]: New VM state is PROLOG. >>>>>>>>> Tue Apr 23 13:12:58 2013 [VM][I]: Virtual Machine has no context >>>>>>>>> Tue Apr 23 13:13:12 2013 [TM][I]: clone: Cloning >>>>>>>>> /vmfs/volumes/1/10fc21f21a452add3838d76d63052457 in >>>>>>>>> vmware02:/vmfs/volumes/0/120/disk.0 >>>>>>>>> Tue Apr 23 13:13:12 2013 [TM][I]: ExitCode: 0 >>>>>>>>> Tue Apr 23 13:13:12 2013 [LCM][I]: New VM state is BOOT >>>>>>>>> Tue Apr 23 13:13:12 2013 [VMM][I]: Generating deployment file: >>>>>>>>> /srv/cloud/one/var/vms/120/deployment.0 >>>>>>>>> Tue Apr 23 13:13:12 2013 [VMM][I]: ExitCode: 0 >>>>>>>>> Tue Apr 23 13:13:12 2013 [VMM][I]: Successfully execute network >>>>>>>>> driver >>>>>>>>> operation: pre. >>>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: Command execution fail: >>>>>>>>> /srv/cloud/one/var/remotes/vmm/vmware/deploy >>>>>>>>> /srv/cloud/one/var/vms/120/deployment.0 vmware02 120 vmware02 >>>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][D]: deploy: Successfully defined >>>>>>>>> domain >>>>>>>>> one-120. >>>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][E]: deploy: Error executing: virsh >>>>>>>>> -c >>>>>>>>> 'esx://vmware02/?no_verify=1&auto_answer=1' start one-120 err: >>>>>>>>> ExitCode: 1 >>>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: out: >>>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: error: Failed to start domain >>>>>>>>> one-120 >>>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: error: internal error Could >>>>>>>>> not >>>>>>>>> start >>>>>>>>> domain: FileNotFound - File [0] 120/disk.0/disk.vmdk was not >>>>>>>>> found >>>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: >>>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: ExitCode: 1 >>>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][I]: Failed to execute >>>>>>>>> virtualization >>>>>>>>> driver >>>>>>>>> operation: deploy. >>>>>>>>> Tue Apr 23 13:13:24 2013 [VMM][E]: Error deploying virtual >>>>>>>>> machine >>>>>>>>> Tue Apr 23 13:13:24 2013 [DiM][I]: New VM state is FAILED >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Chen Xiang >>>>>>>>> >>>>>>>>>> Hi, >>>>>>>>>> >>>>>>>>>> Please use root_squash instead of no_root_squash >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> /srv/cloud/one/var/datastores/0 >>>>>>>>>> *(rw,sync,no_subtree_check,no_root_squash,anonuid=10000,anongid=10000) >>>>>>>>>> /srv/cloud/one/var/datastores/1 >>>>>>>>>> *(rw,sync,no_subtree_check,no_root_squash,anonuid=10000,anongid=10000) >>>>>>>>>> -- >>>>>>>>>> >>>>>>>>>> You will need to force the nfs server to re-read the conf file. >>>>>>>>>> >>>>>>>>>> Regards, >>>>>>>>>> >>>>>>>>>> -Tino >>>>>>>>>> -- >>>>>>>>>> Constantino Vázquez Blanco, PhD, MSc >>>>>>>>>> Project Engineer >>>>>>>>>> OpenNebula - The Open-Source Solution for Data Center >>>>>>>>>> Virtualization >>>>>>>>>> www.OpenNebula.org | @tinova79 | @OpenNebula >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Mon, Apr 22, 2013 at 3:46 PM, <chenxi...@aquala-tech.com> >>>>>>>>>> wrote: >>>>>>>>>>> Here is what I got when trying to define the VM via virsh: >>>>>>>>>>> >>>>>>>>>>> oneadmin@opennebula:~/images$ virsh -c >>>>>>>>>>> 'esx://vmware02/?no_verify=1&auto_answer=1' >>>>>>>>>>> Enter username for vmware02 [root]: >>>>>>>>>>> Enter root's password for vmware02: >>>>>>>>>>> Welcome to virsh, the virtualization interactive terminal. >>>>>>>>>>> >>>>>>>>>>> Type: 'help' for help with commands >>>>>>>>>>> 'quit' to quit >>>>>>>>>>> >>>>>>>>>>> virsh # define /srv/cloud/one/var/vms/111/deployment.0 >>>>>>>>>>> 2013-04-22 13:24:49.391+0000: 17332: info : libvirt version: >>>>>>>>>>> 0.9.10 >>>>>>>>>>> 2013-04-22 13:24:49.391+0000: 17332: warning : >>>>>>>>>>> virVMXFormatVNC:3224 >>>>>>>>>>> : >>>>>>>>>>> VNC >>>>>>>>>>> port 6011 it out of [5900..5964] range >>>>>>>>>>> error: Failed to define domain from >>>>>>>>>>> /srv/cloud/one/var/vms/111/deployment.0 >>>>>>>>>>> error: internal error HTTP response code 403 for upload to >>>>>>>>>>> 'https://vmware02:443/folder/111%2fdisk%2e0/one%2d111.vmx?dcPath=ha%2ddatacenter&dsName=0' >>>>>>>>>>> >>>>>>>>>>> virsh # exit >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> Hi, >>>>>>>>>>>> >>>>>>>>>>>> I am running OpenNebula 3.8.3 on Ubuntu 12.04 (front end) with >>>>>>>>>>>> VMWare >>>>>>>>>>>> ESXi >>>>>>>>>>>> 5.0 (node). I am able to create a VMWare node in SunStone, >>>>>>>>>>>> register >>>>>>>>>>>> the >>>>>>>>>>>> ttylinux images (ttylinux.vmdk.tar download from C12G.com), >>>>>>>>>>>> but >>>>>>>>>>>> failed >>>>>>>>>>>> to >>>>>>>>>>>> instantiate a VM. >>>>>>>>>>>> >>>>>>>>>>>> Here is what I have as the NFS exports: >>>>>>>>>>>> >>>>>>>>>>>> /srv/cloud/one/var/datastores/0 >>>>>>>>>>>> *(rw,sync,no_subtree_check,no_root_squash,anonuid=10000,anongid=10000) >>>>>>>>>>>> /srv/cloud/one/var/datastores/1 >>>>>>>>>>>> *(rw,sync,no_subtree_check,no_root_squash,anonuid=10000,anongid=10000) >>>>>>>>>>>> >>>>>>>>>>>> One the ESXi node I mounted the NFS exports to /vmfs/volumes/0 >>>>>>>>>>>> and >>>>>>>>>>>> /vmfs/volumes/1 respectively. >>>>>>>>>>>> >>>>>>>>>>>> Yes I configured libvirt-0.9.10 and make it worked with >>>>>>>>>>>> VMWare, >>>>>>>>>>>> verified >>>>>>>>>>>> by command lines such as the following command (where vmware02 >>>>>>>>>>>> is >>>>>>>>>>>> my >>>>>>>>>>>> ESXi >>>>>>>>>>>> hostname): >>>>>>>>>>>> >>>>>>>>>>>> virsh -c 'esx://vmware02/?no_verify=1&auto_answer=1' >>>>>>>>>>>> >>>>>>>>>>>> Below is my VM template: >>>>>>>>>>>> >>>>>>>>>>>> CPU="1" >>>>>>>>>>>> DISK=[ >>>>>>>>>>>> IMAGE="tty_vmdk", >>>>>>>>>>>> IMAGE_UNAME="oneadmin" ] >>>>>>>>>>>> GRAPHICS=[ >>>>>>>>>>>> LISTEN="0.0.0.0", >>>>>>>>>>>> TYPE="vnc" ] >>>>>>>>>>>> MEMORY="512" >>>>>>>>>>>> NAME="ttylinux" >>>>>>>>>>>> >>>>>>>>>>>> Below is what I got when trying to instantiate a VM: >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Mon Apr 22 21:27:11 2013 [DiM][I]: New VM state is ACTIVE. >>>>>>>>>>>> Mon Apr 22 21:27:11 2013 [LCM][I]: New VM state is PROLOG. >>>>>>>>>>>> Mon Apr 22 21:27:11 2013 [VM][I]: Virtual Machine has no >>>>>>>>>>>> context >>>>>>>>>>>> Mon Apr 22 21:27:17 2013 [TM][I]: clone: Cloning >>>>>>>>>>>> /vmfs/volumes/1/43352fb75cee9bbc1da3c1e7ff474e26 in >>>>>>>>>>>> vmware02:/vmfs/volumes/0/112/disk.0 >>>>>>>>>>>> Mon Apr 22 21:27:17 2013 [TM][I]: ExitCode: 0 >>>>>>>>>>>> Mon Apr 22 21:27:17 2013 [LCM][I]: New VM state is BOOT >>>>>>>>>>>> Mon Apr 22 21:27:17 2013 [VMM][I]: Generating deployment file: >>>>>>>>>>>> /srv/cloud/one/var/vms/112/deployment.0 >>>>>>>>>>>> Mon Apr 22 21:27:17 2013 [VMM][I]: ExitCode: 0 >>>>>>>>>>>> Mon Apr 22 21:27:17 2013 [VMM][I]: Successfully execute >>>>>>>>>>>> network >>>>>>>>>>>> driver >>>>>>>>>>>> operation: pre. >>>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: Command execution fail: >>>>>>>>>>>> /srv/cloud/one/var/remotes/vmm/vmware/deploy >>>>>>>>>>>> /srv/cloud/one/var/vms/112/deployment.0 vmware02 112 vmware02 >>>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][E]: deploy: Error executing: >>>>>>>>>>>> virsh >>>>>>>>>>>> -c >>>>>>>>>>>> 'esx://vmware02/?no_verify=1&auto_answer=1' define >>>>>>>>>>>> /srv/cloud/one/var/vms/112/deployment.0 err: ExitCode: 1 >>>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: out: >>>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: 2013-04-22 >>>>>>>>>>>> 13:27:21.858+0000: >>>>>>>>>>>> 17586: >>>>>>>>>>>> info : libvirt version: 0.9.10 >>>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: 2013-04-22 >>>>>>>>>>>> 13:27:21.858+0000: >>>>>>>>>>>> 17586: >>>>>>>>>>>> warning : virVMXFormatVNC:3224 : VNC port 6012 it out of >>>>>>>>>>>> [5900..5964] >>>>>>>>>>>> range >>>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: error: Failed to define >>>>>>>>>>>> domain >>>>>>>>>>>> from >>>>>>>>>>>> /srv/cloud/one/var/vms/112/deployment.0 >>>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: error: internal error HTTP >>>>>>>>>>>> response >>>>>>>>>>>> code 403 for upload to >>>>>>>>>>>> 'https://vmware02:443/folder/112%2fdisk%2e0/one%2d112.vmx?dcPath=ha%2ddatacenter&dsName=0' >>>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: >>>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: ExitCode: 255 >>>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][I]: Failed to execute >>>>>>>>>>>> virtualization >>>>>>>>>>>> driver >>>>>>>>>>>> operation: deploy. >>>>>>>>>>>> Mon Apr 22 21:27:22 2013 [VMM][E]: Error deploying virtual >>>>>>>>>>>> machine >>>>>>>>>>>> Mon Apr 22 21:27:22 2013 [DiM][I]: New VM state is FAILED >>>>>>>>>>>> >>>>>>>>>>>> What might be wrong? This looks like that I did not have write >>>>>>>>>>>> access >>>>>>>>>>>> to >>>>>>>>>>>> some place. So I tried both the oneadmin and root acount in >>>>>>>>>>>> etc/vmwarerc, >>>>>>>>>>>> with not much luck. >>>>>>>>>>>> >>>>>>>>>>>> On the front end the oneadmin user belongs to the following >>>>>>>>>>>> groups: >>>>>>>>>>>> cloud, >>>>>>>>>>>> adm, sudo, libvirtd. >>>>>>>>>>>> >>>>>>>>>>>> Best regards, >>>>>>>>>>>> >>>>>>>>>>>> Chen Xiang >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Users mailing list >>>>>>>>>>>> Users@lists.opennebula.org >>>>>>>>>>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Users mailing list >>>>>>>>>>> Users@lists.opennebula.org >>>>>>>>>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org >>>>>>>>>> >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> Users mailing list >>>>>>>>> Users@lists.opennebula.org >>>>>>>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org >>>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Users mailing list >>>>>>>> Users@lists.opennebula.org >>>>>>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org >>>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Users mailing list >>>>>> Users@lists.opennebula.org >>>>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org >>>>> >>>> >>>> >>>> _______________________________________________ >>>> Users mailing list >>>> Users@lists.opennebula.org >>>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org >>> >> >> _______________________________________________ >> Users mailing list >> Users@lists.opennebula.org >> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org >> > > > _______________________________________________ > Users mailing list > Users@lists.opennebula.org > http://lists.opennebula.org/listinfo.cgi/users-opennebula.org > _______________________________________________ Users mailing list Users@lists.opennebula.org http://lists.opennebula.org/listinfo.cgi/users-opennebula.org