OK, Marcus, I made the change to detect my volumes and it seems to work
just fine.

Perhaps another day of testing and we can check this code in. What do you
think?


On Wed, Oct 23, 2013 at 9:14 PM, Mike Tutkowski <
mike.tutkow...@solidfire.com> wrote:

> Thanks, Marcus...I hadn't read that note, but that makes sense.
>
> Yes, that must be the root disk for the VM. I can put in code, as you
> recommend, to handle only my volumes.
>
>
> On Wed, Oct 23, 2013 at 5:37 PM, Marcus Sorensen <shadow...@gmail.com>wrote:
>
>> It should be sending the path info for each disk per the XML of the
>> VM... so it will send all disks regardless of whether or not your
>> adaptor manages that disk, and it's up to your adaptor to ignore any
>> that aren't managed by it. There should be notes to that effect in the
>> code near the disconnectPhysicalDisk interface in StorageAdaptor:
>>
>>     // given local path to file/device (per Libvirt XML), 1) check
>> that device is
>>     // handled by your adaptor, return false if not. 2) clean up
>> device, return true
>>     public boolean disconnectPhysicalDiskByPath(String localPath);
>>
>> Since we only have XML disk definitions when we stop or migrate a VM,
>> we have to try all adaptors against all defined disks. So in your
>> disconnectPhysicalDisk you might do something like check that the path
>> starts with '/dev/disk/by-path' and contains 'iscs-iqn' (maybe there's
>> some way that's more robust like checking the full path against a lun
>> listing or something). If it doesn't match, then your
>> disconnectPhysicalDisk just does nothing.
>>
>> I assume this is a root disk or some other local storage disk. If it's
>> not, then your VM XML is messed up somehow.
>>
>> On Wed, Oct 23, 2013 at 5:01 PM, Mike Tutkowski
>> <mike.tutkow...@solidfire.com> wrote:
>> > I found the problem.
>> >
>> > disconnectPhysicalDiskByPath is being passed in (in my situation) the
>> > following:
>> >
>> > /var/lib/libvirt/images/9887d511-8dc7-4cb4-96f9-01230fe4bbb6
>> >
>> > Due to the name of the method, my code was expecting data such as the
>> > following:
>> >
>> >
>> /dev/disk/by-path/ip-192.168.233.10:3260-iscsi-iqn.2012-03.com.solidfire:volume1-lun-0
>> >
>> > Was it intentional to send the data into this method in the current way?
>> >
>> >
>> > On Wed, Oct 23, 2013 at 1:57 PM, Mike Tutkowski
>> > <mike.tutkow...@solidfire.com> wrote:
>> >>
>> >> You know, I forgot we supposed to be doing that! :) Multi-tasking too
>> much
>> >> today, I guess.
>> >>
>> >> Anyways, it must not be working because I still had a hypervisor
>> >> connection after I shut down the VM.
>> >>
>> >> Let me investigate.
>> >>
>> >>
>> >> On Wed, Oct 23, 2013 at 1:48 PM, Marcus Sorensen <shadow...@gmail.com>
>> >> wrote:
>> >>>
>> >>> Are we not disconnecting when we stop the vm? There's a method for
>> it, we
>> >>> should be. disconnectPhysicalDiskViaVmSpec
>> >>>
>> >>> On Oct 23, 2013 1:28 PM, "Mike Tutkowski" <
>> mike.tutkow...@solidfire.com>
>> >>> wrote:
>> >>>>
>> >>>> I see one problem for us now, Marcus.
>> >>>>
>> >>>> * You have a running VM that you attach a volume to.
>> >>>> * You stop the VM.
>> >>>> * You detach the volume.
>> >>>> * You start up the VM.
>> >>>>
>> >>>> The VM will not be connected to the volume (which is good), but the
>> >>>> hypervisor will still be connected to the volume.
>> >>>>
>> >>>> It would be great if we actually sent a command to the last host ID
>> of
>> >>>> the stopped VM when detaching a volume (to have the hypervisor
>> disconnect
>> >>>> from the volume).
>> >>>>
>> >>>> What do you think about that?
>> >>>>
>> >>>>
>> >>>> On Wed, Oct 23, 2013 at 1:15 PM, Mike Tutkowski
>> >>>> <mike.tutkow...@solidfire.com> wrote:
>> >>>>>
>> >>>>> OK, whatever way you prefer then, Marcus (createVdb first or
>> second).
>> >>>>>
>> >>>>> If I leave createVdb first and return 0, it does seem to work.
>> >>>>>
>> >>>>>
>> >>>>> On Wed, Oct 23, 2013 at 11:13 AM, Marcus Sorensen <
>> shadow...@gmail.com>
>> >>>>> wrote:
>> >>>>>>
>> >>>>>> I think we could flip-flop these two lines if necessary:
>> >>>>>>
>> >>>>>>             createVbd(conn, vmSpec, vmName, vm);
>> >>>>>>
>> >>>>>>             _storagePoolMgr.connectPhysicalDisksViaVmSpec(vmSpec);
>> >>>>>>
>> >>>>>> I haven't actually tried it though. But in general I don't see the
>> >>>>>> Libvirt DiskDef using size at all, which is what createVbd does
>> >>>>>> (creates XML definitions for disks to attach to the VM
>> definition). It
>> >>>>>> just takes the device at it's native advertised size when it
>> actually
>> >>>>>> goes to use it.
>> >>>>>>
>> >>>>>> On Wed, Oct 23, 2013 at 10:52 AM, Mike Tutkowski
>> >>>>>> <mike.tutkow...@solidfire.com> wrote:
>> >>>>>> > Little problem that I wanted to get your take on, Marcus.
>> >>>>>> >
>> >>>>>> > When a VM is being started, we call createVdb before calling
>> >>>>>> > connectPhysicalDisksViaVmSpec.
>> >>>>>> >
>> >>>>>> > The problem is that createVdb calls getPhysicalDisk and my volume
>> >>>>>> > has not
>> >>>>>> > yet been connected because connectPhysicalDisksViaVmSpec has not
>> yet
>> >>>>>> > been
>> >>>>>> > called.
>> >>>>>> >
>> >>>>>> > When I try to read up the size of the disk to populate a
>> >>>>>> > PhysicalDisk, I get
>> >>>>>> > an error, of course, because the path does not yet exist.
>> >>>>>> >
>> >>>>>> > I could populate a 0 for the size of the physical disk and then
>> the
>> >>>>>> > next
>> >>>>>> > time getPhysicalDisk is called, it should be filled in with a
>> proper
>> >>>>>> > size.
>> >>>>>> >
>> >>>>>> > Do you see a problem with that approach?
>> >>>>>> >
>> >>>>>> >
>> >>>>>> > On Tue, Oct 22, 2013 at 6:40 PM, Marcus Sorensen
>> >>>>>> > <shadow...@gmail.com>
>> >>>>>> > wrote:
>> >>>>>> >>
>> >>>>>> >> That's right. All should be well.
>> >>>>>> >>
>> >>>>>> >> On Oct 22, 2013 6:03 PM, "Mike Tutkowski"
>> >>>>>> >> <mike.tutkow...@solidfire.com>
>> >>>>>> >> wrote:
>> >>>>>> >>>
>> >>>>>> >>> Looks like we disconnect physical disks when the VM is stopped.
>> >>>>>> >>>
>> >>>>>> >>> I didn't see that before.
>> >>>>>> >>>
>> >>>>>> >>> I suppose that means the disks are physically disconnected when
>> >>>>>> >>> the VM is
>> >>>>>> >>> stopped, but the CloudStack DB still has the VM associated with
>> >>>>>> >>> the disks
>> >>>>>> >>> for the next time the VM may be started up (unless someone
>> does a
>> >>>>>> >>> disconnect
>> >>>>>> >>> while the VM is in the Stopped State).
>> >>>>>> >>>
>> >>>>>> >>>
>> >>>>>> >>> On Tue, Oct 22, 2013 at 4:19 PM, Mike Tutkowski
>> >>>>>> >>> <mike.tutkow...@solidfire.com> wrote:
>> >>>>>> >>>>
>> >>>>>> >>>> Hey Marcus,
>> >>>>>> >>>>
>> >>>>>> >>>> Quick question for you related to attaching/detaching volumes
>> >>>>>> >>>> when the
>> >>>>>> >>>> VM is in the Stopped State.
>> >>>>>> >>>>
>> >>>>>> >>>> If I detach a volume from a VM that is in the Stopped State,
>> the
>> >>>>>> >>>> DB
>> >>>>>> >>>> seems to get updated, but I don't see a command going to the
>> KVM
>> >>>>>> >>>> hypervisor
>> >>>>>> >>>> that leads to the removal of the iSCSI target.
>> >>>>>> >>>>
>> >>>>>> >>>> It seems the iSCSI target is only removed the next time the
>> VM is
>> >>>>>> >>>> started.
>> >>>>>> >>>>
>> >>>>>> >>>> Do you know if this is true?
>> >>>>>> >>>>
>> >>>>>> >>>> If it is, I'm concerned that the volume could be attached to
>> >>>>>> >>>> another VM
>> >>>>>> >>>> before the Stopped VM is re-started and when the Stopped VM
>> gets
>> >>>>>> >>>> restarted
>> >>>>>> >>>> that it would disconnect the iSCSI volume from underneath the
>> VM
>> >>>>>> >>>> that now
>> >>>>>> >>>> has the volume attached.
>> >>>>>> >>>>
>> >>>>>> >>>> I still want to perform some tests on this, but am first
>> trying
>> >>>>>> >>>> to get a
>> >>>>>> >>>> VM to start after I've attached a volume to it when it was in
>> the
>> >>>>>> >>>> Stopped
>> >>>>>> >>>> State.
>> >>>>>> >>>>
>> >>>>>> >>>> Thanks,
>> >>>>>> >>>> Mike
>> >>>>>> >>>>
>> >>>>>> >>>>
>> >>>>>> >>>> On Mon, Oct 21, 2013 at 10:57 PM, Mike Tutkowski
>> >>>>>> >>>> <mike.tutkow...@solidfire.com> wrote:
>> >>>>>> >>>>>
>> >>>>>> >>>>> Thanks for that info, Marcus.
>> >>>>>> >>>>>
>> >>>>>> >>>>> By the way, I wanted to see if I could attach my volume to a
>> VM
>> >>>>>> >>>>> in the
>> >>>>>> >>>>> Stopped State.
>> >>>>>> >>>>>
>> >>>>>> >>>>> The attach logic didn't trigger any exceptions; however,
>> when I
>> >>>>>> >>>>> started
>> >>>>>> >>>>> the VM, I received an Insufficient Capacity exception.
>> >>>>>> >>>>>
>> >>>>>> >>>>> If I detach the volume and then start the VM, the VM starts
>> just
>> >>>>>> >>>>> fine.
>> >>>>>> >>>>>
>> >>>>>> >>>>> I noticed a problem here (in StoragePoolHostDaoImpl):
>> >>>>>> >>>>>
>> >>>>>> >>>>>     @Override
>> >>>>>> >>>>>
>> >>>>>> >>>>>     public StoragePoolHostVO findByPoolHost(long poolId, long
>> >>>>>> >>>>> hostId) {
>> >>>>>> >>>>>
>> >>>>>> >>>>>         SearchCriteria<StoragePoolHostVO> sc =
>> >>>>>> >>>>> PoolHostSearch.create();
>> >>>>>> >>>>>
>> >>>>>> >>>>>         sc.setParameters("pool_id", poolId);
>> >>>>>> >>>>>
>> >>>>>> >>>>>         sc.setParameters("host_id", hostId);
>> >>>>>> >>>>>
>> >>>>>> >>>>>         return findOneIncludingRemovedBy(sc);
>> >>>>>> >>>>>
>> >>>>>> >>>>>     }
>> >>>>>> >>>>>
>> >>>>>> >>>>> The findOneIncludingRemovedBy method returns null (the
>> poolId is
>> >>>>>> >>>>> my
>> >>>>>> >>>>> storage pool's ID and the hostId is the expected host ID).
>> >>>>>> >>>>>
>> >>>>>> >>>>> I'm not sure what this method is trying to do.
>> >>>>>> >>>>>
>> >>>>>> >>>>> I looked in the storage_pool_host_ref table (is that the
>> correct
>> >>>>>> >>>>> table?) and it only has one row, which maps the local storage
>> >>>>>> >>>>> pool of the
>> >>>>>> >>>>> KVM host to the KVM host (which explains why no match is
>> found
>> >>>>>> >>>>> for my
>> >>>>>> >>>>> situation).
>> >>>>>> >>>>>
>> >>>>>> >>>>> Do you understand what this logic is trying to do?
>> >>>>>> >>>>>
>> >>>>>> >>>>> Thanks!
>> >>>>>> >>>>>
>> >>>>>> >>>>>
>> >>>>>> >>>>>
>> >>>>>> >>>>> On Mon, Oct 21, 2013 at 8:08 PM, Marcus Sorensen
>> >>>>>> >>>>> <shadow...@gmail.com>
>> >>>>>> >>>>> wrote:
>> >>>>>> >>>>>>
>> >>>>>> >>>>>> Do you have the capability to clone the root disk? Normally
>> the
>> >>>>>> >>>>>> template is installed to primary, and then cloned for each
>> root
>> >>>>>> >>>>>> disk.
>> >>>>>> >>>>>> In some cases (such as CLVM), this isn't efficient and so
>> the
>> >>>>>> >>>>>> template
>> >>>>>> >>>>>> is copied fresh to populate each root disk.
>> >>>>>> >>>>>>
>> >>>>>> >>>>>> I'm actually not 100% sure how this works in the new code.
>> It
>> >>>>>> >>>>>> used to
>> >>>>>> >>>>>> be handled by copyPhysicalDisk in the storage adaptor,
>> called
>> >>>>>> >>>>>> by
>> >>>>>> >>>>>> copyTemplateToPrimaryStorage, which runs on the agent. It
>> would
>> >>>>>> >>>>>> pass
>> >>>>>> >>>>>> template/secondary storage info, and the destination
>> >>>>>> >>>>>> volume/primary
>> >>>>>> >>>>>> storage info, and copyPhysicalDisk would do the work of
>> >>>>>> >>>>>> installing the
>> >>>>>> >>>>>> image to the destination.  Then subsequent root disks would
>> be
>> >>>>>> >>>>>> cloned
>> >>>>>> >>>>>> in CreateCommand by calling createDiskFromTemplate.
>> >>>>>> >>>>>>
>> >>>>>> >>>>>> In master it looks like this was moved to
>> KVMStorageProcessor
>> >>>>>> >>>>>> 'cloneVolumeFromBaseTemplate', although I think this just
>> takes
>> >>>>>> >>>>>> over
>> >>>>>> >>>>>> as default, and there's something in your storage driver
>> that
>> >>>>>> >>>>>> should
>> >>>>>> >>>>>> be capable of cloning templates on the mgmt server side. I'm
>> >>>>>> >>>>>> less sure
>> >>>>>> >>>>>> about how the template gets to primary storage in the first
>> >>>>>> >>>>>> place, I
>> >>>>>> >>>>>> assume copyTemplateToPrimaryStorage in KVMStorageProcessor
>> >>>>>> >>>>>> calling
>> >>>>>> >>>>>> copyPhysicalDisk in your adaptor. It's a bit tough for me to
>> >>>>>> >>>>>> tell
>> >>>>>> >>>>>> since our earlier storage adaptor did everything on the
>> host it
>> >>>>>> >>>>>> mostly
>> >>>>>> >>>>>> just worked with the default stuff.
>> >>>>>> >>>>>>
>> >>>>>> >>>>>> On Mon, Oct 21, 2013 at 4:49 PM, Mike Tutkowski
>> >>>>>> >>>>>> <mike.tutkow...@solidfire.com> wrote:
>> >>>>>> >>>>>> > Hey Marcus,
>> >>>>>> >>>>>> >
>> >>>>>> >>>>>> > So...now that this works well for data disks, I was
>> wondering
>> >>>>>> >>>>>> > what
>> >>>>>> >>>>>> > might be
>> >>>>>> >>>>>> > involved in getting this process to work for root disks.
>> >>>>>> >>>>>> >
>> >>>>>> >>>>>> > Can you point me in the right direction as far as what
>> gets
>> >>>>>> >>>>>> > invoked
>> >>>>>> >>>>>> > when a
>> >>>>>> >>>>>> > VM is being created on KVM (so that its root disk can be
>> >>>>>> >>>>>> > created and
>> >>>>>> >>>>>> > the
>> >>>>>> >>>>>> > necessary template laid down or ISO installed)?
>> >>>>>> >>>>>> >
>> >>>>>> >>>>>> > Thanks!
>> >>>>>> >>>>>> >
>> >>>>>> >>>>>> >
>> >>>>>> >>>>>> > On Mon, Oct 21, 2013 at 1:14 PM, Mike Tutkowski
>> >>>>>> >>>>>> > <mike.tutkow...@solidfire.com> wrote:
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >> Hey Marcus,
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >> Just wanted to let you know the branch of mine that has
>> your
>> >>>>>> >>>>>> >> code
>> >>>>>> >>>>>> >> and mine
>> >>>>>> >>>>>> >> appears to work well with regards to attaching a data
>> disk
>> >>>>>> >>>>>> >> to a
>> >>>>>> >>>>>> >> running VM:
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >> fdisk -l from hypervisor:
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >> http://i.imgur.com/NkP5fo0.png
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >> fdisk -l from within VM:
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >> http://i.imgur.com/8YwiiC7.png
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >> I plan to do more testing on this over the coming days.
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >> If all goes well, perhaps we can check this code in by
>> the
>> >>>>>> >>>>>> >> end of
>> >>>>>> >>>>>> >> the
>> >>>>>> >>>>>> >> week?
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >> Talk to you later,
>> >>>>>> >>>>>> >> Mike
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >> On Sun, Oct 20, 2013 at 10:23 PM, Mike Tutkowski
>> >>>>>> >>>>>> >> <mike.tutkow...@solidfire.com> wrote:
>> >>>>>> >>>>>> >>>
>> >>>>>> >>>>>> >>> Don't ask me, but it works now (I've been having this
>> >>>>>> >>>>>> >>> trouble
>> >>>>>> >>>>>> >>> quite a
>> >>>>>> >>>>>> >>> while today).
>> >>>>>> >>>>>> >>>
>> >>>>>> >>>>>> >>> I guess the trick is to send you an e-mail. :)
>> >>>>>> >>>>>> >>>
>> >>>>>> >>>>>> >>>
>> >>>>>> >>>>>> >>> On Sun, Oct 20, 2013 at 10:05 PM, Marcus Sorensen
>> >>>>>> >>>>>> >>> <shadow...@gmail.com>
>> >>>>>> >>>>>> >>> wrote:
>> >>>>>> >>>>>> >>>>
>> >>>>>> >>>>>> >>>> Did you create a service offering that uses local
>> storage,
>> >>>>>> >>>>>> >>>> or add
>> >>>>>> >>>>>> >>>> a
>> >>>>>> >>>>>> >>>> shared primary storage? By default there is no storage
>> >>>>>> >>>>>> >>>> that
>> >>>>>> >>>>>> >>>> matches the
>> >>>>>> >>>>>> >>>> built in offerings.
>> >>>>>> >>>>>> >>>>
>> >>>>>> >>>>>> >>>> On Oct 20, 2013 9:39 PM, "Mike Tutkowski"
>> >>>>>> >>>>>> >>>> <mike.tutkow...@solidfire.com>
>> >>>>>> >>>>>> >>>> wrote:
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>> Hey Marcus,
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>> So, I went back to the branch of mine that has your
>> code
>> >>>>>> >>>>>> >>>>> and
>> >>>>>> >>>>>> >>>>> mine and
>> >>>>>> >>>>>> >>>>> was able to create a new CloudStack install from
>> scratch
>> >>>>>> >>>>>> >>>>> with it
>> >>>>>> >>>>>> >>>>> (once
>> >>>>>> >>>>>> >>>>> again, after manually deleting what was in
>> >>>>>> >>>>>> >>>>> /var/lib/libvirt/images to the
>> >>>>>> >>>>>> >>>>> get system VMs to start).
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>> Anyways, my system VMs are running now and I tried to
>> >>>>>> >>>>>> >>>>> kick off a
>> >>>>>> >>>>>> >>>>> VM
>> >>>>>> >>>>>> >>>>> using the CentOS 6.3 image you provided me a while
>> back.
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>> The virtual router has a Status of Running; however,
>> my
>> >>>>>> >>>>>> >>>>> VM fails
>> >>>>>> >>>>>> >>>>> to
>> >>>>>> >>>>>> >>>>> start (with the generic message of Insufficient
>> >>>>>> >>>>>> >>>>> Capacity).
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>> I've not seen this exception before (related to the
>> VR).
>> >>>>>> >>>>>> >>>>> Do you
>> >>>>>> >>>>>> >>>>> have
>> >>>>>> >>>>>> >>>>> any insight into this?:
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>> com.cloud.exception.ResourceUnavailableException:
>> >>>>>> >>>>>> >>>>> Resource
>> >>>>>> >>>>>> >>>>> [Pod:1] is
>> >>>>>> >>>>>> >>>>> unreachable: Unable to apply userdata and password
>> entry
>> >>>>>> >>>>>> >>>>> on
>> >>>>>> >>>>>> >>>>> router
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> com.cloud.network.router.VirtualNetworkApplianceManagerImpl.applyRules(VirtualNetworkApplianceManagerImpl.java:3793)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> com.cloud.network.router.VirtualNetworkApplianceManagerImpl.applyUserData(VirtualNetworkApplianceManagerImpl.java:3017)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> com.cloud.network.element.VirtualRouterElement.addPasswordAndUserdata(VirtualRouterElement.java:933)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.engine.orchestration.NetworkOrchestrator.prepareElement(NetworkOrchestrator.java:1172)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.engine.orchestration.NetworkOrchestrator.prepareNic(NetworkOrchestrator.java:1288)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.engine.orchestration.NetworkOrchestrator.prepare(NetworkOrchestrator.java:1224)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> com.cloud.vm.VirtualMachineManagerImpl.advanceStart(VirtualMachineManagerImpl.java:826)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> com.cloud.vm.VirtualMachineManagerImpl.start(VirtualMachineManagerImpl.java:508)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.engine.cloud.entity.api.VMEntityManagerImpl.deployVirtualMachine(VMEntityManagerImpl.java:227)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.engine.cloud.entity.api.VirtualMachineEntityImpl.deploy(VirtualMachineEntityImpl.java:209)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> com.cloud.vm.UserVmManagerImpl.startVirtualMachine(UserVmManagerImpl.java:3338)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> com.cloud.vm.UserVmManagerImpl.startVirtualMachine(UserVmManagerImpl.java:2919)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> com.cloud.vm.UserVmManagerImpl.startVirtualMachine(UserVmManagerImpl.java:2905)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> com.cloud.utils.component.ComponentInstantiationPostProcessor$InterceptorDispatcher.intercept(ComponentInstantiationPostProcessor.java:125)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.api.command.user.vm.DeployVMCmd.execute(DeployVMCmd.java:421)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> com.cloud.api.ApiDispatcher.dispatch(ApiDispatcher.java:161)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> com.cloud.api.ApiAsyncJobDispatcher.runJobInContext(ApiAsyncJobDispatcher.java:109)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> com.cloud.api.ApiAsyncJobDispatcher$1.run(ApiAsyncJobDispatcher.java:66)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:56)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.managed.context.impl.DefaultManagedContext.callWithContext(DefaultManagedContext.java:103)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.managed.context.impl.DefaultManagedContext.runWithContext(DefaultManagedContext.java:53)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> com.cloud.api.ApiAsyncJobDispatcher.runJob(ApiAsyncJobDispatcher.java:63)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.framework.jobs.impl.AsyncJobManagerImpl$1.runInContext(AsyncJobManagerImpl.java:532)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:56)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.managed.context.impl.DefaultManagedContext.callWithContext(DefaultManagedContext.java:103)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.managed.context.impl.DefaultManagedContext.runWithContext(DefaultManagedContext.java:53)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> org.apache.cloudstack.managed.context.ManagedContextRunnable.run(ManagedContextRunnable.java:46)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> java.util.concurrent.FutureTask.run(FutureTask.java:262)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>> >>>>>> >>>>>> >>>>> at
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>>
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>> >>>>>> >>>>>> >>>>> at java.lang.Thread.run(Thread.java:724)
>> >>>>>> >>>>>> >>>>>
>> >>>>>> >>>>>> >>>>> Thanks!
>> >>>>>> >>>>>> >>>
>> >>>>>> >>>>>> >>>
>> >>>>>> >>>>>> >>>
>> >>>>>> >>>>>> >>>
>> >>>>>> >>>>>> >>> --
>> >>>>>> >>>>>> >>> Mike Tutkowski
>> >>>>>> >>>>>> >>> Senior CloudStack Developer, SolidFire Inc.
>> >>>>>> >>>>>> >>> e: mike.tutkow...@solidfire.com
>> >>>>>> >>>>>> >>> o: 303.746.7302
>> >>>>>> >>>>>> >>> Advancing the way the world uses the cloud™
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >>
>> >>>>>> >>>>>> >> --
>> >>>>>> >>>>>> >> Mike Tutkowski
>> >>>>>> >>>>>> >> Senior CloudStack Developer, SolidFire Inc.
>> >>>>>> >>>>>> >> e: mike.tutkow...@solidfire.com
>> >>>>>> >>>>>> >> o: 303.746.7302
>> >>>>>> >>>>>> >> Advancing the way the world uses the cloud™
>> >>>>>> >>>>>> >
>> >>>>>> >>>>>> >
>> >>>>>> >>>>>> >
>> >>>>>> >>>>>> >
>> >>>>>> >>>>>> > --
>> >>>>>> >>>>>> > Mike Tutkowski
>> >>>>>> >>>>>> > Senior CloudStack Developer, SolidFire Inc.
>> >>>>>> >>>>>> > e: mike.tutkow...@solidfire.com
>> >>>>>> >>>>>> > o: 303.746.7302
>> >>>>>> >>>>>> > Advancing the way the world uses the cloud™
>> >>>>>> >>>>>
>> >>>>>> >>>>>
>> >>>>>> >>>>>
>> >>>>>> >>>>>
>> >>>>>> >>>>> --
>> >>>>>> >>>>> Mike Tutkowski
>> >>>>>> >>>>> Senior CloudStack Developer, SolidFire Inc.
>> >>>>>> >>>>> e: mike.tutkow...@solidfire.com
>> >>>>>> >>>>> o: 303.746.7302
>> >>>>>> >>>>> Advancing the way the world uses the cloud™
>> >>>>>> >>>>
>> >>>>>> >>>>
>> >>>>>> >>>>
>> >>>>>> >>>>
>> >>>>>> >>>> --
>> >>>>>> >>>> Mike Tutkowski
>> >>>>>> >>>> Senior CloudStack Developer, SolidFire Inc.
>> >>>>>> >>>> e: mike.tutkow...@solidfire.com
>> >>>>>> >>>> o: 303.746.7302
>> >>>>>> >>>> Advancing the way the world uses the cloud™
>> >>>>>> >>>
>> >>>>>> >>>
>> >>>>>> >>>
>> >>>>>> >>>
>> >>>>>> >>> --
>> >>>>>> >>> Mike Tutkowski
>> >>>>>> >>> Senior CloudStack Developer, SolidFire Inc.
>> >>>>>> >>> e: mike.tutkow...@solidfire.com
>> >>>>>> >>> o: 303.746.7302
>> >>>>>> >>> Advancing the way the world uses the cloud™
>> >>>>>> >
>> >>>>>> >
>> >>>>>> >
>> >>>>>> >
>> >>>>>> > --
>> >>>>>> > Mike Tutkowski
>> >>>>>> > Senior CloudStack Developer, SolidFire Inc.
>> >>>>>> > e: mike.tutkow...@solidfire.com
>> >>>>>> > o: 303.746.7302
>> >>>>>> > Advancing the way the world uses the cloud™
>> >>>>>
>> >>>>>
>> >>>>>
>> >>>>>
>> >>>>> --
>> >>>>> Mike Tutkowski
>> >>>>> Senior CloudStack Developer, SolidFire Inc.
>> >>>>> e: mike.tutkow...@solidfire.com
>> >>>>> o: 303.746.7302
>> >>>>> Advancing the way the world uses the cloud™
>> >>>>
>> >>>>
>> >>>>
>> >>>>
>> >>>> --
>> >>>> Mike Tutkowski
>> >>>> Senior CloudStack Developer, SolidFire Inc.
>> >>>> e: mike.tutkow...@solidfire.com
>> >>>> o: 303.746.7302
>> >>>> Advancing the way the world uses the cloud™
>> >>
>> >>
>> >>
>> >>
>> >> --
>> >> Mike Tutkowski
>> >> Senior CloudStack Developer, SolidFire Inc.
>> >> e: mike.tutkow...@solidfire.com
>> >> o: 303.746.7302
>> >> Advancing the way the world uses the cloud™
>> >
>> >
>> >
>> >
>> > --
>> > Mike Tutkowski
>> > Senior CloudStack Developer, SolidFire Inc.
>> > e: mike.tutkow...@solidfire.com
>> > o: 303.746.7302
>> > Advancing the way the world uses the cloud™
>>
>
>
>
> --
> *Mike Tutkowski*
> *Senior CloudStack Developer, SolidFire Inc.*
> e: mike.tutkow...@solidfire.com
> o: 303.746.7302
> Advancing the way the world uses the 
> cloud<http://solidfire.com/solution/overview/?video=play>
> *™*
>



-- 
*Mike Tutkowski*
*Senior CloudStack Developer, SolidFire Inc.*
e: mike.tutkow...@solidfire.com
o: 303.746.7302
Advancing the way the world uses the
cloud<http://solidfire.com/solution/overview/?video=play>
*™*

Reply via email to