Let me clarify this line a bit:

"We get away without this with XenServer and VMware because - as far as I
know - CS delegates HA and live migration to those clusters and they handle
it most likely with some kind of locking protocol on the SR/datastore."

When I set up a XenServer or a VMware cluster, all nodes in the cluster
have the proper CHAP credentials and can access a shared SR/datastore.

HA and live migrations are OK here because the cluster controls access to
the VDI on the SR (or VMDK on the datastore) with some kind of locking
protocol, I expect.

Since KVM isn't really in a cluster outside of the CloudStack world,
CloudStack has to handle these intricacies.

In my case, I'm just presenting a raw disk to a VM on a KVM host.

In that case, HA and live migration depend on the storage plug-in being
able to grant and revoke access to the volume for hosts as needed.

I'd actually rather not even bother with CHAP when using KVM.


On Fri, Sep 27, 2013 at 12:06 AM, Mike Tutkowski <
mike.tutkow...@solidfire.com> wrote:

> Hey Marcus,
>
> I agree that CHAP does not fulfill the same role as fencing.
>
> I think we're going to have trouble with HA and live migrations on KVM if
> the storage plug-in doesn't have a way of knowing when a host wants to
> access a volume and when we want to revoke access to that volume.
>
> We get away without this with XenServer and VMware because - as far as I
> know - CS delegates HA and live migration to those clusters and they handle
> it most likely with some kind of locking protocol on the SR/datastore.
>
> As far as the path field is concerned, I should be able to populate it
> with the IQN of the volume in question.
>
> One problem I do see, however, is in the getPhysicalDisk method.
>
> How are you envisioning I keep track of KVMPhysicalDisks that I create in
> my connect method?
>
> Initially I was thinking I'd just keep them in a map. Storage pool UUID to
> KVMPhysicalDisks.
>
> The problem is, how do I reconstruct that map if the agent is restarted
> (say the host crashes or is restarted).
>
> For storage pools, we get a message (ModifyStoragePoolCommand) from the CS
> MS to tell us about all of the relevant storage pools. With this message, I
> can reconstruct my cache of storage pools. No problem there.
>
> But how will I know which volumes belong to a given storage pool if I have
> to rebuild that map? How will I even know which volumes are in use at all?
>
> Thanks
>
>
> On Thu, Sep 26, 2013 at 11:37 PM, Marcus Sorensen <shadow...@gmail.com>wrote:
>
>> On Thu, Sep 26, 2013 at 10:23 PM, Mike Tutkowski
>> <mike.tutkow...@solidfire.com> wrote:
>> > My comments are inline:
>> >
>> >
>> > On Thu, Sep 26, 2013 at 9:10 PM, Marcus Sorensen <shadow...@gmail.com
>> >wrote:
>> >
>> >> Ok, let me digest this a bit. I got the github responses but I'd also
>> >> like to keep it on-list as well.
>> >>
>> >>  My initial thoughts are:
>> >>
>> >> 1) I don't think disk format and size are necessary parameters for
>> >> connectPhysicalDisk, as the format can be determined by the adaptor,
>> >> and the size is set during the createAsync call in the plugin. We
>> >> really just need the disk path and the pool.
>> >>
>> > [Mike T.] I agree, format is not needed. The only reason I have size
>> passed
>> > in is because I need to create a KVMPhysicalDisk at the end of the
>> connect
>> > method. Since this KVMPhysicalDisk is (in the code on GitHub) being
>> used to
>> > create our XML to attach the disk, I figured we'd need that size. The
>> > KVMPhysicalDisk I produce from my implementation of getPhysicalDisk is
>> not
>> > as good because I don't know the size of the disk at this point (I don't
>> > keep that information around). The reason I don't keep that info around
>> is
>> > because I don't have a good way to reproduce that info if the KVM host
>> is
>> > rebooted. We get info about storage pools in the form of a
>> > ModifyStoragePoolCommand, but nothing about the volumes inside of the
>> > storage pool.
>>
>> getPhysicalDisk is called in a bunch of places. I'd rely on the
>> connectPhysicalDisk to only make the block device appear on the host,
>> and then getPhysicalDisk to find that block device and fill out things
>> like disk size and path (the real path to the local block device) for
>> passing and creating the disk XML. Trust me, unless things have
>> changed significantly you need to be able to identify a given device
>> as a specific local disk by whatever you are setting the 'path'
>> attribute to be.  getPhysicalDisk will be called on your storage pool
>> with simply the path attribute, and via your adaptor with the pool and
>> path.
>>
>> So you may set path as some combination of iqn and target/pool info,
>> or if iqn is enough to identify a unique block device (in
>> /dev/disk/by-id maybe?) on a host then just use that. Path just needs
>> to be something, anything, to identify the disk on the host. In
>> getPhysicalDisk, identify the local block device matching the info,
>> create a new KVMPhysicalDisk with the local path, size, etc, and
>> return it.
>>
>> >
>> >>
>> >> 2) I thought this access group thing you mention are the grantAccess
>> >> and revokeAccess calls in the storage plugin 2.0 design doc. Was that
>> >> not implemented?
>> >>
>> > [Mike T.] Yeah, as I mentioned in an e-mail way back, those methods were
>> > never implemented in 4.2. I think you said you were going to get around
>> > them not being implemented by keeping certain logic that talks to the
>> SAN
>> > in the agent. I don't think we want any SolidFire-specific code in the
>> > agent, however, so I can't go that route. If those methods do not get
>> > implemented in 4.3, then I will need to use CHAP credentials for KVM
>> (just
>> > like I did with XenServer and VMware).
>>
>> I initially figured your StorageAdaptor implementation would be all
>> solidfire specific, just like the mgmt server side plugin is. If it
>> can be generic to all iscsi storage then that's great. I agree that
>> ideally the agent wouldn't be making API calls to your SAN. I don't
>> think it should be necessary given that you're not going to use the
>> ACL route. I'm not sure CHAP fills the same purpose of fencing.
>>
>> >
>> >>
>> >> I see you've  added getters/setters for the attach cmd to pass the
>> >> iscsi info you need. Would it perhaps be possible to send a details
>> >> Map<String, String> instead? Then any plugin implementer could attach
>> >> arbitrary data they need. So it might be
>> >> connectPhysicalDisk(StoragePoolType type, String poolUuid, String
>> >> volPath, Map<String, String> details)?  I'll have to look and see
>> >> where those cmd. attributes are set, ideally it would be all the way
>> >> back in the plugin to avoid custom code for every adaptor that wants
>> >> to set details.
>> >>
>> > [Mike T.] If I'm not using the volumes.path field for anything, I can
>> stick
>> > the IQN in volumes.path (as well as leaving it in volumes.iscsi_name,
>> which
>> > is used elsewhere). That way we only have to ask for getPath().
>>
>> Yeah, whatever it is that you'd need to find the right block device
>> should go in the path. If you look through LibvirtComputingResource
>> you'll see stuff like this sprinkled around:
>>
>>                 KVMPhysicalDisk volume = primaryPool.getPhysicalDisk(cmd
>>                         .getVolumePath());
>>
>>
>> or:
>>         String volid = cmd.getPath();
>>          KVMPhysicalDisk vol = pool.getPhysicalDisk(volid);
>>
>> or:
>>
>>                     KVMPhysicalDisk physicalDisk =
>> _storagePoolMgr.getPhysicalDisk( store.getPoolType(),
>>                             store.getUuid(),
>>                             data.getPath());
>>
>> Maybe some of it is short-circuited by the new KVMStorageProcessor,
>> but I'd still implement a working one, and then attachVolume can call
>> getPhysicalDisk after connectPhysicalDisk, even on your adaptor/pool.
>>
>> >
>> >>
>> >> On Thu, Sep 26, 2013 at 7:35 PM, Mike Tutkowski
>> >> <mike.tutkow...@solidfire.com> wrote:
>> >> > Also, if we went the non-CHAP route, before attaching a volume to a
>> VM,
>> >> > we'd have to tell the plug-in to set up a volume access group.
>> >> >
>> >> > When a volume is detached from a VM, we'd have to tell the plug-in to
>> >> > delete the volume access group.
>> >> >
>> >> >
>> >> > On Thu, Sep 26, 2013 at 7:32 PM, Mike Tutkowski <
>> >> > mike.tutkow...@solidfire.com> wrote:
>> >> >
>> >> >> I mention this is my comments on GitHub, as well, but CHAP info is
>> >> >> associated with an account - not a storage pool.
>> >> >>
>> >> >> Ideally we could do without CHAP info entirely if we had a reliable
>> way
>> >> to
>> >> >> tell the storage plug-in that a given host wants to access a given
>> >> volume.
>> >> >> In this case, my storage plug-in could create what we call a Volume
>> >> Access
>> >> >> Group on the SAN. It would essentially say, "The host with IQN <x>
>> can
>> >> >> access the volume with IQN <y> without using CHAP credentials." Of
>> >> course
>> >> >> we'd need a way to revoke this privilege in the event of a live
>> >> migration
>> >> >> of a VM. Right now, I do not believe such a facility is supported
>> with
>> >> the
>> >> >> storage plug-ins.
>> >> >>
>> >> >>
>> >> >> On Thu, Sep 26, 2013 at 4:56 PM, Marcus Sorensen <
>> shadow...@gmail.com
>> >> >wrote:
>> >> >>
>> >> >>> Looking at your code, is the chap info stored with the pool, so we
>> >> >>> could pass the pool to the adaptor? That would be more agnostic,
>> >> >>> anyone implementing a plugin could pull the specifics they need for
>> >> >>> their stuff out of the pool on the adaptor side, rather than
>> creating
>> >> >>> custom signatures.
>> >> >>>
>> >> >>> Also, I think we may want to consider implementing
>> connect/disconnect
>> >> >>> as just dummy methods in LibvirtStorageAdaptor, so we don't have
>> to be
>> >> >>> picky about which adaptors/types in every single place we may want
>> to
>> >> >>> connect/disconnect (in 4.1 there were several, I'm not sure if
>> >> >>> everything goes through this in 4.2). We can just call
>> >> >>> adaptor.connectPhysicalDisk and the adaptor can decide if it needs
>> to
>> >> >>> do anything.
>> >> >>>
>> >> >>> Comments are attached to your commit, I just wanted to echo them
>> here
>> >> >>> on-list.
>> >> >>>
>> >> >>> On Thu, Sep 26, 2013 at 4:32 PM, Mike Tutkowski
>> >> >>> <mike.tutkow...@solidfire.com> wrote:
>> >> >>> > Oh, SnapshotTestWithFakeData is just modified because the code
>> wasn't
>> >> >>> > building until I corrected this. It has nothing really to do
>> with my
>> >> >>> real
>> >> >>> > changes.
>> >> >>> >
>> >> >>> >
>> >> >>> > On Thu, Sep 26, 2013 at 4:31 PM, Mike Tutkowski <
>> >> >>> > mike.tutkow...@solidfire.com> wrote:
>> >> >>> >
>> >> >>> >> Hey Marcus,
>> >> >>> >>
>> >> >>> >> I implemented your recommendations regarding adding connect and
>> >> >>> disconnect
>> >> >>> >> methods. It is not yet checked in (as you know, having trouble
>> with
>> >> my
>> >> >>> KVM
>> >> >>> >> environment), but it is on GitHub here:
>> >> >>> >>
>> >> >>> >>
>> >> >>> >>
>> >> >>>
>> >>
>> https://github.com/mike-tutkowski/incubator-cloudstack/commit/f2897c65689012e6157c0a0c2ed7e5355900c59a
>> >> >>> >>
>> >> >>> >> Please let me know if you have any more comments.
>> >> >>> >>
>> >> >>> >> Thanks!
>> >> >>> >>
>> >> >>> >>
>> >> >>> >> On Thu, Sep 26, 2013 at 4:05 PM, Marcus Sorensen <
>> >> shadow...@gmail.com
>> >> >>> >wrote:
>> >> >>> >>
>> >> >>> >>> Mike, everyone,
>> >> >>> >>>    As I've mentioned on the board, I'm working on getting our
>> own
>> >> >>> >>> internal KVM storage plugin working on 4.2. In the interest of
>> >> making
>> >> >>> >>> it forward compatible, I just wanted to confirm what you were
>> doing
>> >> >>> >>> with the solidfire plugin as far as attaching your iscsi LUNs.
>> We
>> >> had
>> >> >>> >>> discussed a new connectPhysicalDisk method for the
>> StorageAdaptor
>> >> >>> >>> class, something perhaps like:
>> >> >>> >>>
>> >> >>> >>> public boolean connectPhysicalDisk(String volumeUuid,
>> >> KVMStoragePool
>> >> >>> >>> pool);
>> >> >>> >>>
>> >> >>> >>> then added to KVMStoragePoolManager:
>> >> >>> >>>
>> >> >>> >>> public boolean connectPhysicalDisk(StoragePoolType type, String
>> >> >>> >>> poolUuid, String volPath) {
>> >> >>> >>>         StorageAdaptor adaptor = getStorageAdaptor(type);
>> >> >>> >>>         KVMStoragePool pool = adaptor.getStoragePool(poolUuid);
>> >> >>> >>>         return adaptor.connectPhysicalDisk(volPath, pool);
>> >> >>> >>>     }
>> >> >>> >>>
>> >> >>> >>> Something similar to this for disconnect as well. Then in the
>> >> >>> >>> KVMStorageProcessor these can be called as needed for
>> >> attach/detach.
>> >> >>> >>> We can probably stub out one in LibvirtStorageAdaptor so
>> there's no
>> >> >>> >>> need to switch or if/else for pool types, for example in
>> >> >>> >>> KVMStorageProcessor.attachVolume.
>> >> >>> >>>
>> >> >>> >>> I have debated on whether or not it should just be rolled into
>> >> >>> >>> getPhysicalDisk, having it connect the disk if it's not already
>> >> >>> >>> connected. getPhysicalDisk is called a lot, and I'm not sure it
>> >> always
>> >> >>> >>> needs to connect the disk when it does. In past iterations
>> >> >>> >>> getPhysicalDisk has simply spoken to our SAN api and returned
>> the
>> >> disk
>> >> >>> >>> details, nothing more. So it seemed more flexible and granular
>> to
>> >> do
>> >> >>> >>> the connectPhysicalDisk (we have one now in our 4.1 version).
>> >> >>> >>>
>> >> >>> >>
>> >> >>> >>
>> >> >>> >>
>> >> >>> >> --
>> >> >>> >> *Mike Tutkowski*
>> >> >>> >> *Senior CloudStack Developer, SolidFire Inc.*
>> >> >>> >> e: mike.tutkow...@solidfire.com
>> >> >>> >> o: 303.746.7302
>> >> >>> >> Advancing the way the world uses the cloud<
>> >> >>> http://solidfire.com/solution/overview/?video=play>
>> >> >>> >> *™*
>> >> >>> >>
>> >> >>> >
>> >> >>> >
>> >> >>> >
>> >> >>> > --
>> >> >>> > *Mike Tutkowski*
>> >> >>> > *Senior CloudStack Developer, SolidFire Inc.*
>> >> >>> > e: mike.tutkow...@solidfire.com
>> >> >>> > o: 303.746.7302
>> >> >>> > Advancing the way the world uses the
>> >> >>> > cloud<http://solidfire.com/solution/overview/?video=play>
>> >> >>> > *™*
>> >> >>>
>> >> >>
>> >> >>
>> >> >>
>> >> >> --
>> >> >> *Mike Tutkowski*
>> >> >> *Senior CloudStack Developer, SolidFire Inc.*
>> >> >> e: mike.tutkow...@solidfire.com
>> >> >> o: 303.746.7302
>> >> >> Advancing the way the world uses the cloud<
>> >> http://solidfire.com/solution/overview/?video=play>
>> >> >> *™*
>> >> >>
>> >> >
>> >> >
>> >> >
>> >> > --
>> >> > *Mike Tutkowski*
>> >> > *Senior CloudStack Developer, SolidFire Inc.*
>> >> > e: mike.tutkow...@solidfire.com
>> >> > o: 303.746.7302
>> >> > Advancing the way the world uses the
>> >> > cloud<http://solidfire.com/solution/overview/?video=play>
>> >> > *™*
>> >>
>> >
>> >
>> >
>> > --
>> > *Mike Tutkowski*
>> > *Senior CloudStack Developer, SolidFire Inc.*
>> > e: mike.tutkow...@solidfire.com
>> > o: 303.746.7302
>> > Advancing the way the world uses the
>> > cloud<http://solidfire.com/solution/overview/?video=play>
>> > *™*
>>
>
>
>
> --
> *Mike Tutkowski*
> *Senior CloudStack Developer, SolidFire Inc.*
> e: mike.tutkow...@solidfire.com
> o: 303.746.7302
> Advancing the way the world uses the 
> cloud<http://solidfire.com/solution/overview/?video=play>
> *™*
>



-- 
*Mike Tutkowski*
*Senior CloudStack Developer, SolidFire Inc.*
e: mike.tutkow...@solidfire.com
o: 303.746.7302
Advancing the way the world uses the
cloud<http://solidfire.com/solution/overview/?video=play>
*™*

Reply via email to