Thats great thank you so much. I will try and get this patch in my test env
asap but will likely wait for official release cut for prod. I really
appreciate you adding this in to the product.

Respectfully,

*Wes Dillingham*
w...@wesdillingham.com
LinkedIn <http://www.linkedin.com/in/wesleydillingham>


On Thu, Dec 5, 2019 at 4:14 PM Mike Christie <mchri...@redhat.com> wrote:

> On 12/04/2019 02:34 PM, Wesley Dillingham wrote:
> > I have never had a permanent loss of a gateway but I'm a believer in
> > Murphy's law and want to have a plan. Glad to hear that there is a
> > solution in-the-works, curious when might that be available in a
> > release? If sooner than later I'll plan to upgrade then immediately,
>
> It should be in the next release which I think we would just make when
> the patch gets merged since we have a good number of fixes sitting in
> the repo.
>
> Patch/PR is here
>
> https://github.com/ceph/ceph-iscsi/pull/156
>
> if you have a non production setup and are used to applying patches and
> testing upstream.
>
> > otherwise, if far down the queue I would like to know if I should ready
> > a standby server.
> >
> >  Thanks so much for all your great work on this product.
> >
> >
> > Respectfully,
> >
> > *Wes Dillingham*
> > w...@wesdillingham.com <mailto:w...@wesdillingham.com>
> > LinkedIn <http://www.linkedin.com/in/wesleydillingham>
> >
> >
> > On Wed, Dec 4, 2019 at 11:18 AM Mike Christie <mchri...@redhat.com
> > <mailto:mchri...@redhat.com>> wrote:
> >
> >     On 12/04/2019 08:26 AM, Gesiel Galvão Bernardes wrote:
> >     > Hi,
> >     >
> >     > Em qua., 4 de dez. de 2019 às 00:31, Mike Christie
> >     <mchri...@redhat.com <mailto:mchri...@redhat.com>
> >     > <mailto:mchri...@redhat.com <mailto:mchri...@redhat.com>>>
> escreveu:
> >     >
> >     >     On 12/03/2019 04:19 PM, Wesley Dillingham wrote:
> >     >     > Thanks. If I am reading this correctly the ability to remove
> >     an iSCSI
> >     >     > gateway would allow the remaining iSCSI gateways to take
> >     over for the
> >     >     > removed gateway's LUN's as of > 3.0. Thats good, we run 3.2.
> >     However,
> >     >     > because the actual update of the central config object
> happens
> >     >     from the
> >     >     > to-be-deleted iSCSI gateway, despite where the gwcli command
> is
> >     >     issued,
> >     >     > it will fail to actually remove said gateway from the object
> >     if that
> >     >     > gateway is not functioning.
> >     >
> >     >     Yes.
> >     >
> >     >     >
> >     >     > I guess this leaves the question still of how to proceed
> >     when one
> >     >     of the
> >     >     > iSCSI gateways fails permanently?  Is that possible, or is it
> >     >     > potentially possible other than manually intervening on the
> >     config
> >     >
> >     >     You could edit the gateway.cfg manually, but I would not do
> >     it, because
> >     >     it's error prone.
> >     >
> >     >     It's probably safest to run in degraded mode and wait for an
> >     updated
> >     >     ceph-iscsi package with a fix. If you are running into the
> >     problem right
> >     >     now, I can bump the priority.
> >     >
> >     > I permanently lost a gateway. I can not leave running "degraded"
> >     because
> >     > I need to add another redundancy gateway, and it does not allow
> >     with the
> >     > gateway "offline".
> >     >
> >     > In this case, what can I do? If I create a new gateway with the
> same
> >     > name and IP as the lost one, and then try to use "delete" in
> >     gwcli, will
> >     > it work?
> >
> >     Yes.
> >
> >     If you can have a temp stop in services you can also do the
> following as
> >     a workaround:
> >
> >     0. Stop applications accessing iscsi luns, and have the initiator log
> >     out of the iscsi target.
> >
> >     1. Stop ceph iscsi service. On all iscsi gw nodes do:
> >
> >     systemctl stop rbd-target-api
> >
> >     2. Delete gateway.cfg. This will delete the configuration info like
> the
> >     target and its ACL and LUN mappings. It does not delete the actual
> >     images or pools that you have data on.
> >
> >     rados -p rbd rm gateway.cfg
> >
> >     3. Start ceph iscsi services again. On all iscsi gw nodes do:
> >
> >     systemctl start rbd-target-api
> >
> >     4. Resetup target with gwcli. For the image/disk setup stage,
> instead of
> >     doing the "create" command do the "attach"command:
> >
> >     attach pool=your_pool image=image_name
> >
> >     Then just re-add your target, ACLs and LUN mappings.
> >
> >     5. On the initiator side relogin to the iscsi target.
> >
> >
> >     >
> >     >
> >     >
> >     >     > object? If its not possible would the best course of action
> >     be to have
> >     >     > standby hardware and quickly recreate the node or perhaps
> >     run the
> >     >     > gateways more ephemerally, from a VM or container?
> >     >     >
> >     >     > Thanks again.
> >     >     >
> >     >     > Respectfully,
> >     >     >
> >     >     > *Wes Dillingham*
> >     >     > w...@wesdillingham.com <mailto:w...@wesdillingham.com>
> >     <mailto:w...@wesdillingham.com <mailto:w...@wesdillingham.com>>
> >     >     <mailto:w...@wesdillingham.com <mailto:w...@wesdillingham.com>
> >     <mailto:w...@wesdillingham.com <mailto:w...@wesdillingham.com>>>
> >     >     > LinkedIn <http://www.linkedin.com/in/wesleydillingham>
> >     >     >
> >     >     >
> >     >     > On Tue, Dec 3, 2019 at 2:45 PM Mike Christie
> >     <mchri...@redhat.com <mailto:mchri...@redhat.com>
> >     >     <mailto:mchri...@redhat.com <mailto:mchri...@redhat.com>>
> >     >     > <mailto:mchri...@redhat.com <mailto:mchri...@redhat.com>
> >     <mailto:mchri...@redhat.com <mailto:mchri...@redhat.com>>>> wrote:
> >     >     >
> >     >     >     I do not think it's going to do what you want when the
> node
> >     >     you want to
> >     >     >     delete is down.
> >     >     >
> >     >     >     It looks like we only temporarily stop the gw from being
> >     >     exported. It
> >     >     >     does not update the gateway.cfg, because we do the config
> >     >     removal call
> >     >     >     on the node we want to delete.
> >     >     >
> >     >     >     So gwcli would report success and the ls command will
> >     show it
> >     >     as no
> >     >     >     longer running/exported, but if you restart the
> >     rbd-target-api
> >     >     service
> >     >     >     then it will show up again.
> >     >     >
> >     >     >     There is an internal command to do what you want. I will
> >     post
> >     >     a PR for
> >     >     >     gwlci and so it can be used by dashboard.
> >     >     >
> >     >     >
> >     >     >     On 12/03/2019 01:19 PM, Jason Dillaman wrote:
> >     >     >     > If I recall correctly, the recent ceph-iscsi release
> >     >     supports the
> >     >     >     > removal of a gateway via the "gwcli". I think the Ceph
> >     >     dashboard can
> >     >     >     > do that as well.
> >     >     >     >
> >     >     >     > On Tue, Dec 3, 2019 at 1:59 PM Wesley Dillingham
> >     >     >     <w...@wesdillingham.com <mailto:w...@wesdillingham.com>
> >     <mailto:w...@wesdillingham.com <mailto:w...@wesdillingham.com>>
> >     >     <mailto:w...@wesdillingham.com <mailto:w...@wesdillingham.com>
> >     <mailto:w...@wesdillingham.com <mailto:w...@wesdillingham.com>>>>
> wrote:
> >     >     >     >>
> >     >     >     >> We utilize 4 iSCSI gateways in a cluster and have
> >     noticed the
> >     >     >     following during patching cycles when we sequentially
> reboot
> >     >     single
> >     >     >     iSCSI-gateways:
> >     >     >     >>
> >     >     >     >> "gwcli" often hangs on the still-up iSCSI GWs but
> >     sometimes
> >     >     still
> >     >     >     functions and gives the message:
> >     >     >     >>
> >     >     >     >> "1 gateway is inaccessible - updates will be disabled"
> >     >     >     >>
> >     >     >     >> This got me thinking about what the course of action
> >     would be
> >     >     >     should an iSCSI gateway fail permanently or
> >     semi-permanently,
> >     >     say a
> >     >     >     hardware issue. What would be the best course of action
> to
> >     >     instruct
> >     >     >     the remaining iSCSI gateways that one of them is no
> longer
> >     >     available
> >     >     >     and that they should allow updates again and take
> >     ownership of the
> >     >     >     now-defunct-node's LUNS?
> >     >     >     >>
> >     >     >     >> I'm guessing pulling down the RADOS config object and
> >     rewriting
> >     >     >     it and re-put'ing it followed by a rbd-target-api
> >     restart might do
> >     >     >     the trick but am hoping there is a more "in-band" and
> less
> >     >     >     potentially devastating way to do this.
> >     >     >     >>
> >     >     >     >> Thanks for any insights.
> >     >     >     >>
> >     >     >     >> Respectfully,
> >     >     >     >>
> >     >     >     >> Wes Dillingham
> >     >     >     >> w...@wesdillingham.com <mailto:w...@wesdillingham.com>
> >     <mailto:w...@wesdillingham.com <mailto:w...@wesdillingham.com>>
> >     >     <mailto:w...@wesdillingham.com <mailto:w...@wesdillingham.com>
> >     <mailto:w...@wesdillingham.com <mailto:w...@wesdillingham.com>>>
> >     >     >     >> LinkedIn
> >     >     >     >> _______________________________________________
> >     >     >     >> ceph-users mailing list -- ceph-users@ceph.io
> >     <mailto:ceph-users@ceph.io>
> >     >     <mailto:ceph-users@ceph.io <mailto:ceph-users@ceph.io>>
> >     >     >     <mailto:ceph-users@ceph.io <mailto:ceph-users@ceph.io>
> >     <mailto:ceph-users@ceph.io <mailto:ceph-users@ceph.io>>>
> >     >     >     >> To unsubscribe send an email to
> >     ceph-users-le...@ceph.io <mailto:ceph-users-le...@ceph.io>
> >     >     <mailto:ceph-users-le...@ceph.io
> >     <mailto:ceph-users-le...@ceph.io>>
> >     >     >     <mailto:ceph-users-le...@ceph.io
> >     <mailto:ceph-users-le...@ceph.io>
> >     >     <mailto:ceph-users-le...@ceph.io
> >     <mailto:ceph-users-le...@ceph.io>>>
> >     >     >     >
> >     >     >     >
> >     >     >     >
> >     >     >
> >     >     _______________________________________________
> >     >     ceph-users mailing list -- ceph-users@ceph.io
> >     <mailto:ceph-users@ceph.io>
> >     >     <mailto:ceph-users@ceph.io <mailto:ceph-users@ceph.io>>
> >     >     To unsubscribe send an email to ceph-users-le...@ceph.io
> >     <mailto:ceph-users-le...@ceph.io>
> >     >     <mailto:ceph-users-le...@ceph.io
> >     <mailto:ceph-users-le...@ceph.io>>
> >     >
> >     _______________________________________________
> >     ceph-users mailing list -- ceph-users@ceph.io
> >     <mailto:ceph-users@ceph.io>
> >     To unsubscribe send an email to ceph-users-le...@ceph.io
> >     <mailto:ceph-users-le...@ceph.io>
> >
>
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to