Hi,

Em qua., 4 de dez. de 2019 às 00:31, Mike Christie <mchri...@redhat.com>
escreveu:

> On 12/03/2019 04:19 PM, Wesley Dillingham wrote:
> > Thanks. If I am reading this correctly the ability to remove an iSCSI
> > gateway would allow the remaining iSCSI gateways to take over for the
> > removed gateway's LUN's as of > 3.0. Thats good, we run 3.2. However,
> > because the actual update of the central config object happens from the
> > to-be-deleted iSCSI gateway, despite where the gwcli command is issued,
> > it will fail to actually remove said gateway from the object if that
> > gateway is not functioning.
>
> Yes.
>
> >
> > I guess this leaves the question still of how to proceed when one of the
> > iSCSI gateways fails permanently?  Is that possible, or is it
> > potentially possible other than manually intervening on the config
>
> You could edit the gateway.cfg manually, but I would not do it, because
> it's error prone.
>
> It's probably safest to run in degraded mode and wait for an updated
> ceph-iscsi package with a fix. If you are running into the problem right
> now, I can bump the priority.
>
> I permanently lost a gateway. I can not leave running "degraded" because I
need to add another redundancy gateway, and it does not allow with the
gateway "offline".

In this case, what can I do? If I create a new gateway with the same name
and IP as the lost one, and then try to use "delete" in gwcli, will it work?



> > object? If its not possible would the best course of action be to have
> > standby hardware and quickly recreate the node or perhaps run the
> > gateways more ephemerally, from a VM or container?
> >
> > Thanks again.
> >
> > Respectfully,
> >
> > *Wes Dillingham*
> > w...@wesdillingham.com <mailto:w...@wesdillingham.com>
> > LinkedIn <http://www.linkedin.com/in/wesleydillingham>
> >
> >
> > On Tue, Dec 3, 2019 at 2:45 PM Mike Christie <mchri...@redhat.com
> > <mailto:mchri...@redhat.com>> wrote:
> >
> >     I do not think it's going to do what you want when the node you want
> to
> >     delete is down.
> >
> >     It looks like we only temporarily stop the gw from being exported. It
> >     does not update the gateway.cfg, because we do the config removal
> call
> >     on the node we want to delete.
> >
> >     So gwcli would report success and the ls command will show it as no
> >     longer running/exported, but if you restart the rbd-target-api
> service
> >     then it will show up again.
> >
> >     There is an internal command to do what you want. I will post a PR
> for
> >     gwlci and so it can be used by dashboard.
> >
> >
> >     On 12/03/2019 01:19 PM, Jason Dillaman wrote:
> >     > If I recall correctly, the recent ceph-iscsi release supports the
> >     > removal of a gateway via the "gwcli". I think the Ceph dashboard
> can
> >     > do that as well.
> >     >
> >     > On Tue, Dec 3, 2019 at 1:59 PM Wesley Dillingham
> >     <w...@wesdillingham.com <mailto:w...@wesdillingham.com>> wrote:
> >     >>
> >     >> We utilize 4 iSCSI gateways in a cluster and have noticed the
> >     following during patching cycles when we sequentially reboot single
> >     iSCSI-gateways:
> >     >>
> >     >> "gwcli" often hangs on the still-up iSCSI GWs but sometimes still
> >     functions and gives the message:
> >     >>
> >     >> "1 gateway is inaccessible - updates will be disabled"
> >     >>
> >     >> This got me thinking about what the course of action would be
> >     should an iSCSI gateway fail permanently or semi-permanently, say a
> >     hardware issue. What would be the best course of action to instruct
> >     the remaining iSCSI gateways that one of them is no longer available
> >     and that they should allow updates again and take ownership of the
> >     now-defunct-node's LUNS?
> >     >>
> >     >> I'm guessing pulling down the RADOS config object and rewriting
> >     it and re-put'ing it followed by a rbd-target-api restart might do
> >     the trick but am hoping there is a more "in-band" and less
> >     potentially devastating way to do this.
> >     >>
> >     >> Thanks for any insights.
> >     >>
> >     >> Respectfully,
> >     >>
> >     >> Wes Dillingham
> >     >> w...@wesdillingham.com <mailto:w...@wesdillingham.com>
> >     >> LinkedIn
> >     >> _______________________________________________
> >     >> ceph-users mailing list -- ceph-users@ceph.io
> >     <mailto:ceph-users@ceph.io>
> >     >> To unsubscribe send an email to ceph-users-le...@ceph.io
> >     <mailto:ceph-users-le...@ceph.io>
> >     >
> >     >
> >     >
> >
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to