Hi,

Em qua., 4 de dez. de 2019 às 00:31, Mike Christie <mchristi@redhat.com> escreveu:
On 12/03/2019 04:19 PM, Wesley Dillingham wrote:
> Thanks. If I am reading this correctly the ability to remove an iSCSI
> gateway would allow the remaining iSCSI gateways to take over for the
> removed gateway's LUN's as of > 3.0. Thats good, we run 3.2. However,
> because the actual update of the central config object happens from the
> to-be-deleted iSCSI gateway, despite where the gwcli command is issued,
> it will fail to actually remove said gateway from the object if that
> gateway is not functioning.

Yes.

>
> I guess this leaves the question still of how to proceed when one of the
> iSCSI gateways fails permanently?  Is that possible, or is it
> potentially possible other than manually intervening on the config

You could edit the gateway.cfg manually, but I would not do it, because
it's error prone.

It's probably safest to run in degraded mode and wait for an updated
ceph-iscsi package with a fix. If you are running into the problem right
now, I can bump the priority.

I permanently lost a gateway. I can not leave running "degraded" because I need to add another redundancy gateway, and it does not allow with the gateway "offline".

In this case, what can I do? If I create a new gateway with the same name and IP as the lost one, and then try to use "delete" in gwcli, will it work?
 
 
> object? If its not possible would the best course of action be to have
> standby hardware and quickly recreate the node or perhaps run the
> gateways more ephemerally, from a VM or container?
>
> Thanks again.
>
> Respectfully,
>
> *Wes Dillingham*
> wes@wesdillingham.com <mailto:wes@wesdillingham.com>
> LinkedIn <http://www.linkedin.com/in/wesleydillingham>
>
>
> On Tue, Dec 3, 2019 at 2:45 PM Mike Christie <mchristi@redhat.com
> <mailto:mchristi@redhat.com>> wrote:
>
>     I do not think it's going to do what you want when the node you want to
>     delete is down.
>
>     It looks like we only temporarily stop the gw from being exported. It
>     does not update the gateway.cfg, because we do the config removal call
>     on the node we want to delete.
>
>     So gwcli would report success and the ls command will show it as no
>     longer running/exported, but if you restart the rbd-target-api service
>     then it will show up again.
>
>     There is an internal command to do what you want. I will post a PR for
>     gwlci and so it can be used by dashboard.
>
>
>     On 12/03/2019 01:19 PM, Jason Dillaman wrote:
>     > If I recall correctly, the recent ceph-iscsi release supports the
>     > removal of a gateway via the "gwcli". I think the Ceph dashboard can
>     > do that as well.
>     >
>     > On Tue, Dec 3, 2019 at 1:59 PM Wesley Dillingham
>     <wes@wesdillingham.com <mailto:wes@wesdillingham.com>> wrote:
>     >>
>     >> We utilize 4 iSCSI gateways in a cluster and have noticed the
>     following during patching cycles when we sequentially reboot single
>     iSCSI-gateways:
>     >>
>     >> "gwcli" often hangs on the still-up iSCSI GWs but sometimes still
>     functions and gives the message:
>     >>
>     >> "1 gateway is inaccessible - updates will be disabled"
>     >>
>     >> This got me thinking about what the course of action would be
>     should an iSCSI gateway fail permanently or semi-permanently, say a
>     hardware issue. What would be the best course of action to instruct
>     the remaining iSCSI gateways that one of them is no longer available
>     and that they should allow updates again and take ownership of the
>     now-defunct-node's LUNS?
>     >>
>     >> I'm guessing pulling down the RADOS config object and rewriting
>     it and re-put'ing it followed by a rbd-target-api restart might do
>     the trick but am hoping there is a more "in-band" and less
>     potentially devastating way to do this.
>     >>
>     >> Thanks for any insights.
>     >>
>     >> Respectfully,
>     >>
>     >> Wes Dillingham
>     >> wes@wesdillingham.com <mailto:wes@wesdillingham.com>
>     >> LinkedIn
>     >> _______________________________________________
>     >> ceph-users mailing list -- ceph-users@ceph.io
>     <mailto:ceph-users@ceph.io>
>     >> To unsubscribe send an email to ceph-users-leave@ceph.io
>     <mailto:ceph-users-leave@ceph.io>
>     >
>     >
>     >
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-leave@ceph.io