[ceph-users] osd removal leaves 'stray daemon'

2022-11-25 Thread Holger Naundorf

Hello,
I have a question about osd removal/replacement:

I just removed an osd where the disk was still running but had read 
errors, leading to failed deep scrubs - as the intent is to replace this 
as soon as we manage to get a spare I removed it with the '--replace' flag:


# ceph orch osd rm 224 --replace

After all placement groups are evacuated I now have 1 osd down/out
and showing as 'destroyed':

# ceph osd tree
ID   CLASS  WEIGHT  TYPE NAMESTATUS REWEIGHT  PRI-AFF
(...)
214hdd14.55269  osd.214 up   1.0  1.0
224hdd14.55269  osd.224  destroyed 0  1.0
234hdd14.55269  osd.234 up   1.0  1.0
(...)

All as expected - but now the health check complains that the 
(destroyed) osd is not managed:


# ceph health detail
HEALTH_WARN 1 stray daemon(s) not managed by cephadm
[WRN] CEPHADM_STRAY_DAEMON: 1 stray daemon(s) not managed by cephadm
stray daemon osd.224 on host ceph19 not managed by cephadm

Is this expected behaviour and I have to live with the yellow check 
until we get a replacement disk and recreate the osd or did something 
not finish correctly?


Regards,
Holger

--
Dr. Holger Naundorf
Christian-Albrechts-Universität zu Kiel
Rechenzentrum / HPC / Server und Storage
Tel: +49 431 880-1990
Fax:  +49 431 880-1523
naund...@rz.uni-kiel.de
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: osd removal leaves 'stray daemon'

2022-12-05 Thread Holger Naundorf

Hello,
a mgr failover did not change the situation - the osd still shows up in 
the 'ceph node ls' - I assume that this is more or less 'working as 
intended' as I did ask for the OSD to be kept in the CRUSH map to be 
replacd later - but as we are still not so experienced with Ceph here I 
wanted to get some input from other sites.


Regards,
Holger

On 30.11.22 16:28, Adam King wrote:
I typically don't see this when I do OSD replacement. If you do a mgr 
failover ("ceph mgr fail") and wait a few minutes does this still show 
up? The stray daemon/host warning is roughly equivalent to comparing the 
daemons in `ceph node ls` and `ceph orch ps` and seeing if there's 
anything in the former but not the latter. Sometimes I have seen the mgr 
will have some out of data info in the node ls and a failover will 
refresh it.


On Fri, Nov 25, 2022 at 6:07 AM Holger Naundorf <mailto:naund...@rz.uni-kiel.de>> wrote:


Hello,
I have a question about osd removal/replacement:

I just removed an osd where the disk was still running but had read
errors, leading to failed deep scrubs - as the intent is to replace
this
as soon as we manage to get a spare I removed it with the
'--replace' flag:

# ceph orch osd rm 224 --replace

After all placement groups are evacuated I now have 1 osd down/out
and showing as 'destroyed':

# ceph osd tree
ID   CLASS  WEIGHT      TYPE NAME        STATUS     REWEIGHT  PRI-AFF
(...)
214    hdd    14.55269          osd.214         up   1.0  1.0
224    hdd    14.55269          osd.224  destroyed         0  1.0
234    hdd    14.55269          osd.234         up   1.0  1.0
(...)

All as expected - but now the health check complains that the
(destroyed) osd is not managed:

# ceph health detail
HEALTH_WARN 1 stray daemon(s) not managed by cephadm
[WRN] CEPHADM_STRAY_DAEMON: 1 stray daemon(s) not managed by cephadm
      stray daemon osd.224 on host ceph19 not managed by cephadm

Is this expected behaviour and I have to live with the yellow check
until we get a replacement disk and recreate the osd or did something
not finish correctly?

Regards,
Holger

-- 
Dr. Holger Naundorf

Christian-Albrechts-Universität zu Kiel
Rechenzentrum / HPC / Server und Storage
Tel: +49 431 880-1990
Fax:  +49 431 880-1523
naund...@rz.uni-kiel.de <mailto:naund...@rz.uni-kiel.de>
___
ceph-users mailing list -- ceph-users@ceph.io
<mailto:ceph-users@ceph.io>
To unsubscribe send an email to ceph-users-le...@ceph.io
<mailto:ceph-users-le...@ceph.io>



--
Dr. Holger Naundorf
Christian-Albrechts-Universität zu Kiel
Rechenzentrum / HPC / Server und Storage
Tel: +49 431 880-1990
Fax:  +49 431 880-1523
naund...@rz.uni-kiel.de
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] cephfs snap-mirror stalled

2022-12-06 Thread Holger Naundorf

Hello,
we have set up a snap-mirror for a directory on one of our clusters - 
running ceph version


ceph version 16.2.7 (dd0603118f56ab514f133c8d2e3adfc983942503) pacific 
(stable)


to get mirrorred our other cluster - running ceph version

ceph version 16.2.9 (4c3647a322c0ff5a1dd2344e039859dcbd28c830) pacific 
(stable)


The initial setup went ok, when the first snapshot was created data 
started to flow at a decent (for our HW) rate of 100-200MB/s. As the 
directory contains  ~200TB this was expected to take some time - but now 
the process has stalled completely after ~100TB were mirrored and ~7d 
running.


Up to now I do not have any hints why it has stopped - I do not see any 
error messages from the cephfs-mirror daemon. Can the small version 
mismatch be a problem?


Any hints where to look to find out what has got stuck are welcome.

Regards,
Holger

--
Dr. Holger Naundorf
Christian-Albrechts-Universität zu Kiel
Rechenzentrum / HPC / Server und Storage
Tel: +49 431 880-1990
Fax:  +49 431 880-1523
naund...@rz.uni-kiel.de
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: cephfs snap-mirror stalled

2022-12-06 Thread Holger Naundorf



On 06.12.22 09:54, Venky Shankar wrote:

Hi Holger,

On Tue, Dec 6, 2022 at 1:42 PM Holger Naundorf  wrote:


Hello,
we have set up a snap-mirror for a directory on one of our clusters -
running ceph version

ceph version 16.2.7 (dd0603118f56ab514f133c8d2e3adfc983942503) pacific
(stable)

to get mirrorred our other cluster - running ceph version

ceph version 16.2.9 (4c3647a322c0ff5a1dd2344e039859dcbd28c830) pacific
(stable)

The initial setup went ok, when the first snapshot was created data
started to flow at a decent (for our HW) rate of 100-200MB/s. As the
directory contains  ~200TB this was expected to take some time - but now
the process has stalled completely after ~100TB were mirrored and ~7d
running.

Up to now I do not have any hints why it has stopped - I do not see any
error messages from the cephfs-mirror daemon. Can the small version
mismatch be a problem?

Any hints where to look to find out what has got stuck are welcome.


I'd look at the mirror daemon logs for any errors to start with. You
might want to crank up the log level for debugging (debug
cephfs_mirror=20).



Even on max debug I do not see anything which looks like an error - but 
as this is the first time I try to dig into any cephfs-mirror logs I 
might not notice (as long as it is not red and flashing).


The Log basically this type of sequence, repeating forever:

(...)
cephfs::mirror::MirrorWatcher handle_notify
cephfs::mirror::Mirror update_fs_mirrors
cephfs::mirror::Mirror schedule_mirror_update_task: scheduling fs mirror 
update (0x556fe3a7f130) after 2 seconds
cephfs::mirror::Watcher handle_notify: notify_id=751516198184655, 
handle=93939050205568, notifier_id=25504530

cephfs::mirror::MirrorWatcher handle_notify
cephfs::mirror::PeerReplayer(19361031-928d-4366-99bd-50df70d3adf1) run: 
trying to pick from 1 directories
cephfs::mirror::PeerReplayer(19361031-928d-4366-99bd-50df70d3adf1) 
pick_directory
cephfs::mirror::Watcher handle_notify: notify_id=751516198184656, 
handle=93939050205568, notifier_id=25504530

cephfs::mirror::MirrorWatcher handle_notify
cephfs::mirror::Mirror update_fs_mirrors
cephfs::mirror::Mirror schedule_mirror_update_task: scheduling fs mirror 
update (0x556fe3a7fc70) after 2 seconds
cephfs::mirror::Watcher handle_notify: notify_id=751516198184657, 
handle=93939050205568, notifier_id=25504530

cephfs::mirror::MirrorWatcher handle_notify
(...)





Regards,
Holger

--
Dr. Holger Naundorf
Christian-Albrechts-Universität zu Kiel
Rechenzentrum / HPC / Server und Storage
Tel: +49 431 880-1990
Fax:  +49 431 880-1523
naund...@rz.uni-kiel.de
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io






--
Dr. Holger Naundorf
Christian-Albrechts-Universität zu Kiel
Rechenzentrum / HPC / Server und Storage
Tel: +49 431 880-1990
Fax:  +49 431 880-1523
naund...@rz.uni-kiel.de
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: cephfs snap-mirror stalled

2022-12-07 Thread Holger Naundorf

On 06.12.22 14:17, Venky Shankar wrote:

On Tue, Dec 6, 2022 at 6:34 PM Holger Naundorf  wrote:




On 06.12.22 09:54, Venky Shankar wrote:

Hi Holger,

On Tue, Dec 6, 2022 at 1:42 PM Holger Naundorf  wrote:


Hello,
we have set up a snap-mirror for a directory on one of our clusters -
running ceph version

ceph version 16.2.7 (dd0603118f56ab514f133c8d2e3adfc983942503) pacific
(stable)

to get mirrorred our other cluster - running ceph version

ceph version 16.2.9 (4c3647a322c0ff5a1dd2344e039859dcbd28c830) pacific
(stable)

The initial setup went ok, when the first snapshot was created data
started to flow at a decent (for our HW) rate of 100-200MB/s. As the
directory contains  ~200TB this was expected to take some time - but now
the process has stalled completely after ~100TB were mirrored and ~7d
running.

Up to now I do not have any hints why it has stopped - I do not see any
error messages from the cephfs-mirror daemon. Can the small version
mismatch be a problem?

Any hints where to look to find out what has got stuck are welcome.


I'd look at the mirror daemon logs for any errors to start with. You
might want to crank up the log level for debugging (debug
cephfs_mirror=20).



Even on max debug I do not see anything which looks like an error - but
as this is the first time I try to dig into any cephfs-mirror logs I
might not notice (as long as it is not red and flashing).

The Log basically this type of sequence, repeating forever:

(...)
cephfs::mirror::MirrorWatcher handle_notify
cephfs::mirror::Mirror update_fs_mirrors
cephfs::mirror::Mirror schedule_mirror_update_task: scheduling fs mirror
update (0x556fe3a7f130) after 2 seconds
cephfs::mirror::Watcher handle_notify: notify_id=751516198184655,
handle=93939050205568, notifier_id=25504530
cephfs::mirror::MirrorWatcher handle_notify
cephfs::mirror::PeerReplayer(19361031-928d-4366-99bd-50df70d3adf1) run:
trying to pick from 1 directories
cephfs::mirror::PeerReplayer(19361031-928d-4366-99bd-50df70d3adf1)
pick_directory
cephfs::mirror::Watcher handle_notify: notify_id=751516198184656,
handle=93939050205568, notifier_id=25504530
cephfs::mirror::MirrorWatcher handle_notify
cephfs::mirror::Mirror update_fs_mirrors
cephfs::mirror::Mirror schedule_mirror_update_task: scheduling fs mirror
update (0x556fe3a7fc70) after 2 seconds
cephfs::mirror::Watcher handle_notify: notify_id=751516198184657,
handle=93939050205568, notifier_id=25504530
cephfs::mirror::MirrorWatcher handle_notify
(...)


Basically, the interesting bit is not captured since it probably
happened sometime back. Could you please set the following:

debug cephfs_mirror = 20
debug client = 20

and restart the mirror daemon? The daemon would start synchronizing
again. When synchronizing stalls, please share the daemon logs. If the
log is huge, you could upload them via ceph-post-file.


If I set debug_client to 20 'huge' is an understatement.

I now have three huge logfiles - one pair with debug_mirror set to 20 
capturing the restart and the point where the sync stalls again and one 
with both mirror and client debug at 20 capturing the  restart - but as 
this setting created ~10GB logs within 20min I reset the client logging 
again to spare our small system disks - if these logs are needed I think 
I will have to set up some remote logging facility.


The observation I made from the scanning the logs:

After the restart the mirror daemon spends some hours comparing the 
incomplete transfers

(Lots of limes with

do_synchronize: 18 entries in stack
do_synchronize: top of stack path=./...(FILENAME)...
do_synchronize: entry=./(FILENAME), data_sync=0, attr_sync=1

then there is the point where
the number of items in the stack goes down:
cephfs::mirror::PeerReplayer () do_synchronize: 8 entries in stack

The top of stack moves up in the diretory levesl:

cephfs::mirror::PeerReplayer () do_synchronize: top of stack 
path=./...FILENAME...


but then it just stops, without any error message visible in the logfile

and switches to the repeating sequence I posted already.

Should I try to upload the logs - even gzipped they are quite huge:
 388M syslog.restart-mirror-client_debug_20.gz
  98M syslog.restart-mirror-no_client_debug.gz
  54M syslog.stalled-no_client_debug.gz

(as our servers are in an isolated net I will have to see if the
'ceph-post-file' method works from another system.

Regards,
Holger Naundorf







Regards,
Holger

--
Dr. Holger Naundorf
Christian-Albrechts-Universität zu Kiel
Rechenzentrum / HPC / Server und Storage
Tel: +49 431 880-1990
Fax:  +49 431 880-1523
naund...@rz.uni-kiel.de
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io






--
Dr. Holger Naundorf
Christian-Albrechts-Universität zu Kiel
Rechenzentrum / HPC / Server und Storage
Tel: +49 431 880-1990
Fax:  +49 431 880-1523
naund...@rz.uni-kiel.de






--
Dr. Holger Naundorf
Christian-Albr

[ceph-users] Re: cephfs snap-mirror stalled

2022-12-15 Thread Holger Naundorf



On 15.12.22 14:06, Venky Shankar wrote:

Hi Holger,

(sorry for the late reply)

On Fri, Dec 9, 2022 at 6:22 PM Holger Naundorf  wrote:


As an update:
After the third restart now the mirror-daemon is running normal again -
only change to the restarts before was that during the restart
dbug_client was set to 20. (First restart was after ~48h of no data
movemnent to the receiver side and no changes, 2nd was with debug_mirror
set to 20 and the thisrd - maybe to quickly - after only ~12h of no data
movement with debug_mirror + _client set to 20). Currently I reset the
debug levels back down to reduce the load on the system disks.


Do you see ESTALE related messages in mirror daemon logs (with
debug_client: 20)? We have run into a couple of instances where the
mirror daemon would be stuck on a directory entry. The workaround for
that is to find the directory path (where the daemon is stuck) and
from another client (mount), list the entries in that directory (or at
times restarting the daemon works, as in your case).


Sounds like our bug - I do have the ESTALE messages with debug 20:

syslog.restart-mirror-client_debug_20.gz:Dec  7 10:05:30 ceph07 
bash[1740788]: debug 2022-12-07T09:05:30.299+ 7f05907a7700 20 
client.29690499 got ESTALE on tid 2258671 from mds.0
syslog.restart-mirror-client_debug_20.gz:Dec  7 10:05:30 ceph07 
bash[1740788]: debug 2022-12-07T09:05:30.299+ 7f05907a7700 20 
client.29690499 got ESTALE on tid 2258671 from mds.0
syslog.restart-mirror-client_debug_20.gz:Dec  7 10:05:30 ceph07 
bash[1740788]: debug 2022-12-07T09:05:30.299+ 7f05907a7700 20 
client.29690499 got ESTALE on tid 2258671 from mds.0

(...)

And when the mirror got stuck again after some time it mysteriously (at 
that time) started to work again while I was poking around in the 
logfiles and the system - and doing an 'ls' in the stuck dir was 
definitly included in the poking.




This will be fixed in the next pacific release (tracker:
https://tracker.ceph.com/issues/55935).


Next means the .11 release or the current .10 we have not yet upgraded to?

Thanks for the update.

Regards,
Holger Naundorf




Regards,
Holger Naundorf


On 07.12.22 15:53, Holger Naundorf wrote:

On 06.12.22 14:17, Venky Shankar wrote:

On Tue, Dec 6, 2022 at 6:34 PM Holger Naundorf
 wrote:




On 06.12.22 09:54, Venky Shankar wrote:

Hi Holger,

On Tue, Dec 6, 2022 at 1:42 PM Holger Naundorf
 wrote:


Hello,
we have set up a snap-mirror for a directory on one of our clusters -
running ceph version

ceph version 16.2.7 (dd0603118f56ab514f133c8d2e3adfc983942503) pacific
(stable)

to get mirrorred our other cluster - running ceph version

ceph version 16.2.9 (4c3647a322c0ff5a1dd2344e039859dcbd28c830) pacific
(stable)

The initial setup went ok, when the first snapshot was created data
started to flow at a decent (for our HW) rate of 100-200MB/s. As the
directory contains  ~200TB this was expected to take some time -
but now
the process has stalled completely after ~100TB were mirrored and ~7d
running.

Up to now I do not have any hints why it has stopped - I do not see
any
error messages from the cephfs-mirror daemon. Can the small version
mismatch be a problem?

Any hints where to look to find out what has got stuck are welcome.


I'd look at the mirror daemon logs for any errors to start with. You
might want to crank up the log level for debugging (debug
cephfs_mirror=20).



Even on max debug I do not see anything which looks like an error - but
as this is the first time I try to dig into any cephfs-mirror logs I
might not notice (as long as it is not red and flashing).

The Log basically this type of sequence, repeating forever:

(...)
cephfs::mirror::MirrorWatcher handle_notify
cephfs::mirror::Mirror update_fs_mirrors
cephfs::mirror::Mirror schedule_mirror_update_task: scheduling fs mirror
update (0x556fe3a7f130) after 2 seconds
cephfs::mirror::Watcher handle_notify: notify_id=751516198184655,
handle=93939050205568, notifier_id=25504530
cephfs::mirror::MirrorWatcher handle_notify
cephfs::mirror::PeerReplayer(19361031-928d-4366-99bd-50df70d3adf1) run:
trying to pick from 1 directories
cephfs::mirror::PeerReplayer(19361031-928d-4366-99bd-50df70d3adf1)
pick_directory
cephfs::mirror::Watcher handle_notify: notify_id=751516198184656,
handle=93939050205568, notifier_id=25504530
cephfs::mirror::MirrorWatcher handle_notify
cephfs::mirror::Mirror update_fs_mirrors
cephfs::mirror::Mirror schedule_mirror_update_task: scheduling fs mirror
update (0x556fe3a7fc70) after 2 seconds
cephfs::mirror::Watcher handle_notify: notify_id=751516198184657,
handle=93939050205568, notifier_id=25504530
cephfs::mirror::MirrorWatcher handle_notify
(...)


Basically, the interesting bit is not captured since it probably
happened sometime back. Could you please set the following:

debug cephfs_mirror = 20
debug client = 20

and restart the mirror daemon? The daemon would start synchronizing
again. When synchronizing stalls, please share the daem