[ceph-users] Re: OSDs cannot join cluster anymore

2023-06-24 Thread Adam King
Reminds me of https://tracker.ceph.com/issues/57007 which wasn't fixed in
pacific until 16.2.11, so this is probably just the result of a cephadm bug
unfortunately.

On Fri, Jun 23, 2023 at 5:16 PM Malte Stroem  wrote:

> Hello Eugen,
>
> thanks.
>
> We found the cause.
>
> Somehow all
>
> /var/lib/ceph/fsid/osd.XX/config
>
> files on every host were still filled with expired information about the
> mons.
>
> So refreshing the files helped to bring the osds up again. Damn.
>
> All other configs for the mons, mds', rgws and so on were up to date.
>
> I do not know why the osd config files did not get refreshed however I
> guess something went wrong draining the nodes we removed from the cluster.
>
> Best regards,
> Malte
>
> Am 21.06.23 um 22:11 schrieb Eugen Block:
> > I still can’t really grasp what might have happened here. But could you
> > please clarify which of the down OSDs (or Hosts) are supposed to be down
> > and which you’re trying to bring back online? Obviously osd.40 is one of
> > your attempts. But what about the hosts cephx01 and cephx08? Are those
> > the ones refusing to start their OSDs? And the remaining up OSDs you
> > haven’t touched yet, correct?
> > And regarding debug logs, you should set it with ceph config set because
> > the local ceph.conf won’t have an effect. It could help to have the
> > startup debug logs from one of the OSDs.
> >
> > Zitat von Malte Stroem :
> >
> >> Hello Eugen,
> >>
> >> recovery and rebalancing was finished however now all PGs show missing
> >> OSDs.
> >>
> >> Everything looks like the PGs are missing OSDs although it finished
> >> correctly.
> >>
> >> As if we shut down the servers immediately.
> >>
> >> But we removed the nodes the way it is described in the documentation.
> >>
> >> We just added new disks and they join the cluster immediately.
> >>
> >> So the old OSDs removed from the cluster are available, I restored
> >> OSD.40 but it does not want to join the cluster.
> >>
> >> Following are the outputs of the mentioned commands:
> >>
> >> ceph -s
> >>
> >>   cluster:
> >> id: X
> >> health: HEALTH_WARN
> >> 1 failed cephadm daemon(s)
> >> 1 filesystem is degraded
> >> 1 MDSs report slow metadata IOs
> >> 19 osds down
> >> 4 hosts (50 osds) down
> >> Reduced data availability: 1220 pgs inactive
> >> Degraded data redundancy: 132 pgs undersized
> >>
> >>   services:
> >> mon: 3 daemons, quorum cephx02,cephx04,cephx06 (age 4m)
> >> mgr: cephx02.xx(active, since 92s), standbys: cephx04.yy,
> >> cephx06.zz mds: 2/2 daemons up, 2 standby
> >> osd: 130 osds: 78 up (since 13m), 97 in (since 35m); 171 remapped
> pgs
> >> rgw: 1 daemon active (1 hosts, 1 zones)
> >>
> >>   data:
> >> volumes: 1/2 healthy, 1 recovering
> >> pools:   12 pools, 1345 pgs
> >> objects: 11.02k objects, 1.9 GiB
> >> usage:   145 TiB used, 669 TiB / 814 TiB avail
> >> pgs: 86.617% pgs unknown
> >>  4.089% pgs not active
> >>  39053/33069 objects misplaced (118.095%)
> >>  1165 unknown
> >>  77   active+undersized+remapped
> >>  55   undersized+remapped+peered
> >>  38   active+clean+remapped
> >>  10   active+clean
> >>
> >> ceph osd tree
> >>
> >> ID   CLASS  WEIGHT  TYPE NAMESTATUS  REWEIGHT
> >> PRI-AFF
> >> -214.36646  root ssds
> >> -610.87329  host cephx01-ssd
> >> 186ssd 0.87329  osd.186down   1.0
> >> 1.0
> >> -760.87329  host cephx02-ssd
> >> 263ssd 0.87329  osd.263  up   1.0
> >> 1.0
> >> -850.87329  host cephx04-ssd
> >> 237ssd 0.87329  osd.237  up   1.0
> >> 1.0
> >> -880.87329  host cephx06-ssd
> >> 236ssd 0.87329  osd.236  up   1.0
> >> 1.0
> >> -940.87329  host cephx08-ssd
> >> 262ssd 0.87329  osd.262down   1.0
> >> 1.0
> >>  -1 1347.07397  root default
> >> -62  261.93823  host cephx01
> >> 139hdd10.91409  osd.139down 0
> >> 1.0
> >> 140hdd10.91409  osd.140down 0
> >> 1.0
> >> 142hdd10.91409  osd.142down 0
> >> 1.0
> >> 144hdd10.91409  osd.144down 0
> >> 1.0
> >> 146hdd10.91409  osd.146down 0
> >> 1.0
> >> 148hdd10.91409  osd.148down 0
> >> 1.0
> >> 150hdd10.91409  osd.150down 0
> >> 1.0
> >> 152hdd10.91409  osd.152down 0
> >> 1.0
> >> 154hdd10.91409  osd.154down   1.0
> >> 

[ceph-users] Re: OSDs cannot join cluster anymore

2023-06-24 Thread Eugen Block
Oh, okay. I believe there was a thread reporting something very  
similar as well some time ago. I don’t remember the details but having  
outdated information on the OSDs was part of it. Were the nodes you  
removed also MON nodes?

But it’s great that you found the root cause.

Zitat von Malte Stroem :


Hello Eugen,

thanks.

We found the cause.

Somehow all

/var/lib/ceph/fsid/osd.XX/config

files on every host were still filled with expired information about  
the mons.


So refreshing the files helped to bring the osds up again. Damn.

All other configs for the mons, mds', rgws and so on were up to date.

I do not know why the osd config files did not get refreshed however  
I guess something went wrong draining the nodes we removed from the  
cluster.


Best regards,
Malte

Am 21.06.23 um 22:11 schrieb Eugen Block:
I still can’t really grasp what might have happened here. But could  
you please clarify which of the down OSDs (or Hosts) are supposed  
to be down and which you’re trying to bring back online? Obviously  
osd.40 is one of your attempts. But what about the hosts cephx01  
and cephx08? Are those the ones refusing to start their OSDs? And  
the remaining up OSDs you haven’t touched yet, correct?
And regarding debug logs, you should set it with ceph config set  
because the local ceph.conf won’t have an effect. It could help to  
have the startup debug logs from one of the OSDs.


Zitat von Malte Stroem :


Hello Eugen,

recovery and rebalancing was finished however now all PGs show  
missing OSDs.


Everything looks like the PGs are missing OSDs although it  
finished correctly.


As if we shut down the servers immediately.

But we removed the nodes the way it is described in the documentation.

We just added new disks and they join the cluster immediately.

So the old OSDs removed from the cluster are available, I restored  
OSD.40 but it does not want to join the cluster.


Following are the outputs of the mentioned commands:

ceph -s

  cluster:
    id: X
    health: HEALTH_WARN
    1 failed cephadm daemon(s)
    1 filesystem is degraded
    1 MDSs report slow metadata IOs
    19 osds down
    4 hosts (50 osds) down
    Reduced data availability: 1220 pgs inactive
    Degraded data redundancy: 132 pgs undersized

  services:
    mon: 3 daemons, quorum cephx02,cephx04,cephx06 (age 4m)
    mgr: cephx02.xx(active, since 92s), standbys:  
cephx04.yy, cephx06.zz mds: 2/2 daemons up, 2 standby

    osd: 130 osds: 78 up (since 13m), 97 in (since 35m); 171 remapped pgs
    rgw: 1 daemon active (1 hosts, 1 zones)

  data:
    volumes: 1/2 healthy, 1 recovering
    pools:   12 pools, 1345 pgs
    objects: 11.02k objects, 1.9 GiB
    usage:   145 TiB used, 669 TiB / 814 TiB avail
    pgs: 86.617% pgs unknown
 4.089% pgs not active
 39053/33069 objects misplaced (118.095%)
 1165 unknown
 77   active+undersized+remapped
 55   undersized+remapped+peered
 38   active+clean+remapped
 10   active+clean

ceph osd tree

ID   CLASS  WEIGHT  TYPE NAME    STATUS  REWEIGHT  PRI-AFF
-21    4.36646  root ssds
-61    0.87329  host cephx01-ssd
186    ssd 0.87329  osd.186    down   1.0  1.0
-76    0.87329  host cephx02-ssd
263    ssd 0.87329  osd.263  up   1.0  1.0
-85    0.87329  host cephx04-ssd
237    ssd 0.87329  osd.237  up   1.0  1.0
-88    0.87329  host cephx06-ssd
236    ssd 0.87329  osd.236  up   1.0  1.0
-94    0.87329  host cephx08-ssd
262    ssd 0.87329  osd.262    down   1.0  1.0
 -1 1347.07397  root default
-62  261.93823  host cephx01
139    hdd    10.91409  osd.139    down 0  1.0
140    hdd    10.91409  osd.140    down 0  1.0
142    hdd    10.91409  osd.142    down 0  1.0
144    hdd    10.91409  osd.144    down 0  1.0
146    hdd    10.91409  osd.146    down 0  1.0
148    hdd    10.91409  osd.148    down 0  1.0
150    hdd    10.91409  osd.150    down 0  1.0
152    hdd    10.91409  osd.152    down 0  1.0
154    hdd    10.91409  osd.154    down   1.0  1.0
156    hdd    10.91409  osd.156    down   1.0  1.0
158    hdd    10.91409  osd.158    down   1.0  1.0
160    hdd    10.91409  osd.160    down   1.0  1.0
162    hdd    10.91409  osd.162    down   1.0  1.0
164    hdd    10.91409  osd.164    down   1.0  1.0
166    hdd    10.91409 

[ceph-users] Re: OSDs cannot join cluster anymore

2023-06-23 Thread Malte Stroem

Hello Eugen,

thanks.

We found the cause.

Somehow all

/var/lib/ceph/fsid/osd.XX/config

files on every host were still filled with expired information about the 
mons.


So refreshing the files helped to bring the osds up again. Damn.

All other configs for the mons, mds', rgws and so on were up to date.

I do not know why the osd config files did not get refreshed however I 
guess something went wrong draining the nodes we removed from the cluster.


Best regards,
Malte

Am 21.06.23 um 22:11 schrieb Eugen Block:
I still can’t really grasp what might have happened here. But could you 
please clarify which of the down OSDs (or Hosts) are supposed to be down 
and which you’re trying to bring back online? Obviously osd.40 is one of 
your attempts. But what about the hosts cephx01 and cephx08? Are those 
the ones refusing to start their OSDs? And the remaining up OSDs you 
haven’t touched yet, correct?
And regarding debug logs, you should set it with ceph config set because 
the local ceph.conf won’t have an effect. It could help to have the 
startup debug logs from one of the OSDs.


Zitat von Malte Stroem :


Hello Eugen,

recovery and rebalancing was finished however now all PGs show missing 
OSDs.


Everything looks like the PGs are missing OSDs although it finished 
correctly.


As if we shut down the servers immediately.

But we removed the nodes the way it is described in the documentation.

We just added new disks and they join the cluster immediately.

So the old OSDs removed from the cluster are available, I restored 
OSD.40 but it does not want to join the cluster.


Following are the outputs of the mentioned commands:

ceph -s

  cluster:
    id: X
    health: HEALTH_WARN
    1 failed cephadm daemon(s)
    1 filesystem is degraded
    1 MDSs report slow metadata IOs
    19 osds down
    4 hosts (50 osds) down
    Reduced data availability: 1220 pgs inactive
    Degraded data redundancy: 132 pgs undersized

  services:
    mon: 3 daemons, quorum cephx02,cephx04,cephx06 (age 4m)
    mgr: cephx02.xx(active, since 92s), standbys: cephx04.yy, 
cephx06.zz mds: 2/2 daemons up, 2 standby

    osd: 130 osds: 78 up (since 13m), 97 in (since 35m); 171 remapped pgs
    rgw: 1 daemon active (1 hosts, 1 zones)

  data:
    volumes: 1/2 healthy, 1 recovering
    pools:   12 pools, 1345 pgs
    objects: 11.02k objects, 1.9 GiB
    usage:   145 TiB used, 669 TiB / 814 TiB avail
    pgs: 86.617% pgs unknown
 4.089% pgs not active
 39053/33069 objects misplaced (118.095%)
 1165 unknown
 77   active+undersized+remapped
 55   undersized+remapped+peered
 38   active+clean+remapped
 10   active+clean

ceph osd tree

ID   CLASS  WEIGHT  TYPE NAME    STATUS  REWEIGHT  
PRI-AFF

-21    4.36646  root ssds
-61    0.87329  host cephx01-ssd
186    ssd 0.87329  osd.186    down   1.0  
1.0

-76    0.87329  host cephx02-ssd
263    ssd 0.87329  osd.263  up   1.0  
1.0

-85    0.87329  host cephx04-ssd
237    ssd 0.87329  osd.237  up   1.0  
1.0

-88    0.87329  host cephx06-ssd
236    ssd 0.87329  osd.236  up   1.0  
1.0

-94    0.87329  host cephx08-ssd
262    ssd 0.87329  osd.262    down   1.0  
1.0

 -1 1347.07397  root default
-62  261.93823  host cephx01
139    hdd    10.91409  osd.139    down 0  
1.0
140    hdd    10.91409  osd.140    down 0  
1.0
142    hdd    10.91409  osd.142    down 0  
1.0
144    hdd    10.91409  osd.144    down 0  
1.0
146    hdd    10.91409  osd.146    down 0  
1.0
148    hdd    10.91409  osd.148    down 0  
1.0
150    hdd    10.91409  osd.150    down 0  
1.0
152    hdd    10.91409  osd.152    down 0  
1.0
154    hdd    10.91409  osd.154    down   1.0  
1.0
156    hdd    10.91409  osd.156    down   1.0  
1.0
158    hdd    10.91409  osd.158    down   1.0  
1.0
160    hdd    10.91409  osd.160    down   1.0  
1.0
162    hdd    10.91409  osd.162    down   1.0  
1.0
164    hdd    10.91409  osd.164    down   1.0  
1.0
166    hdd    10.91409  osd.166    down   1.0  
1.0
168    hdd    10.91409  osd.168    down   1.0  
1.0
170    hdd    10.91409  osd.170    down   1.0  
1.0
172    hdd    10.91409  osd.172    down   1.0  
1.0
174    hdd    

[ceph-users] Re: OSDs cannot join cluster anymore

2023-06-22 Thread Stefan Kooman

On 6/21/23 11:20, Malte Stroem wrote:

Hello Eugen,

recovery and rebalancing was finished however now all PGs show missing 
OSDs.


Everything looks like the PGs are missing OSDs although it finished 
correctly.


As if we shut down the servers immediately.

But we removed the nodes the way it is described in the documentation.

We just added new disks and they join the cluster immediately.

So the old OSDs removed from the cluster are available, I restored 
OSD.40 but it does not want to join the cluster.



Are the osd.$id keys still there of the removed OSDs (check with ceph 
auth list)? Otherwise you might need to import the keyring into the 
cluster (/var/lib/ceph/osd/ceph-$id/keyring) and provide it proper CAPS.


Gr. Stefan
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: OSDs cannot join cluster anymore

2023-06-21 Thread Eugen Block
I still can’t really grasp what might have happened here. But could  
you please clarify which of the down OSDs (or Hosts) are supposed to  
be down and which you’re trying to bring back online? Obviously osd.40  
is one of your attempts. But what about the hosts cephx01 and cephx08?  
Are those the ones refusing to start their OSDs? And the remaining up  
OSDs you haven’t touched yet, correct?
And regarding debug logs, you should set it with ceph config set  
because the local ceph.conf won’t have an effect. It could help to  
have the startup debug logs from one of the OSDs.


Zitat von Malte Stroem :


Hello Eugen,

recovery and rebalancing was finished however now all PGs show missing OSDs.

Everything looks like the PGs are missing OSDs although it finished  
correctly.


As if we shut down the servers immediately.

But we removed the nodes the way it is described in the documentation.

We just added new disks and they join the cluster immediately.

So the old OSDs removed from the cluster are available, I restored  
OSD.40 but it does not want to join the cluster.


Following are the outputs of the mentioned commands:

ceph -s

  cluster:
id: X
health: HEALTH_WARN
1 failed cephadm daemon(s)
1 filesystem is degraded
1 MDSs report slow metadata IOs
19 osds down
4 hosts (50 osds) down
Reduced data availability: 1220 pgs inactive
Degraded data redundancy: 132 pgs undersized

  services:
mon: 3 daemons, quorum cephx02,cephx04,cephx06 (age 4m)
mgr: cephx02.xx(active, since 92s), standbys:  
cephx04.yy, cephx06.zz mds: 2/2 daemons up, 2 standby

osd: 130 osds: 78 up (since 13m), 97 in (since 35m); 171 remapped pgs
rgw: 1 daemon active (1 hosts, 1 zones)

  data:
volumes: 1/2 healthy, 1 recovering
pools:   12 pools, 1345 pgs
objects: 11.02k objects, 1.9 GiB
usage:   145 TiB used, 669 TiB / 814 TiB avail
pgs: 86.617% pgs unknown
 4.089% pgs not active
 39053/33069 objects misplaced (118.095%)
 1165 unknown
 77   active+undersized+remapped
 55   undersized+remapped+peered
 38   active+clean+remapped
 10   active+clean

ceph osd tree

ID   CLASS  WEIGHT  TYPE NAMESTATUS  REWEIGHT  PRI-AFF
-214.36646  root ssds
-610.87329  host cephx01-ssd
186ssd 0.87329  osd.186down   1.0  1.0
-760.87329  host cephx02-ssd
263ssd 0.87329  osd.263  up   1.0  1.0
-850.87329  host cephx04-ssd
237ssd 0.87329  osd.237  up   1.0  1.0
-880.87329  host cephx06-ssd
236ssd 0.87329  osd.236  up   1.0  1.0
-940.87329  host cephx08-ssd
262ssd 0.87329  osd.262down   1.0  1.0
 -1 1347.07397  root default
-62  261.93823  host cephx01
139hdd10.91409  osd.139down 0  1.0
140hdd10.91409  osd.140down 0  1.0
142hdd10.91409  osd.142down 0  1.0
144hdd10.91409  osd.144down 0  1.0
146hdd10.91409  osd.146down 0  1.0
148hdd10.91409  osd.148down 0  1.0
150hdd10.91409  osd.150down 0  1.0
152hdd10.91409  osd.152down 0  1.0
154hdd10.91409  osd.154down   1.0  1.0
156hdd10.91409  osd.156down   1.0  1.0
158hdd10.91409  osd.158down   1.0  1.0
160hdd10.91409  osd.160down   1.0  1.0
162hdd10.91409  osd.162down   1.0  1.0
164hdd10.91409  osd.164down   1.0  1.0
166hdd10.91409  osd.166down   1.0  1.0
168hdd10.91409  osd.168down   1.0  1.0
170hdd10.91409  osd.170down   1.0  1.0
172hdd10.91409  osd.172down   1.0  1.0
174hdd10.91409  osd.174down   1.0  1.0
176hdd10.91409  osd.176down   1.0  1.0
178hdd10.91409  osd.178down   1.0  1.0
180hdd10.91409  osd.180down   1.0  1.0
182hdd10.91409  osd.182down   1.0  1.0
184hdd10.91409  osd.184down   1.0  1.0
-67  261.93823  host cephx02
138hdd10.91409  osd.138  up   1.0  

[ceph-users] Re: OSDs cannot join cluster anymore

2023-06-21 Thread Malte Stroem

Hello Eugen,

recovery and rebalancing was finished however now all PGs show missing OSDs.

Everything looks like the PGs are missing OSDs although it finished 
correctly.


As if we shut down the servers immediately.

But we removed the nodes the way it is described in the documentation.

We just added new disks and they join the cluster immediately.

So the old OSDs removed from the cluster are available, I restored 
OSD.40 but it does not want to join the cluster.


Following are the outputs of the mentioned commands:

ceph -s

  cluster:
id: X
health: HEALTH_WARN
1 failed cephadm daemon(s)
1 filesystem is degraded
1 MDSs report slow metadata IOs
19 osds down
4 hosts (50 osds) down
Reduced data availability: 1220 pgs inactive
Degraded data redundancy: 132 pgs undersized

  services:
mon: 3 daemons, quorum cephx02,cephx04,cephx06 (age 4m)
mgr: cephx02.xx(active, since 92s), standbys: cephx04.yy, 
cephx06.zz 


mds: 2/2 daemons up, 2 standby
osd: 130 osds: 78 up (since 13m), 97 in (since 35m); 171 remapped pgs
rgw: 1 daemon active (1 hosts, 1 zones)

  data:
volumes: 1/2 healthy, 1 recovering
pools:   12 pools, 1345 pgs
objects: 11.02k objects, 1.9 GiB
usage:   145 TiB used, 669 TiB / 814 TiB avail
pgs: 86.617% pgs unknown
 4.089% pgs not active
 39053/33069 objects misplaced (118.095%)
 1165 unknown
 77   active+undersized+remapped
 55   undersized+remapped+peered
 38   active+clean+remapped
 10   active+clean

ceph osd tree

ID   CLASS  WEIGHT  TYPE NAMESTATUS  REWEIGHT  PRI-AFF
-214.36646  root ssds
-610.87329  host cephx01-ssd
186ssd 0.87329  osd.186down   1.0  1.0
-760.87329  host cephx02-ssd
263ssd 0.87329  osd.263  up   1.0  1.0
-850.87329  host cephx04-ssd
237ssd 0.87329  osd.237  up   1.0  1.0
-880.87329  host cephx06-ssd
236ssd 0.87329  osd.236  up   1.0  1.0
-940.87329  host cephx08-ssd
262ssd 0.87329  osd.262down   1.0  1.0
 -1 1347.07397  root default
-62  261.93823  host cephx01
139hdd10.91409  osd.139down 0  1.0
140hdd10.91409  osd.140down 0  1.0
142hdd10.91409  osd.142down 0  1.0
144hdd10.91409  osd.144down 0  1.0
146hdd10.91409  osd.146down 0  1.0
148hdd10.91409  osd.148down 0  1.0
150hdd10.91409  osd.150down 0  1.0
152hdd10.91409  osd.152down 0  1.0
154hdd10.91409  osd.154down   1.0  1.0
156hdd10.91409  osd.156down   1.0  1.0
158hdd10.91409  osd.158down   1.0  1.0
160hdd10.91409  osd.160down   1.0  1.0
162hdd10.91409  osd.162down   1.0  1.0
164hdd10.91409  osd.164down   1.0  1.0
166hdd10.91409  osd.166down   1.0  1.0
168hdd10.91409  osd.168down   1.0  1.0
170hdd10.91409  osd.170down   1.0  1.0
172hdd10.91409  osd.172down   1.0  1.0
174hdd10.91409  osd.174down   1.0  1.0
176hdd10.91409  osd.176down   1.0  1.0
178hdd10.91409  osd.178down   1.0  1.0
180hdd10.91409  osd.180down   1.0  1.0
182hdd10.91409  osd.182down   1.0  1.0
184hdd10.91409  osd.184down   1.0  1.0
-67  261.93823  host cephx02
138hdd10.91409  osd.138  up   1.0  1.0
141hdd10.91409  osd.141  up   1.0  1.0
143hdd10.91409  osd.143  up   1.0  1.0
145hdd10.91409  osd.145  up   1.0  1.0
147hdd10.91409  osd.147  up   1.0  1.0
149hdd10.91409  osd.149  up   1.0  1.0
151hdd10.91409  osd.151  up   1.0  1.0
153hdd10.91409  osd.153  up   1.0  1.0
155hdd10.91409  osd.155  up   1.0  1.0
157

[ceph-users] Re: OSDs cannot join cluster anymore

2023-06-21 Thread Eugen Block

Hi,

Yes, we drained the nodes. It needed two weeks to finish the  
process, and yes, I think this is the root cause.
So we still have the nodes but when I try to restart one of those  
OSDs it still cannot join:


if the nodes were drained successfully (can you confirm that all PGs  
were active+clean after draining before you removed the nodes?) then  
the disks on the removed nodes wouldn't have any data to bring back.  
The question would be, why do the remaining OSDs still reference  
removed OSDs. Or am I misunderstanding something? I think it would  
help to know the whole story, can you provide more details? Also some  
more general cluster info would be helpful:

$ ceph -s
$ ceph osd tree
$ ceph health detail


Zitat von Malte Stroem :


Hello Eugen,

thank you. Yesterday I thought: Well, Eugen can help!

Yes, we drained the nodes. It needed two weeks to finish the  
process, and yes, I think this is the root cause.


So we still have the nodes but when I try to restart one of those  
OSDs it still cannot join:


Jun 21 09:46:03 ceph-node bash[2323668]: Running command:  
/usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-66/block
Jun 21 09:46:03 ceph-node bash[2323668]: Running command:  
/usr/bin/chown -R ceph:ceph /dev/dm-19
Jun 21 09:46:03 ceph-node bash[2323668]: Running command:  
/usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-66
Jun 21 09:46:03 ceph-node bash[2323668]: --> ceph-volume lvm  
activate successful for osd ID: 66 Jun 21 09:51:04 ceph-node  
bash[2323668]: debug 2023-06-21T07:51:04.176+ 7fabef5a1200  0  
monclient(hunting): authenticate timed out after 300
Jun 21 09:56:04 ceph-node bash[2323668]: debug  
2023-06-21T07:56:04.179+ 7fabef5a1200  0 monclient(hunting):  
authenticate timed out after 300
Jun 21 10:01:04 ceph-node bash[2323668]: debug  
2023-06-21T08:01:04.177+ 7fabef5a1200  0 monclient(hunting):  
authenticate timed out after 300
Jun 21 10:06:04 ceph-node bash[2323668]: debug  
2023-06-21T08:06:04.179+ 7fabef5a1200  0 monclient(hunting):  
authenticate timed out after 300
Jun 21 10:11:04 ceph-node bash[2323668]: debug  
2023-06-21T08:11:04.174+ 7fabef5a1200  0 monclient(hunting):  
authenticate timed out after 300


Same messages on all OSDs.

We still have some nodes running and did not restart those OSDs.

Best,
Malte

Am 21.06.23 um 09:50 schrieb Eugen Block:

Hi,
can you share more details what exactly you did? How did you remove  
the nodes? Hopefully, you waited for the draining to finish? But if  
the remaining OSDs wait for removed OSDs it sounds like the  
draining was not finished.


Zitat von Malte Stroem :


Hello,

we removed some nodes from our cluster. This worked without problems.

Now, lots of OSDs do not want to join the cluster anymore if we  
reboot one of the still available nodes.


It always runs into timeouts:

--> ceph-volume lvm activate successful for osd ID: XX
monclient(hunting): authenticate timed out after 300

MONs and MGRs are running fine.

Network is working, netcat to the MONs' ports are open.

Setting a higher debug level has no effect even if we add it to  
the ceph.conf file.


The PGs are pretty unhappy, e. g.:

7.143  87771   0 0  0    0  
314744902235    0   0  10081 10081  down   
2023-06-20T09:16:03.546158+    961275'1395646 961300:9605547   
[209,NONE,NONE] 209  [209,NONE,NONE] 209    961231'1395512  
 2023-06-19T23:46:40.101791+    961231'1395512   
2023-06-19T23:46:40.101791+


PG query wants us to set an OSD lost however I do not want to do this.

OSDs are blocked by OSDs from the removed nodes:

ceph osd blocked-by
osd  num_blocked
152   38
244   41
144   54
...

We added the removed hosts again and tried to start the OSDs on  
this node and they also failed into the timeout mentioned above.


This is a containerized cluster running version 16.2.10.

Replication is 3, some pools use an erasure coded profile.

Best regards,
Malte


___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: OSDs cannot join cluster anymore

2023-06-21 Thread Malte Stroem

Hello Eugen,

thank you. Yesterday I thought: Well, Eugen can help!

Yes, we drained the nodes. It needed two weeks to finish the process, 
and yes, I think this is the root cause.


So we still have the nodes but when I try to restart one of those OSDs 
it still cannot join:


Jun 21 09:46:03 ceph-node bash[2323668]: Running command: /usr/bin/chown 
-h ceph:ceph /var/lib/ceph/osd/ceph-66/block
Jun 21 09:46:03 ceph-node bash[2323668]: Running command: /usr/bin/chown 
-R ceph:ceph /dev/dm-19
Jun 21 09:46:03 ceph-node bash[2323668]: Running command: /usr/bin/chown 
-R ceph:ceph /var/lib/ceph/osd/ceph-66
Jun 21 09:46:03 ceph-node bash[2323668]: --> ceph-volume lvm activate 
successful for osd ID: 66 

Jun 21 09:51:04 ceph-node bash[2323668]: debug 
2023-06-21T07:51:04.176+ 7fabef5a1200  0 monclient(hunting): 
authenticate timed out after 300
Jun 21 09:56:04 ceph-node bash[2323668]: debug 
2023-06-21T07:56:04.179+ 7fabef5a1200  0 monclient(hunting): 
authenticate timed out after 300
Jun 21 10:01:04 ceph-node bash[2323668]: debug 
2023-06-21T08:01:04.177+ 7fabef5a1200  0 monclient(hunting): 
authenticate timed out after 300
Jun 21 10:06:04 ceph-node bash[2323668]: debug 
2023-06-21T08:06:04.179+ 7fabef5a1200  0 monclient(hunting): 
authenticate timed out after 300
Jun 21 10:11:04 ceph-node bash[2323668]: debug 
2023-06-21T08:11:04.174+ 7fabef5a1200  0 monclient(hunting): 
authenticate timed out after 300


Same messages on all OSDs.

We still have some nodes running and did not restart those OSDs.

Best,
Malte

Am 21.06.23 um 09:50 schrieb Eugen Block:

Hi,
can you share more details what exactly you did? How did you remove the 
nodes? Hopefully, you waited for the draining to finish? But if the 
remaining OSDs wait for removed OSDs it sounds like the draining was not 
finished.


Zitat von Malte Stroem :


Hello,

we removed some nodes from our cluster. This worked without problems.

Now, lots of OSDs do not want to join the cluster anymore if we reboot 
one of the still available nodes.


It always runs into timeouts:

--> ceph-volume lvm activate successful for osd ID: XX
monclient(hunting): authenticate timed out after 300

MONs and MGRs are running fine.

Network is working, netcat to the MONs' ports are open.

Setting a higher debug level has no effect even if we add it to the 
ceph.conf file.


The PGs are pretty unhappy, e. g.:

7.143  87771   0 0  0    0 
314744902235    0   0  10081 10081  down  
2023-06-20T09:16:03.546158+    961275'1395646 961300:9605547  
[209,NONE,NONE] 209  [209,NONE,NONE] 209    961231'1395512  
2023-06-19T23:46:40.101791+    961231'1395512  
2023-06-19T23:46:40.101791+


PG query wants us to set an OSD lost however I do not want to do this.

OSDs are blocked by OSDs from the removed nodes:

ceph osd blocked-by
osd  num_blocked
152   38
244   41
144   54
...

We added the removed hosts again and tried to start the OSDs on this 
node and they also failed into the timeout mentioned above.


This is a containerized cluster running version 16.2.10.

Replication is 3, some pools use an erasure coded profile.

Best regards,
Malte


___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: OSDs cannot join cluster anymore

2023-06-21 Thread Eugen Block

Hi,
can you share more details what exactly you did? How did you remove  
the nodes? Hopefully, you waited for the draining to finish? But if  
the remaining OSDs wait for removed OSDs it sounds like the draining  
was not finished.


Zitat von Malte Stroem :


Hello,

we removed some nodes from our cluster. This worked without problems.

Now, lots of OSDs do not want to join the cluster anymore if we  
reboot one of the still available nodes.


It always runs into timeouts:

--> ceph-volume lvm activate successful for osd ID: XX
monclient(hunting): authenticate timed out after 300

MONs and MGRs are running fine.

Network is working, netcat to the MONs' ports are open.

Setting a higher debug level has no effect even if we add it to the  
ceph.conf file.


The PGs are pretty unhappy, e. g.:

7.143  87771   0 0  00  
3147449022350   0  10081 10081  down   
2023-06-20T09:16:03.546158+961275'1395646 961300:9605547   
[209,NONE,NONE] 209  [209,NONE,NONE] 209961231'1395512   
2023-06-19T23:46:40.101791+961231'1395512   
2023-06-19T23:46:40.101791+


PG query wants us to set an OSD lost however I do not want to do this.

OSDs are blocked by OSDs from the removed nodes:

ceph osd blocked-by
osd  num_blocked
152   38
244   41
144   54
...

We added the removed hosts again and tried to start the OSDs on this  
node and they also failed into the timeout mentioned above.


This is a containerized cluster running version 16.2.10.

Replication is 3, some pools use an erasure coded profile.

Best regards,
Malte


___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io