[ceph-users] OSD hearbeat_check failure while using 10Gb/s

2024-06-17 Thread Sarunas Burdulis
r keeps running, but it is not booted. Has anyone experienced this? Any ideas on what should be fixed? Please let me know what other info would be useful. Best regards, -- Sarunas Burdulis Dartmouth Mathematics math.dartmouth.edu/~sarunas · https://useplaintext.email · OpenPGP_signature.as

[ceph-users] Re: OSD hearbeat_check failure while using 10Gb/s

2024-06-18 Thread Sarunas Burdulis
gone. -- Sarunas Burdulis Dartmouth Mathematics math.dartmouth.edu/~sarunas · https://useplaintext.email · OpenPGP_signature.asc Description: OpenPGP digital signature ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to

[ceph-users] Re: Cluster healthy, but 16.2.7 osd daemon upgrade says its unsafe to stop them?

2022-05-25 Thread Sarunas Burdulis
s are or would become offline)' appears in the logs. Hi, Has there been any solution or workaround to this? We have a seemingly healthy cluster, which is stuck on OSD upgrade step when upgrading from 15.2.16 to 16.2.8 with the same error(s). -- Sarunas Burdulis Dartmouth M

[ceph-users] Re: Cluster healthy, but 16.2.7 osd daemon upgrade says its unsafe to stop them?

2022-05-25 Thread Sarunas Burdulis
On 25/05/2022 15.39, Tim Olow wrote: Do you have any pools with only one replica? All pools are 'replicated size' 2 or 3, 'min_size' 1 or 2. -- Sarunas Burdulis Dartmouth Mathematics https://math.dartmouth.edu/~sarunas · https:

[ceph-users] Re: Cluster healthy, but 16.2.7 osd daemon upgrade says its unsafe to stop them?

2022-05-26 Thread Sarunas Burdulis
36.13","41.7","41.1b","42.6","42.1a"]} Error EBUSY: unsafe to stop osd(s) at this time (2 PGs are or would become offline) What are “bad_become_inactive” PGs? What can be done to make OSD into “ok-to-stop” (or override it)? `ceph -s` still reports HE

[ceph-users] Re: Cluster healthy, but 16.2.7 osd daemon upgrade says its unsafe to stop them?

2022-05-26 Thread Sarunas Burdulis
On 5/26/22 14:09, Wesley Dillingham wrote: What does "ceph osd pool ls detail" say? $ ceph osd pool ls detail pool 0 'rbd' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 64 pgp_num 64 autoscale_mode on last_change 44740 flags hashpspool,selfmanaged_snaps stripe_width 0

[ceph-users] Re: Cluster healthy, but 16.2.7 osd daemon upgrade says its unsafe to stop them?

2022-05-27 Thread Sarunas Burdulis
k-to-stop and upgrade to 16.2.8 completed). -- Sarunas Burdulis Dartmouth Mathematics math.dartmouth.edu/~sarunas · https://useplaintext.email · ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Cluster healthy, but 16.2.7 osd daemon upgrade says its unsafe to stop them?

2022-05-27 Thread Sarunas Burdulis
On 5/27/22 04:54, Robert Sander wrote: Am 26.05.22 um 20:21 schrieb Sarunas Burdulis: size 2 min_size 1 With such a setting you are guaranteed to lose data. What would you suggest? -- Sarunas Burdulis Dartmouth Mathematics math.dartmouth.edu/~sarunas · https://useplaintext.email

[ceph-users] Re: Cluster healthy, but 16.2.7 osd daemon upgrade says its unsafe to stop them?

2022-05-27 Thread Sarunas Burdulis
to have a min_size of 2 and replica 3 per replicated pool. Thanks. I don't recall creating any of the default.* pools, so they might have created by ceph-deploy, years ago (kraken?). They all have min_size 1, replica 2. -- Sarunas Burdulis Dartmouth Mathematics math.dartmouth.edu/~sa