[ceph-users] rbd_mirroring_delete_delay not removing images with snaps

2023-01-23 Thread Tyler Brekke
We use the rbd-mirror as a way to migrate volumes between clusters.
The process is enable mirroring on the image to migrate, demote on the
primary cluster, promote on the secondary cluster, and then disable
mirroring on the image.
When we started using `rbd_mirroring_delete_delay` so we could retain a
backup of the source image, we noticed volumes with unprotected snaps do
not get purged from the trash. Previously, the image and all its snaps
would be successfully removed after disabling mirroring.
I would expect a similar function when using `rbd_mirroring_delete_delay`
as well. Is rbd trash just overly cautious here?

-- 
Tyler Brekke
Senior Engineer I
tbre...@digitalocean.com
--
We're Hiring! <https://do.co/careers> | @digitalocean
<https://twitter.com/digitalocean> | YouTube
<https://www.youtube.com/digitalocean>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Odd 10-minute delay before recovery IO begins

2022-12-05 Thread Tyler Brekke
Sounds like your OSDs were down, but not marked out. Recovery will only
occur once they are actually marked out. The default
mon_osd_down_out_interval is 10 minutes.

You can mark them out explicitly with ceph osd out 

On Mon, Dec 5, 2022 at 2:20 PM Sean Matheny 
wrote:

> Hi all,
>
> New Quincy cluster here that I'm just running through some benchmarks
> against:
>
> ceph version 17.2.3 (dff484dfc9e19a9819f375586300b3b79d80034d) quincy
> (stable)
> 11 nodes of 24x 18TB HDD OSDs, 2x 2.9TB SSD OSDs
>
> I'm seeing a delay of almost exactly 10 minutes when I remove an OSD/node
> from the cluster until actual recovery IO begins. This is much different
> behaviour that what I'm used to in Nautilus previously, where recovery IO
> would commence within seconds. Downed OSDs are reflected in ceph health
> within a few seconds (as expected), and affected PGs show as undersized a
> few seconds later (as expected). I guess this 10-minute delay may even be a
> feature-- accidentally rebooting a node before setting recovery flags would
> prevent rebalancing, for example. Just thought it was worth asking in case
> it's a bug or something to look deeper into.
>
> I've read through the OSD config and all of my recovery tuneables look ok,
> for example:
> https://docs.ceph.com/en/latest/rados/configuration/osd-config-ref/
>
> [ceph: root@ /]# ceph config get osd osd_recovery_delay_start
> 20.00
> 3[ceph: root@ /]# ceph config get osd osd_recovery_sleep
> 40.00
> 5[ceph: root@ /]# ceph config get osd osd_recovery_sleep_hdd
> 60.10
> 7[ceph: root@ /]# ceph config get osd osd_recovery_sleep_ssd
> 80.00
> 9[ceph: root@ /]# ceph config get osd osd_recovery_sleep_hybrid
> 100.025000
>
> Thanks in advance.
>
> Ngā mihi,
>
> Sean Matheny
> HPC Cloud Platform DevOps Lead
> New Zealand eScience Infrastructure (NeSI)
>
> e: sean.math...@nesi.org.nz
>
>
>
> _______
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>


-- 
Tyler Brekke
Senior Engineer I
tbre...@digitalocean.com
--
We're Hiring! <https://do.co/careers> | @digitalocean
<https://twitter.com/digitalocean> | YouTube
<https://www.youtube.com/digitalocean>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: all monitors deleted, state recovered using documentation .. at what point to start osds ?

2022-11-10 Thread Tyler Brekke
Hi Shashi, I think you need to have a mgr running to get updated reporting,
which would explain the incorrect ceph status output.

Since you have a monitor quorum 1 out of 1, you can start up OSDs. but I
would recommend getting all your mons/mgrs back up first.

On Tue, Nov 8, 2022 at 5:56 PM Shashi Dahal  wrote:

> Hi,
>
> Unfortunately, all 3 monitors were lost.
> I followed this ->
>
> https://docs.ceph.com/en/quincy/rados/troubleshooting/troubleshooting-mon/#mon-store-recovery-using-osds
> and it is in the current state now.
>
> id: 234c6a96-8101-49d1-b354-1110e759d572
> health: HEALTH_WARN
> mon is allowing insecure global_id reclaim
> no active mgr
>
>   services:
> mon: 1 daemons, quorum mon1 (age 8m)
> mgr: no daemons active
> osd: 40 osds: 40 up (since 5M), 40 in (since 5M)
>
>   data:
> pools:   0 pools, 0 pgs
> objects: 0 objects, 0 B
> usage:   0 B used, 0 B / 0 B avail
> pgs:
>
>
> all OSD daemons are turned off.
>
> It says all 40 osds are up, but the osd services are down.
> If I run  ceph osd dump, it shows all the volumes and pg numbers .. so it
> looks like it knows of all those.
>
> My question is, is it now in a state where it's safe to start an OSD daemon
> ?  Because  ceph status shows no pools, no pgs, I have not turned it on to
> ensure no data loss occurs. At what point would I start the osd daemon?
>
>
> Note:
> If anyone has done something like this before, and can offer (paid)
> assistance/consultation, that is also welcome .
>
> Thanks,
> _______
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>


-- 
Tyler Brekke
Senior Engineer I
tbre...@digitalocean.com
--
We're Hiring! <https://do.co/careers> | @digitalocean
<https://twitter.com/digitalocean> | YouTube
<https://www.youtube.com/digitalocean>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Question about quorum

2022-11-03 Thread Tyler Brekke
Hi Murilo,

Since we need a majority to maintain a quorum when you lost 2 mons, you
only had 50% available and lost quorum. This is why all recommendations
specify having an odd number of mons. As you do not get any added
availability with 4 instead of 3. If you had 5 mons, you can lose two
without losing availability.


On Thu, Nov 3, 2022, 2:55 PM Murilo Morais  wrote:

> Good afternoon everyone!
>
> I have a lab with 4 mons, I was testing the behavior in case a certain
> amount of hosts went offline, as soon as the second one went offline
> everything stopped. It would be interesting if there was a fifth node to
> ensure that, if two fall, everything will work, but why did everything stop
> with only 2 nodes when if there were 3 nodes in the cluster and one fell,
> everything would still be working? Is there no way to get this behavior
> with 4 nodes?
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: 2-Layer CRUSH Map Rule?

2022-09-27 Thread Tyler Brekke
Hi Matthew,

You just have to take two steps when writing your crush rule. First you
want to get 3 different hosts, then you need 2 osd from each host.

ceph osd getcrushmap -o /tmp/crush
crushtool -d /tmp/crush -o /tmp/crush.txt

#edit it / make new rule

rule custom-ec-ruleset {
id 3
type erasure
min_size 4
max_size 6
step take your-root
step choose indep 3 type host
step chooseleaf indep 2 type osd
step emit
}

crushtool -c /tmp/crush.txt -o /tmp/crush.new

You can use the crushtool to test the mappings and make sure they are
working as you expect.

crushtool  -i /tmp/crush.new --test --show-mappings --rule 3 --num-rep 6

You can then compare the OSD id and make sure its exactly what you are
looking for.

You can set the crushmap with

ceph osd setcrushmap -i /tmp/crush.new

Note: If you are overwriting your current rule, your data will need to
rebalance as soon as your set the crushmap, close to 100% of your objects
will move. If you create a new rule, you can set your pool to use the new
pool id anytime you are ready.

On Sun, Sep 25, 2022 at 12:49 AM duluxoz  wrote:

> Hi Everybody (Hi Dr. Nick),
>
> TL/DR: Is is possible to have a "2-Layer" Crush Map?
>
> I think it is (although I'm not sure about how to set it up).
>
> My issue is that we're using 4-2 Erasure coding on our OSDs, with 7-OSDs
> per OSD-Node (yes, the Cluster is handling things AOK - we're running at
> about 65-70% utilisation of CPU, RAM, etc, so no problem there).
>
> However, we only have 3 OSD-Nodes, and I'd like to ensure that each Node
> has 2 of each pool's OSDs so that if we loose a Node the other 2 can
> take up the slack. I know(?) that with 4-2 EC we can loose 2 out of the
> 6 OSDs, but I'm worried that if we loose a Node it'll take more than 2
> OSDs with it, rendering us "stuffed" (stuffed: a technical term which is
> used as a substitute for a four-letter word rhyming with "truck") 😁
>
> Anyone have any pointers?
>
> Cheers
>
> Dulux-Oz
>
> _______
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>


-- 
Tyler Brekke
Senior Engineer I
tbre...@digitalocean.com
--
We're Hiring! <https://do.co/careers> | @digitalocean
<https://twitter.com/digitalocean> | YouTube
<https://www.youtube.com/digitalocean>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io