Here's a crush ruleset for 8+2 that will choose 2 osds per host:

rule cephfs_data_82 {
        id 4
        type erasure
        min_size 3
        max_size 10
        step set_chooseleaf_tries 5
        step set_choose_tries 100
        step take default class hdd
        step choose indep 5 type host
        step choose indep 2 type osd
        step emit
}



This is kind of useful because if you set min_size to 8, you could even
lose an entire host and stay online.

Cheers, dan





On Thu, Mar 25, 2021, 7:02 PM by morphin <morphinwith...@gmail.com> wrote:

> Hello.
>
> I have 5 node Cluster in A datacenter. Also I have same 5 node in B
> datacenter.
> They're gonna be 10 node 8+2 EC cluster for backup but I need to add
> the 5 node later.
> I have to sync my S3 data with multisite on the 5 node cluster in A
> datacenter and move
> them to the B and add the other 5 node to the same cluster.
>
> The question is: Can I create 8+2 ec pool on 5 node cluster and add
> the 5 node later? How can I rebalance the data after that?
> Or is there any better solution in my case? what should I do?
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to