On Wed, Apr 19, 2023 at 5:57 PM Reto Gysi <rlg...@gmail.com> wrote:
>
>
> Hi,
>
> Am Mi., 19. Apr. 2023 um 11:02 Uhr schrieb Ilya Dryomov <idryo...@gmail.com>:
>>
>> On Wed, Apr 19, 2023 at 10:29 AM Reto Gysi <rlg...@gmail.com> wrote:
>> >
>> > yes, I used the same ecpool_hdd also for cephfs file systems. The new pool 
>> > ecpool_test I've created for a test, I've also created it with application 
>> > profile 'cephfs', but there aren't any cephfs filesystem attached to it.
>>
>> This is not and has never been supported.
>
>
> Do you mean 1) using the same erasure coded pool for both rbd and cephfs, or  
> 2) multiple cephfs filesystem using the same erasure coded pool via 
> ceph.dir.layout.pool="ecpool_hdd"?

(1), using the same EC pool for both RBD and CephFS.

>
> 1)
>
>
> 2)
> rgysi cephfs filesystem
> rgysi - 5 clients
> =====
> RANK  STATE           MDS             ACTIVITY     DNS    INOS   DIRS   CAPS
> 0    active  rgysi.debian.uhgqen  Reqs:    0 /s   409k   408k  40.8k  16.5k
>       POOL          TYPE     USED  AVAIL
> cephfs.rgysi.meta  metadata  1454M  2114G
> cephfs.rgysi.data    data    4898G  17.6T
>    ecpool_hdd       data    29.3T  29.6T
>
> root@zephir:~# getfattr -n ceph.dir.layout /home/rgysi/am/ecpool/
> getfattr: Removing leading '/' from absolute path names
> # file: home/rgysi/am/ecpool/
> ceph.dir.layout="stripe_unit=4194304 stripe_count=1 object_size=4194304 
> pool=ecpool_hdd"
>
> root@zephir:~#
>
> backups cephfs filesystem
> backups - 2 clients
> =======
> RANK  STATE            MDS              ACTIVITY     DNS    INOS   DIRS   CAPS
> 0    active  backups.debian.runngh  Reqs:    0 /s   253k   253k  21.3k   899
>        POOL           TYPE     USED  AVAIL
> cephfs.backups.meta  metadata  1364M  2114G
> cephfs.backups.data    data    16.7T  16.4T
>     ecpool_hdd        data    29.3T  29.6T
>
> root@zephir:~# getfattr -n ceph.dir.layout 
> /mnt/backups/windows/windows-drives/
> getfattr: Removing leading '/' from absolute path names
> # file: mnt/backups/windows/windows-drives/
> ceph.dir.layout="stripe_unit=4194304 stripe_count=1 object_size=4194304 
> pool=ecpool_hdd"
>
> root@zephir:~#
>
>
>
> So I guess I should use a different ec datapool for rbd and for each of the 
> cephfs filesystems in the future, correct?

Definitely a different EC pool for RBD (i.e. don't mix with CephFS).
Not sure about the _each_ of the filesystems bit -- Venky or Patrick can
comment on whether sharing an EC pool between filesystems is OK.

Thanks,

                Ilya
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to