Hi Vivien,
You need to create first the OSDs, then corresponding crush rules and after
that pools with that crush rule.
This might work if each host contains of 12 HDDs and 8 SSDs:
# OSD spec yaml
---
service_type: osd
service_id: osd_1_ssd
placement:
host_pattern: '*'
spec:
data_devices:
rotational: 0
limit: 6
encrypted: true
---
service_type: osd
service_id: osd_2_hdd_blockdb_on_ssd
placement:
host_pattern: '*'
spec:
data_devices:
rotational: 1
limit: 6
encrypted: true
block_db_size: 549755813888
db_slots: 3
db_devices:
rotational: 0
---
service_type: osd
service_id: osd_3_hddonly
placement:
host_pattern: '*'
spec:
data_devices:
rotational: 1
limit: 6
crush_device_class: hddonly
# Crush rules
ceph osd crush rule create-replicated replicated_ssd default host ssd
ceph osd crush rule create-replicated replicated_hdd default host hdd
ceph osd crush rule create-replicated replicated_hddonly default host
hddonly
# Pools
~# ceph osd pool create {pool-name} [{pg-num} [{pgp-num}]] [replicated] \
[crush-rule-name]
ceph osd pool create ssd-pool 32 32 replicated_ssd
ceph osd pool create hdd-pool 32 32 replicated_hdd
ceph osd pool create hddonly-pool 32 32 replicated_hddonly
Best Stephan
Am Di., 30. Sept. 2025 um 14:06 Uhr schrieb GLE, Vivien <[email protected]
>:
> Hi,
>
>
> For testing purpose we need to deploy :
>
>
> - 1 pool of 6 SSD OSD
>
> - 1 pool of 6 HDD OSD
>
> - 1 pool of 6 HDD OSD with 2 SSD for DB+WAL
>
>
> I tried to orch apply this yaml but it doesn't work as expected
>
>
> Osd part of the yaml ->
>
>
> service_type: osd
> service_id: osd_spec
> placement:
> hosts:
> - host1
> - host2
> - host3
> - host4
> - host5
> - host6
> data_devices:
> paths:
> - /dev/sda
> - /dev/sdc
> spec:
> data_devices:
> all: true
> filter_logic: AND
> objectstore: bluestore
> ---
> service_type: osd
> service_id: osd_spec_wall
> placement:
> hosts:
> - host1
> - host2
> - host3
> - host4
> - host5
> - host6
>
> spec:
> data_devices:
> paths:
> - /dev/sdf
> db_devices:
> paths:
> - /dev/sde
> limit: 2
> db_slots: 3
>
>
>
> Only 1 db on /dev/sde from host1 has been created and this OSD showed up
> as half full at his creation:
>
> ceph osd df | grep 25
>
> 25 hdd 0.63669 1.00000 652 GiB 373 GiB 1.5 MiB 1 KiB 38 MiB
> 279 GiB 57.15 35.93 0 up
>
> ceph-volume lvm list
>
> ====== osd.25 ======
>
> [block]
>
> /dev/ceph-23d9297a-d0e1-47be-8650-5c8ccae4fe0e/osd-block-2f009760-fc2b-46d5-984d-e8200dfd9d9d
>
> block device
> /dev/ceph-23d9297a-d0e1-47be-8650-5c8ccae4fe0e/osd-block-2f009760-fc2b-46d5-984d-e8200dfd9d9d
> block uuid vqJGf8-V5g0-S1cA-BAcN-Qm9D-9VTx-xsc8wk
> cephx lockbox secret
> cluster fsid id
> cluster name ceph
> crush device class
> db device
>
> /dev/ceph-9c39b87c-a39c-413f-b1ef-07881195fcb8/osd-db-feee5095-b5e7-47a0-ae87-8e5039512661
> db uuid Rtj5KA-Qxjk-IFmY-3ffQ-gSAu-Snte-uku8HC
> encrypted 0
> osd fsid 2f009760-fc2b-46d5-984d-e8200dfd9d9d
> osd id 25
> osdspec affinity osd_spec_wall
> type block
> vdo 0
> with tpm 0
> devices /dev/sdf
>
> [db]
> /dev/ceph-9c39b87c-a39c-413f-b1ef-07881195fcb8/osd-db-feee5095-b5e7-47a0-ae87-8e5039512661
>
> block device
> /dev/ceph-23d9297a-d0e1-47be-8650-5c8ccae4fe0e/osd-block-2f009760-fc2b-46d5-984d-e8200dfd9d9d
> block uuid vqJGf8-V5g0-S1cA-BAcN-Qm9D-9VTx-xsc8wk
> cephx lockbox secret
> cluster fsid id
> cluster name ceph
> crush device class
> db device
>
> /dev/ceph-9c39b87c-a39c-413f-b1ef-07881195fcb8/osd-db-feee5095-b5e7-47a0-ae87-8e5039512661
> db uuid Rtj5KA-Qxjk-IFmY-3ffQ-gSAu-Snte-uku8HC
> encrypted 0
> osd fsid 2f009760-fc2b-46d5-984d-e8200dfd9d9d
> osd id 25
> osdspec affinity osd_spec_wall
> type db
> vdo 0
> with tpm 0
> devices /dev/sde
>
>
>
> The other /dev/sde device did show up as data_device instead of db_device
> (example here on host2) :
>
> ceph-volume lvm list
>
> ====== osd.17 ======
>
> [block]
>
> /dev/ceph-629f98b0-5ed4-4e75-81b9-e85ca76afb15/osd-block-5d43d683-1f7f-4dc1-935e-6a79745252f9
>
> block device
> /dev/ceph-629f98b0-5ed4-4e75-81b9-e85ca76afb15/osd-block-5d43d683-1f7f-4dc1-935e-6a79745252f9
> block uuid HQpp1l-x7IB-kA2W-6gWO-BGlM-VN2k-vYf43R
> cephx lockbox secret
> cluster fsid id
> cluster name ceph
> crush device class
> encrypted 0
> osd fsid 5d43d683-1f7f-4dc1-935e-6a79745252f9
> osd id 17
> osdspec affinity osd_spec
> type block
> vdo 0
> with tpm 0
> devices /dev/sde
>
> Thx for your help
> Vivien
>
>
>
>
>
>
>
> _______________________________________________
> ceph-users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
>
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]