In our deployments it works like this. It grabs what you get from "ceph orch ls" from top to bottom. That's why we use the numbering (osd_1... osd_2... osd_3...). Also db_slots work quite well.
Am Di., 30. Sept. 2025 um 15:57 Uhr schrieb Eugen Block <[email protected]>: > That's not going to work as you expect: if the same host is applicable > to multiple osd specs, only one of the specs will be applied (I think > it's the last one in the list of 'ceph orch ls osd --export'). > > I would also recommend to not use device paths, those can change after > a reboot (although cephadm works with LVM and labels). > > The config option "db_slots" has never worked reliably, it's mentioned > in the docs somewhere, so it most likely won't work either as you > expect. > > If you can distinguish your OSDs using models, sizes or some other > parameter, you could use several osd specs. If you can't, you might > need to create the daemons manually like this: > > ceph01:~ # ceph orch daemon add osd > ceph01:data_devices=/dev/vdb,/dev/vdc,/dev/vdd,db_devices=/dev/vde,block_db_size=64G > (I'm not sure about that last parameter, writing this from the top of my > head) > > But note that these OSDs will be displayed as "unmanaged" in 'ceph > orch ls osd' output. > > > Zitat von "GLE, Vivien" <[email protected]>: > > > Hi, > > > > > > For testing purpose we need to deploy : > > > > > > - 1 pool of 6 SSD OSD > > > > - 1 pool of 6 HDD OSD > > > > - 1 pool of 6 HDD OSD with 2 SSD for DB+WAL > > > > > > I tried to orch apply this yaml but it doesn't work as expected > > > > > > Osd part of the yaml -> > > > > > > service_type: osd > > service_id: osd_spec > > placement: > > hosts: > > - host1 > > - host2 > > - host3 > > - host4 > > - host5 > > - host6 > > data_devices: > > paths: > > - /dev/sda > > - /dev/sdc > > spec: > > data_devices: > > all: true > > filter_logic: AND > > objectstore: bluestore > > --- > > service_type: osd > > service_id: osd_spec_wall > > placement: > > hosts: > > - host1 > > - host2 > > - host3 > > - host4 > > - host5 > > - host6 > > > > spec: > > data_devices: > > paths: > > - /dev/sdf > > db_devices: > > paths: > > - /dev/sde > > limit: 2 > > db_slots: 3 > > > > > > > > Only 1 db on /dev/sde from host1 has been created and this OSD > > showed up as half full at his creation: > > > > ceph osd df | grep 25 > > > > 25 hdd 0.63669 1.00000 652 GiB 373 GiB 1.5 MiB 1 KiB 38 > > MiB 279 GiB 57.15 35.93 0 up > > > > ceph-volume lvm list > > > > ====== osd.25 ====== > > > > [block] > > > /dev/ceph-23d9297a-d0e1-47be-8650-5c8ccae4fe0e/osd-block-2f009760-fc2b-46d5-984d-e8200dfd9d9d > > > > block device > > > /dev/ceph-23d9297a-d0e1-47be-8650-5c8ccae4fe0e/osd-block-2f009760-fc2b-46d5-984d-e8200dfd9d9d > > block uuid vqJGf8-V5g0-S1cA-BAcN-Qm9D-9VTx-xsc8wk > > cephx lockbox secret > > cluster fsid id > > cluster name ceph > > crush device class > > db device > > > /dev/ceph-9c39b87c-a39c-413f-b1ef-07881195fcb8/osd-db-feee5095-b5e7-47a0-ae87-8e5039512661 > > db uuid Rtj5KA-Qxjk-IFmY-3ffQ-gSAu-Snte-uku8HC > > encrypted 0 > > osd fsid 2f009760-fc2b-46d5-984d-e8200dfd9d9d > > osd id 25 > > osdspec affinity osd_spec_wall > > type block > > vdo 0 > > with tpm 0 > > devices /dev/sdf > > > > [db] > > > /dev/ceph-9c39b87c-a39c-413f-b1ef-07881195fcb8/osd-db-feee5095-b5e7-47a0-ae87-8e5039512661 > > > > block device > > > /dev/ceph-23d9297a-d0e1-47be-8650-5c8ccae4fe0e/osd-block-2f009760-fc2b-46d5-984d-e8200dfd9d9d > > block uuid vqJGf8-V5g0-S1cA-BAcN-Qm9D-9VTx-xsc8wk > > cephx lockbox secret > > cluster fsid id > > cluster name ceph > > crush device class > > db device > > > /dev/ceph-9c39b87c-a39c-413f-b1ef-07881195fcb8/osd-db-feee5095-b5e7-47a0-ae87-8e5039512661 > > db uuid Rtj5KA-Qxjk-IFmY-3ffQ-gSAu-Snte-uku8HC > > encrypted 0 > > osd fsid 2f009760-fc2b-46d5-984d-e8200dfd9d9d > > osd id 25 > > osdspec affinity osd_spec_wall > > type db > > vdo 0 > > with tpm 0 > > devices /dev/sde > > > > > > > > The other /dev/sde device did show up as data_device instead of > > db_device (example here on host2) : > > > > ceph-volume lvm list > > > > ====== osd.17 ====== > > > > [block] > > > /dev/ceph-629f98b0-5ed4-4e75-81b9-e85ca76afb15/osd-block-5d43d683-1f7f-4dc1-935e-6a79745252f9 > > > > block device > > > /dev/ceph-629f98b0-5ed4-4e75-81b9-e85ca76afb15/osd-block-5d43d683-1f7f-4dc1-935e-6a79745252f9 > > block uuid HQpp1l-x7IB-kA2W-6gWO-BGlM-VN2k-vYf43R > > cephx lockbox secret > > cluster fsid id > > cluster name ceph > > crush device class > > encrypted 0 > > osd fsid 5d43d683-1f7f-4dc1-935e-6a79745252f9 > > osd id 17 > > osdspec affinity osd_spec > > type block > > vdo 0 > > with tpm 0 > > devices /dev/sde > > > > Thx for your help > > Vivien > > > > > > > > > > > > > > > > _______________________________________________ > > ceph-users mailing list -- [email protected] > > To unsubscribe send an email to [email protected] > > > _______________________________________________ > ceph-users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > _______________________________________________ ceph-users mailing list -- [email protected] To unsubscribe send an email to [email protected]
