In our deployments it works like this.

It grabs what you get from "ceph orch ls" from top to bottom.
That's why we use the numbering (osd_1... osd_2... osd_3...).
Also db_slots work quite well.



Am Di., 30. Sept. 2025 um 15:57 Uhr schrieb Eugen Block <[email protected]>:

> That's not going to work as you expect: if the same host is applicable
> to multiple osd specs, only one of the specs will be applied (I think
> it's the last one in the list of 'ceph orch ls osd --export').
>
> I would also recommend to not use device paths, those can change after
> a reboot (although cephadm works with LVM and labels).
>
> The config option "db_slots" has never worked reliably, it's mentioned
> in the docs somewhere, so it most likely won't work either as you
> expect.
>
> If you can distinguish your OSDs using models, sizes or some other
> parameter, you could use several osd specs. If you can't, you might
> need to create the daemons manually like this:
>
> ceph01:~ # ceph orch daemon add osd
> ceph01:data_devices=/dev/vdb,/dev/vdc,/dev/vdd,db_devices=/dev/vde,block_db_size=64G
> (I'm not sure about that last parameter, writing this from the top of my
> head)
>
> But note that these OSDs will be displayed as "unmanaged" in 'ceph
> orch ls osd' output.
>
>
> Zitat von "GLE, Vivien" <[email protected]>:
>
> > Hi,
> >
> >
> > For testing purpose we need to deploy :
> >
> >
> > - 1 pool of 6 SSD OSD
> >
> > - 1 pool of 6 HDD OSD
> >
> > - 1 pool of 6 HDD OSD with 2 SSD for DB+WAL
> >
> >
> > I tried to orch apply this yaml but it doesn't work as expected
> >
> >
> > Osd part of the yaml ->
> >
> >
> > service_type: osd
> > service_id: osd_spec
> > placement:
> >   hosts:
> >     -  host1
> >     -  host2
> >     -  host3
> >     -  host4
> >     -  host5
> >     -  host6
> > data_devices:
> >   paths:
> >     - /dev/sda
> >     - /dev/sdc
> > spec:
> >   data_devices:
> >     all: true
> >   filter_logic: AND
> >   objectstore: bluestore
> > ---
> > service_type: osd
> > service_id: osd_spec_wall
> > placement:
> >   hosts:
> >     -  host1
> >     -  host2
> >     -  host3
> >     -  host4
> >     -  host5
> >     -  host6
> >
> > spec:
> >   data_devices:
> >     paths:
> >       - /dev/sdf
> >   db_devices:
> >     paths:
> >       - /dev/sde
> >     limit: 2
> >   db_slots: 3
> >
> >
> >
> > Only 1 db on /dev/sde from host1 has been created and this OSD
> > showed up as half full at his creation:
> >
> > ceph osd df | grep 25
> >
> > 25    hdd  0.63669   1.00000  652 GiB  373 GiB  1.5 MiB   1 KiB   38
> > MiB  279 GiB  57.15  35.93    0      up
> >
> > ceph-volume lvm list
> >
> > ====== osd.25 ======
> >
> >   [block]
> >
> /dev/ceph-23d9297a-d0e1-47be-8650-5c8ccae4fe0e/osd-block-2f009760-fc2b-46d5-984d-e8200dfd9d9d
> >
> >       block device
> >
> /dev/ceph-23d9297a-d0e1-47be-8650-5c8ccae4fe0e/osd-block-2f009760-fc2b-46d5-984d-e8200dfd9d9d
> >       block uuid                vqJGf8-V5g0-S1cA-BAcN-Qm9D-9VTx-xsc8wk
> >       cephx lockbox secret
> >       cluster fsid              id
> >       cluster name              ceph
> >       crush device class
> >       db device
> >
> /dev/ceph-9c39b87c-a39c-413f-b1ef-07881195fcb8/osd-db-feee5095-b5e7-47a0-ae87-8e5039512661
> >       db uuid                   Rtj5KA-Qxjk-IFmY-3ffQ-gSAu-Snte-uku8HC
> >       encrypted                 0
> >       osd fsid                  2f009760-fc2b-46d5-984d-e8200dfd9d9d
> >       osd id                    25
> >       osdspec affinity          osd_spec_wall
> >       type                      block
> >       vdo                       0
> >       with tpm                  0
> >       devices                   /dev/sdf
> >
> >   [db]
> >
> /dev/ceph-9c39b87c-a39c-413f-b1ef-07881195fcb8/osd-db-feee5095-b5e7-47a0-ae87-8e5039512661
> >
> >       block device
> >
> /dev/ceph-23d9297a-d0e1-47be-8650-5c8ccae4fe0e/osd-block-2f009760-fc2b-46d5-984d-e8200dfd9d9d
> >       block uuid                vqJGf8-V5g0-S1cA-BAcN-Qm9D-9VTx-xsc8wk
> >       cephx lockbox secret
> >       cluster fsid              id
> >       cluster name              ceph
> >       crush device class
> >       db device
> >
> /dev/ceph-9c39b87c-a39c-413f-b1ef-07881195fcb8/osd-db-feee5095-b5e7-47a0-ae87-8e5039512661
> >       db uuid                   Rtj5KA-Qxjk-IFmY-3ffQ-gSAu-Snte-uku8HC
> >       encrypted                 0
> >       osd fsid                  2f009760-fc2b-46d5-984d-e8200dfd9d9d
> >       osd id                    25
> >       osdspec affinity          osd_spec_wall
> >       type                      db
> >       vdo                       0
> >       with tpm                  0
> >       devices                   /dev/sde
> >
> >
> >
> > The other /dev/sde device did show up as data_device instead of
> > db_device (example here on host2) :
> >
> > ceph-volume lvm list
> >
> > ====== osd.17 ======
> >
> >   [block]
> >
> /dev/ceph-629f98b0-5ed4-4e75-81b9-e85ca76afb15/osd-block-5d43d683-1f7f-4dc1-935e-6a79745252f9
> >
> >       block device
> >
> /dev/ceph-629f98b0-5ed4-4e75-81b9-e85ca76afb15/osd-block-5d43d683-1f7f-4dc1-935e-6a79745252f9
> >       block uuid                HQpp1l-x7IB-kA2W-6gWO-BGlM-VN2k-vYf43R
> >       cephx lockbox secret
> >       cluster fsid              id
> >       cluster name              ceph
> >       crush device class
> >       encrypted                 0
> >       osd fsid                  5d43d683-1f7f-4dc1-935e-6a79745252f9
> >       osd id                    17
> >       osdspec affinity          osd_spec
> >       type                      block
> >       vdo                       0
> >       with tpm                  0
> >       devices                   /dev/sde
> >
> > Thx for your help
> > Vivien
> >
> >
> >
> >
> >
> >
> >
> > _______________________________________________
> > ceph-users mailing list -- [email protected]
> > To unsubscribe send an email to [email protected]
>
>
> _______________________________________________
> ceph-users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
>
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to