That's not going to work as you expect: if the same host is applicable to multiple osd specs, only one of the specs will be applied (I think it's the last one in the list of 'ceph orch ls osd --export').

I would also recommend to not use device paths, those can change after a reboot (although cephadm works with LVM and labels).

The config option "db_slots" has never worked reliably, it's mentioned in the docs somewhere, so it most likely won't work either as you expect.

If you can distinguish your OSDs using models, sizes or some other parameter, you could use several osd specs. If you can't, you might need to create the daemons manually like this:

ceph01:~ # ceph orch daemon add osd ceph01:data_devices=/dev/vdb,/dev/vdc,/dev/vdd,db_devices=/dev/vde,block_db_size=64G (I'm not sure about that last parameter, writing this from the top of my head)

But note that these OSDs will be displayed as "unmanaged" in 'ceph orch ls osd' output.


Zitat von "GLE, Vivien" <[email protected]>:

Hi,


For testing purpose we need to deploy :


- 1 pool of 6 SSD OSD

- 1 pool of 6 HDD OSD

- 1 pool of 6 HDD OSD with 2 SSD for DB+WAL


I tried to orch apply this yaml but it doesn't work as expected


Osd part of the yaml ->


service_type: osd
service_id: osd_spec
placement:
  hosts:
    -  host1
    -  host2
    -  host3
    -  host4
    -  host5
    -  host6
data_devices:
  paths:
    - /dev/sda
    - /dev/sdc
spec:
  data_devices:
    all: true
  filter_logic: AND
  objectstore: bluestore
---
service_type: osd
service_id: osd_spec_wall
placement:
  hosts:
    -  host1
    -  host2
    -  host3
    -  host4
    -  host5
    -  host6

spec:
  data_devices:
    paths:
      - /dev/sdf
  db_devices:
    paths:
      - /dev/sde
    limit: 2
  db_slots: 3



Only 1 db on /dev/sde from host1 has been created and this OSD showed up as half full at his creation:

ceph osd df | grep 25

25 hdd 0.63669 1.00000 652 GiB 373 GiB 1.5 MiB 1 KiB 38 MiB 279 GiB 57.15 35.93 0 up

ceph-volume lvm list

====== osd.25 ======

[block] /dev/ceph-23d9297a-d0e1-47be-8650-5c8ccae4fe0e/osd-block-2f009760-fc2b-46d5-984d-e8200dfd9d9d

block device /dev/ceph-23d9297a-d0e1-47be-8650-5c8ccae4fe0e/osd-block-2f009760-fc2b-46d5-984d-e8200dfd9d9d
      block uuid                vqJGf8-V5g0-S1cA-BAcN-Qm9D-9VTx-xsc8wk
      cephx lockbox secret
      cluster fsid              id
      cluster name              ceph
      crush device class
db device /dev/ceph-9c39b87c-a39c-413f-b1ef-07881195fcb8/osd-db-feee5095-b5e7-47a0-ae87-8e5039512661
      db uuid                   Rtj5KA-Qxjk-IFmY-3ffQ-gSAu-Snte-uku8HC
      encrypted                 0
      osd fsid                  2f009760-fc2b-46d5-984d-e8200dfd9d9d
      osd id                    25
      osdspec affinity          osd_spec_wall
      type                      block
      vdo                       0
      with tpm                  0
      devices                   /dev/sdf

[db] /dev/ceph-9c39b87c-a39c-413f-b1ef-07881195fcb8/osd-db-feee5095-b5e7-47a0-ae87-8e5039512661

block device /dev/ceph-23d9297a-d0e1-47be-8650-5c8ccae4fe0e/osd-block-2f009760-fc2b-46d5-984d-e8200dfd9d9d
      block uuid                vqJGf8-V5g0-S1cA-BAcN-Qm9D-9VTx-xsc8wk
      cephx lockbox secret
      cluster fsid              id
      cluster name              ceph
      crush device class
db device /dev/ceph-9c39b87c-a39c-413f-b1ef-07881195fcb8/osd-db-feee5095-b5e7-47a0-ae87-8e5039512661
      db uuid                   Rtj5KA-Qxjk-IFmY-3ffQ-gSAu-Snte-uku8HC
      encrypted                 0
      osd fsid                  2f009760-fc2b-46d5-984d-e8200dfd9d9d
      osd id                    25
      osdspec affinity          osd_spec_wall
      type                      db
      vdo                       0
      with tpm                  0
      devices                   /dev/sde



The other /dev/sde device did show up as data_device instead of db_device (example here on host2) :

ceph-volume lvm list

====== osd.17 ======

[block] /dev/ceph-629f98b0-5ed4-4e75-81b9-e85ca76afb15/osd-block-5d43d683-1f7f-4dc1-935e-6a79745252f9

block device /dev/ceph-629f98b0-5ed4-4e75-81b9-e85ca76afb15/osd-block-5d43d683-1f7f-4dc1-935e-6a79745252f9
      block uuid                HQpp1l-x7IB-kA2W-6gWO-BGlM-VN2k-vYf43R
      cephx lockbox secret
      cluster fsid              id
      cluster name              ceph
      crush device class
      encrypted                 0
      osd fsid                  5d43d683-1f7f-4dc1-935e-6a79745252f9
      osd id                    17
      osdspec affinity          osd_spec
      type                      block
      vdo                       0
      with tpm                  0
      devices                   /dev/sde

Thx for your help
Vivien







_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]


_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to