Here is the log from ceph-volume.
```
[2021-02-05 04:03:17,000][ceph_volume.process][INFO  ] Running command: 
/usr/sbin/vgcreate --force --yes ceph-a3886f74-3de9-4e6e-a983-8330eda0bd64 
/dev/sdd
[2021-02-05 04:03:17,134][ceph_volume.process][INFO  ] stdout Physical volume 
"/dev/sdd" successfully created.
[2021-02-05 04:03:17,166][ceph_volume.process][INFO  ] stdout Volume group 
"ceph-a3886f74-3de9-4e6e-a983-8330eda0bd64" successfully created
[2021-02-05 04:03:17,189][ceph_volume.process][INFO  ] Running command: 
/usr/sbin/vgs --noheadings --readonly --units=b --nosuffix --separator=";" -S 
vg_name=ceph-a3886f74-3de9-4e6e-a983-8330eda0bd64 -o 
vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_size
[2021-02-05 04:03:17,229][ceph_volume.process][INFO  ] stdout 
ceph-a3886f74-3de9-4e6e-a983-8330eda0bd64";"1";"0";"wz--n-";"572317";"572317";"4194304
[2021-02-05 04:03:17,229][ceph_volume.api.lvm][DEBUG ] size was passed: 2.18 TB 
-> 572318
[2021-02-05 04:03:17,235][ceph_volume.process][INFO  ] Running command: 
/usr/sbin/lvcreate --yes -l 572318 -n 
osd-block-b05c3c90-b7d5-4f13-8a58-f72761c1971b 
ceph-a3886f74-3de9-4e6e-a983-8330eda0bd64
[2021-02-05 04:03:17,244][ceph_volume.process][INFO  ] stderr Volume group 
"ceph-a3886f74-3de9-4e6e-a983-8330eda0bd64" has insufficient free space (572317 
extents): 572318 required.
```
size was passed: 2.18 TB -> 572318
How is  this calculated?


Thanks!
Tony
> -----Original Message-----
> From: Tony Liu <tonyliu0...@hotmail.com>
> Sent: Thursday, February 4, 2021 8:34 PM
> To: ceph-users@ceph.io
> Subject: [ceph-users] replace OSD failed
> 
> Hi,
> 
> With 15.2.8, run "ceph orch rm osd 12 --replace --force", PGs on osd.12
> are remapped, osd.12 is removed from "ceph osd tree", the daemon is
> removed from "ceph orch ps", the device is "available"
> in "ceph orch device ls". Everything seems good at this point.
> 
> Then dry-run service spec.
> ```
> # cat osd-spec.yaml
> service_type: osd
> service_id: osd-spec
> placement:
>   hosts:
>   - ceph-osd-1
> data_devices:
>   rotational: 1
> db_devices:
>   rotational: 0
> 
> # ceph orch apply osd -i osd-spec.yaml --dry-run
> +---------+----------+------------+----------+----------+-----+
> |SERVICE  |NAME      |HOST        |DATA      |DB        |WAL  |
> +---------+----------+------------+----------+----------+-----+
> |osd      |osd-spec  |ceph-osd-3  |/dev/sdd  |/dev/sdb  |-    |
> +---------+----------+------------+----------+----------+-----+
> ```
> It looks as expected.
> 
> Then "ceph orch apply osd -i osd-spec.yaml".
> Here is the log of cephadm.
> ```
> /bin/docker:stderr --> relative data size: 1.0 /bin/docker:stderr -->
> passed block_db devices: 1 physical, 0 LVM /bin/docker:stderr Running
> command: /usr/bin/ceph-authtool --gen-print-key /bin/docker:stderr
> Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-
> osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd tree -f json
> /bin/docker:stderr Running command: /usr/bin/ceph --cluster ceph --name
> client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring
> -i - osd new b05c3c90-b7d5-4f13-8a58-f72761c1971b 12 /bin/docker:stderr
> Running command: /usr/sbin/vgcreate --force --yes ceph-a3886f74-3de9-
> 4e6e-a983-8330eda0bd64 /dev/sdd /bin/docker:stderr  stdout: Physical
> volume "/dev/sdd" successfully created.
> /bin/docker:stderr  stdout: Volume group "ceph-a3886f74-3de9-4e6e-a983-
> 8330eda0bd64" successfully created /bin/docker:stderr Running command:
> /usr/sbin/lvcreate --yes -l 572318 -n osd-block-b05c3c90-b7d5-4f13-8a58-
> f72761c1971b ceph-a3886f74-3de9-4e6e-a983-8330eda0bd64
> /bin/docker:stderr  stderr: Volume group "ceph-a3886f74-3de9-4e6e-a983-
> 8330eda0bd64" has insufficient free space (572317 extents): 572318
> required.
> /bin/docker:stderr --> Was unable to complete a new OSD, will rollback
> changes ``` Q1, why VG name (ceph-<id>) is different from others (ceph-
> block-<id>)?
> Q2, where is that 572318 from? Since all HDDs are the same model, VG
> "Total PE" of all HDDs is 572317.
> Has anyone seen similar issues? Anything I am missing?
> 
> 
> Thanks!
> Tony
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an
> email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to