hi, I've forget to write the command, I've used:
===== ceph osd crush move fc-r02-ceph-osd-01 root=default ceph osd crush move fc-r02-ceph-osd-01 root=default ... ===== and I've found also this param: =========== root@fc-r02-ceph-osd-01:[~]: ceph osd crush tree --show-shadow ID CLASS WEIGHT TYPE NAME -39 nvme 1.81938 root default~nvme -30 nvme 0 host fc-r02-ceph-osd-01~nvme -31 nvme 0.36388 host fc-r02-ceph-osd-02~nvme 36 nvme 0.36388 osd.36 -32 nvme 0.36388 host fc-r02-ceph-osd-03~nvme 40 nvme 0.36388 osd.40 -33 nvme 0.36388 host fc-r02-ceph-osd-04~nvme 37 nvme 0.36388 osd.37 -34 nvme 0.36388 host fc-r02-ceph-osd-05~nvme 38 nvme 0.36388 osd.38 -35 nvme 0.36388 host fc-r02-ceph-osd-06~nvme 39 nvme 0.36388 osd.39 -38 nvme 0 root ssds~nvme -37 nvme 0 datacenter fc-ssds~nvme -36 nvme 0 rack r02-ssds~nvme -29 nvme 0 root sata~nvme -28 nvme 0 datacenter fc-sata~nvme -27 nvme 0 rack r02-sata~nvme -24 ssd 0 root ssds~ssd -23 ssd 0 datacenter fc-ssds~ssd -21 ssd 0 rack r02-ssds~ssd -22 ssd 0 root sata~ssd -19 ssd 0 datacenter fc-sata~ssd -20 ssd 0 rack r02-sata~ssd -14 0 root sata -18 0 datacenter fc-sata -16 0 rack r02-sata -13 0 root ssds -17 0 datacenter fc-ssds -15 0 rack r02-ssds -4 ssd 22.17122 root default~ssd -7 ssd 4.00145 host fc-r02-ceph-osd-01~ssd 0 ssd 0.45470 osd.0 1 ssd 0.45470 osd.1 2 ssd 0.45470 osd.2 3 ssd 0.45470 osd.3 4 ssd 0.45470 osd.4 5 ssd 0.45470 osd.5 41 ssd 0.36388 osd.41 42 ssd 0.45470 osd.42 48 ssd 0.45470 osd.48 -3 ssd 3.61948 host fc-r02-ceph-osd-02~ssd 6 ssd 0.45470 osd.6 7 ssd 0.45470 osd.7 8 ssd 0.45470 osd.8 9 ssd 0.45470 osd.9 10 ssd 0.43660 osd.10 29 ssd 0.45470 osd.29 43 ssd 0.45470 osd.43 49 ssd 0.45470 osd.49 -8 ssd 3.63757 host fc-r02-ceph-osd-03~ssd 11 ssd 0.45470 osd.11 12 ssd 0.45470 osd.12 13 ssd 0.45470 osd.13 14 ssd 0.45470 osd.14 15 ssd 0.45470 osd.15 16 ssd 0.45470 osd.16 44 ssd 0.45470 osd.44 50 ssd 0.45470 osd.50 -10 ssd 3.63757 host fc-r02-ceph-osd-04~ssd 30 ssd 0.45470 osd.30 31 ssd 0.45470 osd.31 32 ssd 0.45470 osd.32 33 ssd 0.45470 osd.33 34 ssd 0.45470 osd.34 35 ssd 0.45470 osd.35 45 ssd 0.45470 osd.45 51 ssd 0.45470 osd.51 -12 ssd 3.63757 host fc-r02-ceph-osd-05~ssd 17 ssd 0.45470 osd.17 18 ssd 0.45470 osd.18 19 ssd 0.45470 osd.19 20 ssd 0.45470 osd.20 21 ssd 0.45470 osd.21 22 ssd 0.45470 osd.22 46 ssd 0.45470 osd.46 52 ssd 0.45470 osd.52 -26 ssd 3.63757 host fc-r02-ceph-osd-06~ssd 23 ssd 0.45470 osd.23 24 ssd 0.45470 osd.24 25 ssd 0.45470 osd.25 26 ssd 0.45470 osd.26 27 ssd 0.45470 osd.27 28 ssd 0.45470 osd.28 47 ssd 0.45470 osd.47 53 ssd 0.45470 osd.53 -1 23.99060 root default -6 4.00145 host fc-r02-ceph-osd-01 0 ssd 0.45470 osd.0 1 ssd 0.45470 osd.1 2 ssd 0.45470 osd.2 3 ssd 0.45470 osd.3 4 ssd 0.45470 osd.4 5 ssd 0.45470 osd.5 41 ssd 0.36388 osd.41 42 ssd 0.45470 osd.42 48 ssd 0.45470 osd.48 -2 3.98335 host fc-r02-ceph-osd-02 36 nvme 0.36388 osd.36 6 ssd 0.45470 osd.6 7 ssd 0.45470 osd.7 8 ssd 0.45470 osd.8 9 ssd 0.45470 osd.9 10 ssd 0.43660 osd.10 29 ssd 0.45470 osd.29 43 ssd 0.45470 osd.43 49 ssd 0.45470 osd.49 -5 4.00145 host fc-r02-ceph-osd-03 40 nvme 0.36388 osd.40 11 ssd 0.45470 osd.11 12 ssd 0.45470 osd.12 13 ssd 0.45470 osd.13 14 ssd 0.45470 osd.14 15 ssd 0.45470 osd.15 16 ssd 0.45470 osd.16 44 ssd 0.45470 osd.44 50 ssd 0.45470 osd.50 -9 4.00145 host fc-r02-ceph-osd-04 37 nvme 0.36388 osd.37 30 ssd 0.45470 osd.30 31 ssd 0.45470 osd.31 32 ssd 0.45470 osd.32 33 ssd 0.45470 osd.33 34 ssd 0.45470 osd.34 35 ssd 0.45470 osd.35 45 ssd 0.45470 osd.45 51 ssd 0.45470 osd.51 -11 4.00145 host fc-r02-ceph-osd-05 38 nvme 0.36388 osd.38 17 ssd 0.45470 osd.17 18 ssd 0.45470 osd.18 19 ssd 0.45470 osd.19 20 ssd 0.45470 osd.20 21 ssd 0.45470 osd.21 22 ssd 0.45470 osd.22 46 ssd 0.45470 osd.46 52 ssd 0.45470 osd.52 -25 4.00145 host fc-r02-ceph-osd-06 39 nvme 0.36388 osd.39 23 ssd 0.45470 osd.23 24 ssd 0.45470 osd.24 25 ssd 0.45470 osd.25 26 ssd 0.45470 osd.26 27 ssd 0.45470 osd.27 28 ssd 0.45470 osd.28 47 ssd 0.45470 osd.47 53 ssd 0.45470 osd.53 ===================== cu denny _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io