BTW: The osdmaptool doesn't see anything to do either: $ ceph osd getmap -o om $ osdmaptool om --upmap /tmp/upmap.sh --upmap-pool cephfs_data osdmaptool: osdmap file 'om' writing upmap command output to: /tmp/upmap.sh checking for upmap cleanups upmap, max-count 100, max deviation 0.01 limiting to pools cephfs_data (1) no upmaps proposed
Tue, 3 Dec 2019 07:30:24 +0100 Lars Täuber <taeu...@bbaw.de> ==> Konstantin Shalygin <k0...@k0ste.ru> : > Hi Konstantin, > > > Tue, 3 Dec 2019 10:01:34 +0700 > Konstantin Shalygin <k0...@k0ste.ru> ==> Lars Täuber <taeu...@bbaw.de>, > ceph-users@ceph.io : > > Please paste your `ceph osd df tree`, `ceph osd pool ls detail`, `ceph > > osd crush rule dump`. > > here it comes: > > $ ceph osd df tree > ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL > %USE VAR PGS STATUS TYPE NAME > -1 195.40730 - 195 TiB 130 TiB 128 TiB 58 GiB 476 GiB 66 TiB > 66.45 1.00 - root default > -25 195.40730 - 195 TiB 130 TiB 128 TiB 58 GiB 476 GiB 66 TiB > 66.45 1.00 - room PRZ > -26 195.40730 - 195 TiB 130 TiB 128 TiB 58 GiB 476 GiB 66 TiB > 66.45 1.00 - row rechts > -27 83.74599 - 84 TiB 57 TiB 56 TiB 25 GiB 206 GiB 27 TiB > 67.51 1.02 - rack 1-eins > -3 27.91533 - 28 TiB 18 TiB 17 TiB 8.4 GiB 66 GiB 10 TiB > 62.80 0.95 - host onode1 > 0 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.8 MiB 14 GiB 2.1 TiB > 62.48 0.94 163 up osd.0 > 1 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 6.5 MiB 12 GiB 2.1 TiB > 62.47 0.94 163 up osd.1 > 2 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 7.1 MiB 12 GiB 2.1 TiB > 62.53 0.94 163 up osd.2 > 3 hdd 5.51459 1.00000 5.5 TiB 3.5 TiB 3.4 TiB 7.5 MiB 12 GiB 2.0 TiB > 62.90 0.95 164 up osd.3 > 37 hdd 5.51459 1.00000 5.5 TiB 3.7 TiB 3.7 TiB 6.4 MiB 13 GiB 1.8 TiB > 67.32 1.01 176 up osd.37 > 4 ssd 0.34239 1.00000 351 GiB 11 GiB 187 MiB 8.3 GiB 2.0 GiB 340 GiB > 3.01 0.05 110 up osd.4 > -13 27.91533 - 28 TiB 17 TiB 17 TiB 8.2 GiB 66 GiB 10 TiB > 62.64 0.94 - host onode4 > 13 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.0 MiB 13 GiB 2.1 TiB > 62.49 0.94 163 up osd.13 > 14 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.2 MiB 13 GiB 2.1 TiB > 62.49 0.94 163 up osd.14 > 15 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.4 MiB 12 GiB 2.1 TiB > 62.43 0.94 163 up osd.15 > 16 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 5.8 MiB 12 GiB 2.1 TiB > 62.13 0.94 162 up osd.16 > 40 hdd 5.51459 1.00000 5.5 TiB 3.7 TiB 3.7 TiB 3.6 MiB 13 GiB 1.8 TiB > 67.36 1.01 176 up osd.40 > 33 ssd 0.34239 1.00000 351 GiB 11 GiB 201 MiB 8.2 GiB 2.2 GiB 340 GiB > 3.02 0.05 110 up osd.33 > -22 27.91533 - 28 TiB 22 TiB 21 TiB 8.1 GiB 74 GiB 6.4 TiB > 77.10 1.16 - host onode7 > 25 hdd 5.51459 1.00000 5.5 TiB 4.3 TiB 4.2 TiB 7.2 MiB 14 GiB 1.2 TiB > 77.59 1.17 203 up osd.25 > 26 hdd 5.51459 1.00000 5.5 TiB 4.3 TiB 4.3 TiB 4.7 MiB 14 GiB 1.2 TiB > 78.40 1.18 205 up osd.26 > 27 hdd 5.51459 1.00000 5.5 TiB 4.2 TiB 4.1 TiB 3.8 MiB 14 GiB 1.3 TiB > 75.80 1.14 198 up osd.27 > 28 hdd 5.51459 1.00000 5.5 TiB 4.2 TiB 4.1 TiB 4.5 MiB 14 GiB 1.3 TiB > 76.13 1.15 199 up osd.28 > 30 hdd 5.51459 1.00000 5.5 TiB 4.5 TiB 4.5 TiB 8.2 MiB 15 GiB 1006 GiB > 82.18 1.24 215 up osd.30 > 36 ssd 0.34239 1.00000 351 GiB 10 GiB 184 MiB 8.1 GiB 2.0 GiB 340 GiB > 2.92 0.04 110 up osd.36 > -28 55.83066 - 56 TiB 35 TiB 34 TiB 17 GiB 132 GiB 21 TiB > 62.36 0.94 - rack 2-zwei > -7 27.91533 - 28 TiB 17 TiB 17 TiB 8.2 GiB 66 GiB 11 TiB > 62.27 0.94 - host onode2 > 5 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.0 MiB 12 GiB 2.1 TiB > 62.08 0.93 162 up osd.5 > 6 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.9 MiB 13 GiB 2.1 TiB > 62.13 0.93 162 up osd.6 > 7 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.3 TiB 3.7 MiB 12 GiB 2.1 TiB > 61.77 0.93 161 up osd.7 > 8 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.3 TiB 3.2 MiB 12 GiB 2.1 TiB > 61.75 0.93 161 up osd.8 > 38 hdd 5.51459 1.00000 5.5 TiB 3.7 TiB 3.7 TiB 3.7 MiB 14 GiB 1.8 TiB > 67.31 1.01 176 up osd.38 > 31 ssd 0.34239 1.00000 351 GiB 11 GiB 166 MiB 8.1 GiB 2.4 GiB 340 GiB > 3.04 0.05 110 up osd.31 > -16 27.91533 - 28 TiB 17 TiB 17 TiB 8.7 GiB 66 GiB 10 TiB > 62.44 0.94 - host onode5 > 17 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4 MiB 12 GiB 2.1 TiB > 62.15 0.94 162 up osd.17 > 18 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.0 MiB 13 GiB 2.1 TiB > 62.16 0.94 162 up osd.18 > 19 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.5 MiB 13 GiB 2.1 TiB > 62.14 0.94 162 up osd.19 > 20 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.5 MiB 13 GiB 2.1 TiB > 62.12 0.93 162 up osd.20 > 41 hdd 5.51459 1.00000 5.5 TiB 3.7 TiB 3.7 TiB 3.9 MiB 14 GiB 1.8 TiB > 67.31 1.01 176 up osd.41 > 34 ssd 0.34239 1.00000 351 GiB 11 GiB 192 MiB 8.7 GiB 1.8 GiB 340 GiB > 3.04 0.05 109 up osd.34 > -29 55.83066 - 56 TiB 38 TiB 38 TiB 16 GiB 138 GiB 17 TiB > 68.95 1.04 - rack 3-drei > -10 27.91533 - 28 TiB 17 TiB 17 TiB 8.1 GiB 63 GiB 11 TiB > 61.02 0.92 - host onode3 > 9 hdd 5.51459 1.00000 5.5 TiB 3.3 TiB 3.3 TiB 3.7 MiB 12 GiB 2.2 TiB > 60.63 0.91 158 up osd.9 > 10 hdd 5.51459 1.00000 5.5 TiB 3.3 TiB 3.3 TiB 3.1 MiB 12 GiB 2.2 TiB > 60.19 0.91 157 up osd.10 > 11 hdd 5.51459 1.00000 5.5 TiB 3.3 TiB 3.3 TiB 6.7 MiB 12 GiB 2.2 TiB > 60.27 0.91 157 up osd.11 > 12 hdd 5.51459 1.00000 5.5 TiB 3.3 TiB 3.3 TiB 4.1 MiB 12 GiB 2.2 TiB > 60.28 0.91 157 up osd.12 > 39 hdd 5.51459 1.00000 5.5 TiB 3.7 TiB 3.7 TiB 4.6 MiB 13 GiB 1.8 TiB > 67.34 1.01 176 up osd.39 > 32 ssd 0.34239 1.00000 351 GiB 10 GiB 271 MiB 8.1 GiB 1.8 GiB 341 GiB > 2.88 0.04 109 up osd.32 > -19 27.91533 - 28 TiB 21 TiB 21 TiB 8.1 GiB 74 GiB 6.5 TiB > 76.89 1.16 - host onode6 > 21 hdd 5.51459 1.00000 5.5 TiB 4.0 TiB 4.0 TiB 6.2 MiB 13 GiB 1.5 TiB > 72.79 1.10 190 up osd.21 > 22 hdd 5.51459 1.00000 5.5 TiB 4.5 TiB 4.5 TiB 5.1 MiB 16 GiB 1.0 TiB > 81.79 1.23 214 up osd.22 > 23 hdd 5.51459 1.00000 5.5 TiB 4.4 TiB 4.4 TiB 4.4 MiB 16 GiB 1.1 TiB > 80.29 1.21 210 up osd.23 > 24 hdd 5.51459 1.00000 5.5 TiB 4.3 TiB 4.2 TiB 6.7 MiB 14 GiB 1.3 TiB > 77.31 1.16 202 up osd.24 > 29 hdd 5.51459 1.00000 5.5 TiB 4.2 TiB 4.2 TiB 4.6 MiB 14 GiB 1.3 TiB > 76.86 1.16 201 up osd.29 > 35 ssd 0.34239 1.00000 351 GiB 10 GiB 208 MiB 8.1 GiB 1.9 GiB 340 GiB > 2.89 0.04 110 up osd.35 > TOTAL 195 TiB 130 TiB 128 TiB 58 GiB 476 GiB 66 TiB > 66.45 > MIN/MAX VAR: 0.04/1.24 STDDEV: 26.74 > > > better only for the class hdd > > $ ceph osd df tree class hdd > ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL > %USE VAR PGS STATUS TYPE NAME > -1 195.40730 - 193 TiB 130 TiB 128 TiB 169 MiB 462 GiB 63 TiB > 67.24 1.00 - root default > -25 195.40730 - 193 TiB 130 TiB 128 TiB 169 MiB 462 GiB 63 TiB > 67.24 1.00 - room PRZ > -26 195.40730 - 193 TiB 130 TiB 128 TiB 169 MiB 462 GiB 63 TiB > 67.24 1.00 - row rechts > -27 83.74599 - 83 TiB 57 TiB 56 TiB 81 MiB 200 GiB 26 TiB > 68.31 1.02 - rack 1-eins > -3 27.91533 - 28 TiB 18 TiB 17 TiB 31 MiB 64 GiB 10 TiB > 63.54 0.94 - host onode1 > 0 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.8 MiB 14 GiB 2.1 TiB > 62.48 0.93 163 up osd.0 > 1 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 6.5 MiB 12 GiB 2.1 TiB > 62.47 0.93 163 up osd.1 > 2 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 7.1 MiB 12 GiB 2.1 TiB > 62.53 0.93 163 up osd.2 > 3 hdd 5.51459 1.00000 5.5 TiB 3.5 TiB 3.4 TiB 7.5 MiB 12 GiB 2.0 TiB > 62.90 0.94 164 up osd.3 > 37 hdd 5.51459 1.00000 5.5 TiB 3.7 TiB 3.7 TiB 6.4 MiB 13 GiB 1.8 TiB > 67.32 1.00 176 up osd.37 > -13 27.91533 - 28 TiB 17 TiB 17 TiB 21 MiB 64 GiB 10 TiB > 63.38 0.94 - host onode4 > 13 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.0 MiB 13 GiB 2.1 TiB > 62.49 0.93 163 up osd.13 > 14 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.2 MiB 13 GiB 2.1 TiB > 62.49 0.93 163 up osd.14 > 15 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.4 MiB 12 GiB 2.1 TiB > 62.43 0.93 163 up osd.15 > 16 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 5.8 MiB 12 GiB 2.1 TiB > 62.13 0.92 162 up osd.16 > 40 hdd 5.51459 1.00000 5.5 TiB 3.7 TiB 3.7 TiB 3.6 MiB 13 GiB 1.8 TiB > 67.36 1.00 176 up osd.40 > -22 27.91533 - 28 TiB 22 TiB 21 TiB 28 MiB 72 GiB 6.1 TiB > 78.02 1.16 - host onode7 > 25 hdd 5.51459 1.00000 5.5 TiB 4.3 TiB 4.2 TiB 7.2 MiB 14 GiB 1.2 TiB > 77.59 1.15 203 up osd.25 > 26 hdd 5.51459 1.00000 5.5 TiB 4.3 TiB 4.3 TiB 4.7 MiB 14 GiB 1.2 TiB > 78.40 1.17 205 up osd.26 > 27 hdd 5.51459 1.00000 5.5 TiB 4.2 TiB 4.1 TiB 3.8 MiB 14 GiB 1.3 TiB > 75.80 1.13 198 up osd.27 > 28 hdd 5.51459 1.00000 5.5 TiB 4.2 TiB 4.1 TiB 4.5 MiB 14 GiB 1.3 TiB > 76.13 1.13 199 up osd.28 > 30 hdd 5.51459 1.00000 5.5 TiB 4.5 TiB 4.5 TiB 8.2 MiB 15 GiB 1006 GiB > 82.18 1.22 215 up osd.30 > -28 55.83066 - 55 TiB 35 TiB 34 TiB 38 MiB 128 GiB 20 TiB > 63.09 0.94 - rack 2-zwei > -7 27.91533 - 28 TiB 17 TiB 17 TiB 18 MiB 63 GiB 10 TiB > 63.01 0.94 - host onode2 > 5 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.0 MiB 12 GiB 2.1 TiB > 62.08 0.92 162 up osd.5 > 6 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.9 MiB 13 GiB 2.1 TiB > 62.13 0.92 162 up osd.6 > 7 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.3 TiB 3.7 MiB 12 GiB 2.1 TiB > 61.77 0.92 161 up osd.7 > 8 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.3 TiB 3.2 MiB 12 GiB 2.1 TiB > 61.75 0.92 161 up osd.8 > 38 hdd 5.51459 1.00000 5.5 TiB 3.7 TiB 3.7 TiB 3.7 MiB 14 GiB 1.8 TiB > 67.31 1.00 176 up osd.38 > -16 27.91533 - 28 TiB 17 TiB 17 TiB 20 MiB 65 GiB 10 TiB > 63.18 0.94 - host onode5 > 17 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4 MiB 12 GiB 2.1 TiB > 62.15 0.92 162 up osd.17 > 18 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.0 MiB 13 GiB 2.1 TiB > 62.16 0.92 162 up osd.18 > 19 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 4.5 MiB 13 GiB 2.1 TiB > 62.14 0.92 162 up osd.19 > 20 hdd 5.51459 1.00000 5.5 TiB 3.4 TiB 3.4 TiB 3.5 MiB 13 GiB 2.1 TiB > 62.12 0.92 162 up osd.20 > 41 hdd 5.51459 1.00000 5.5 TiB 3.7 TiB 3.7 TiB 3.9 MiB 14 GiB 1.8 TiB > 67.31 1.00 176 up osd.41 > -29 55.83066 - 55 TiB 38 TiB 38 TiB 49 MiB 134 GiB 17 TiB > 69.77 1.04 - rack 3-drei > -10 27.91533 - 28 TiB 17 TiB 17 TiB 22 MiB 62 GiB 11 TiB > 61.74 0.92 - host onode3 > 9 hdd 5.51459 1.00000 5.5 TiB 3.3 TiB 3.3 TiB 3.7 MiB 12 GiB 2.2 TiB > 60.63 0.90 158 up osd.9 > 10 hdd 5.51459 1.00000 5.5 TiB 3.3 TiB 3.3 TiB 3.1 MiB 12 GiB 2.2 TiB > 60.19 0.90 157 up osd.10 > 11 hdd 5.51459 1.00000 5.5 TiB 3.3 TiB 3.3 TiB 6.7 MiB 12 GiB 2.2 TiB > 60.27 0.90 157 up osd.11 > 12 hdd 5.51459 1.00000 5.5 TiB 3.3 TiB 3.3 TiB 4.1 MiB 12 GiB 2.2 TiB > 60.28 0.90 157 up osd.12 > 39 hdd 5.51459 1.00000 5.5 TiB 3.7 TiB 3.7 TiB 4.6 MiB 13 GiB 1.8 TiB > 67.34 1.00 176 up osd.39 > -19 27.91533 - 28 TiB 21 TiB 21 TiB 27 MiB 72 GiB 6.1 TiB > 77.81 1.16 - host onode6 > 21 hdd 5.51459 1.00000 5.5 TiB 4.0 TiB 4.0 TiB 6.2 MiB 13 GiB 1.5 TiB > 72.79 1.08 190 up osd.21 > 22 hdd 5.51459 1.00000 5.5 TiB 4.5 TiB 4.5 TiB 5.1 MiB 16 GiB 1.0 TiB > 81.79 1.22 214 up osd.22 > 23 hdd 5.51459 1.00000 5.5 TiB 4.4 TiB 4.4 TiB 4.4 MiB 16 GiB 1.1 TiB > 80.29 1.19 210 up osd.23 > 24 hdd 5.51459 1.00000 5.5 TiB 4.3 TiB 4.2 TiB 6.7 MiB 14 GiB 1.3 TiB > 77.31 1.15 202 up osd.24 > 29 hdd 5.51459 1.00000 5.5 TiB 4.2 TiB 4.2 TiB 4.6 MiB 14 GiB 1.3 TiB > 76.86 1.14 201 up osd.29 > TOTAL 193 TiB 130 TiB 128 TiB 169 MiB 462 GiB 63 TiB > 67.24 > MIN/MAX VAR: 0.90/1.22 STDDEV: 7.17 > > > > > ceph osd pool ls detail > pool 1 'cephfs_data' erasure size 6 min_size 5 crush_rule 1 object_hash > rjenkins pg_num 1024 pgp_num 1024 autoscale_mode on last_change 20353 lfor > 0/0/2366 flags hashpspool,ec_overwrites,selfmanaged_snaps max_bytes > 119457034600410 stripe_width 16384 target_size_ratio 0.85 application cephfs > removed_snaps > [2~4,7~27,2f~1e,4f~1f,6f~39,a9~5,af~1,b1~1,b3~1,b5~1,b7~1,b9~1,bb~1,bd~1,bf~1,c1~1,c3~1,c5~1,c7~1,c9~1] > pool 2 'cephfs_metadata' replicated size 3 min_size 2 crush_rule 2 > object_hash rjenkins pg_num 256 pgp_num 256 autoscale_mode on last_change 261 > lfor 0/0/259 flags hashpspool stripe_width 0 pg_autoscale_bias 4 pg_num_min > 16 recovery_priority 5 target_size_ratio 0.3 application cephfs > > > $ ceph osd crush rule dump > [ > { > "rule_id": 0, > "rule_name": "replicated_rule", > "ruleset": 0, > "type": 1, > "min_size": 1, > "max_size": 10, > "steps": [ > { > "op": "take", > "item": -1, > "item_name": "default" > }, > { > "op": "chooseleaf_firstn", > "num": 0, > "type": "host" > }, > { > "op": "emit" > } > ] > }, > { > "rule_id": 1, > "rule_name": "cephfs_data", > "ruleset": 1, > "type": 3, > "min_size": 3, > "max_size": 6, > "steps": [ > { > "op": "set_chooseleaf_tries", > "num": 5 > }, > { > "op": "set_choose_tries", > "num": 100 > }, > { > "op": "take", > "item": -2, > "item_name": "default~hdd" > }, > { > "op": "chooseleaf_indep", > "num": 0, > "type": "host" > }, > { > "op": "emit" > } > ] > }, > { > "rule_id": 2, > "rule_name": "rep_3_ssd", > "ruleset": 2, > "type": 1, > "min_size": 1, > "max_size": 10, > "steps": [ > { > "op": "take", > "item": -6, > "item_name": "default~ssd" > }, > { > "op": "chooseleaf_firstn", > "num": 0, > "type": "host" > }, > { > "op": "emit" > } > ] > } > ] > > > Tanks, > Lars > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email to ceph-users-le...@ceph.io -- Informationstechnologie Berlin-Brandenburgische Akademie der Wissenschaften Jägerstraße 22-23 10117 Berlin Tel.: +49 30 20370-352 http://www.bbaw.de _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io