ceph df detail:
[root@k8s-1 ~]# ceph df detail
--- RAW STORAGE ---
CLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED
hdd    600 GiB  600 GiB  157 MiB   157 MiB       0.03
TOTAL  600 GiB  600 GiB  157 MiB   157 MiB       0.03
 
--- POOLS ---
POOL                                    ID  PGS   STORED   (DATA)  (OMAP)  
OBJECTS     USED   (DATA)  (OMAP)  %USED  MAX AVAIL  QUOTA OBJECTS  QUOTA BYTES 
 DIRTY  USED COMPR  UNDER COMPR
device_health_metrics                    1    1      0 B      0 B     0 B       
 0      0 B      0 B     0 B      0    158 GiB            N/A          N/A    
N/A         0 B          0 B
os-gwbtwlltuklwxrpl.rgw.buckets.index    2  230      0 B      0 B     0 B       
11      0 B      0 B     0 B      0    158 GiB            N/A          N/A    
N/A         0 B          0 B
os-gwbtwlltuklwxrpl.rgw.control          3    8      0 B      0 B     0 B       
 8      0 B      0 B     0 B      0    158 GiB            N/A          N/A    
N/A         0 B          0 B
os-gwbtwlltuklwxrpl.rgw.log              4    8  3.7 KiB  3.7 KiB     0 B      
180  420 KiB  420 KiB     0 B      0    158 GiB            N/A          N/A    
N/A         0 B          0 B
os-gwbtwlltuklwxrpl.rgw.meta             5    8  1.9 KiB  1.9 KiB     0 B       
 7   72 KiB   72 KiB     0 B      0    158 GiB            N/A          N/A    
N/A         0 B          0 B
.rgw.root                                6    8  4.9 KiB  4.9 KiB     0 B       
16  180 KiB  180 KiB     0 B      0    158 GiB            N/A          N/A    
N/A         0 B          0 B
os-gwbtwlltuklwxrpl.rgw.buckets.non-ec   7    8      0 B      0 B     0 B       
 0      0 B      0 B     0 B      0    158 GiB            N/A          N/A    
N/A         0 B          0 B
os-gwbtwlltuklwxrpl.rgw.otp              8    8      0 B      0 B     0 B       
 0      0 B      0 B     0 B      0    158 GiB            N/A          N/A    
N/A         0 B          0 B
os-gwbtwlltuklwxrpl.rgw.buckets.data     9   32      0 B      0 B     0 B       
 0      0 B      0 B     0 B      0    317 GiB            N/A          N/A    
N/A         0 B          0 B

ceph osd pool ls detail:
[root@k8s-1 ~]# ceph osd pool ls detail
pool 1 'device_health_metrics' replicated size 3 min_size 2 crush_rule 1 
object_hash rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 25 flags 
hashpspool stripe_width 0 pg_num_max 32 pg_num_min 1 application 
mgr_devicehealth
pool 2 'os-gwbtwlltuklwxrpl.rgw.buckets.index' replicated size 3 min_size 2 
crush_rule 5 object_hash rjenkins pg_num 217 pgp_num 209 pg_num_target 32 
pgp_num_target 32 autoscale_mode off last_change 322 lfor 0/322/320 flags 
hashpspool stripe_width 0 compression_mode none pg_num_min 8 target_size_ratio 
0.5 application rook-ceph-rgw
pool 3 'os-gwbtwlltuklwxrpl.rgw.control' replicated size 3 min_size 2 
crush_rule 7 object_hash rjenkins pg_num 8 pgp_num 8 autoscale_mode on 
last_change 36 flags hashpspool stripe_width 0 compression_mode none pg_num_min 
8 target_size_ratio 0.5 application rook-ceph-rgw
pool 4 'os-gwbtwlltuklwxrpl.rgw.log' replicated size 3 min_size 2 crush_rule 6 
object_hash rjenkins pg_num 8 pgp_num 8 autoscale_mode on last_change 35 flags 
hashpspool stripe_width 0 compression_mode none pg_num_min 8 target_size_ratio 
0.5 application rook-ceph-rgw
pool 5 'os-gwbtwlltuklwxrpl.rgw.meta' replicated size 3 min_size 2 crush_rule 2 
object_hash rjenkins pg_num 8 pgp_num 8 autoscale_mode on last_change 35 flags 
hashpspool stripe_width 0 compression_mode none pg_num_min 8 target_size_ratio 
0.5 application rook-ceph-rgw
pool 6 '.rgw.root' replicated size 3 min_size 2 crush_rule 4 object_hash 
rjenkins pg_num 8 pgp_num 8 autoscale_mode on last_change 35 flags hashpspool 
stripe_width 0 compression_mode none pg_num_min 8 target_size_ratio 0.5 
application rook-ceph-rgw
pool 7 'os-gwbtwlltuklwxrpl.rgw.buckets.non-ec' replicated size 3 min_size 2 
crush_rule 8 object_hash rjenkins pg_num 8 pgp_num 8 autoscale_mode on 
last_change 35 flags hashpspool stripe_width 0 compression_mode none pg_num_min 
8 target_size_ratio 0.5 application rook-ceph-rgw
pool 8 'os-gwbtwlltuklwxrpl.rgw.otp' replicated size 3 min_size 2 crush_rule 3 
object_hash rjenkins pg_num 8 pgp_num 8 autoscale_mode on last_change 36 flags 
hashpspool stripe_width 0 compression_mode none pg_num_min 8 target_size_ratio 
0.5 application rook-ceph-rgw
pool 9 'os-gwbtwlltuklwxrpl.rgw.buckets.data' erasure profile 
os-gwbtwlltuklwxrpl.rgw.buckets.data_ecprofile size 3 min_size 2 crush_rule 9 
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode off last_change 70 
flags hashpspool,ec_overwrites stripe_width 8192 compression_mode none 
target_size_ratio 0.5 application rook-ceph-rgw

I tested under a smaller ceph cluster with 6 osds, and set the pg num of the 
index pool to 256, the reduce it to 32, it still slowly.
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to