On 5/26/22 14:09, Wesley Dillingham wrote:
What does "ceph osd pool ls detail" say?
$ ceph osd pool ls detail
pool 0 'rbd' replicated size 2 min_size 1 crush_rule 0 object_hash
rjenkins pg_num 64 pgp_num 64 autoscale_mode on last_change 44740 flags
hashpspool,selfmanaged_snaps stripe_width 0 application rbd
pool 1 '.rgw.root' replicated size 2 min_size 1 crush_rule 0 object_hash
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 44740 lfor
0/0/31483 owner 18446744073709551615 flags hashpspool stripe_width 0
application rgw
pool 2 'default.rgw.control' replicated size 2 min_size 1 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
44740 lfor 0/0/31469 owner 18446744073709551615 flags hashpspool
stripe_width 0 application rgw
pool 3 'default.rgw.data.root' replicated size 2 min_size 1 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
44740 lfor 0/0/31471 owner 18446744073709551615 flags hashpspool
stripe_width 0 application rgw
pool 4 'default.rgw.gc' replicated size 2 min_size 1 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
44740 lfor 0/0/31471 owner 18446744073709551615 flags hashpspool
stripe_width 0 application rgw
pool 5 'default.rgw.log' replicated size 2 min_size 1 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
44740 lfor 0/0/31387 owner 18446744073709551615 flags hashpspool
stripe_width 0 application rgw
pool 6 'default.rgw.users.uid' replicated size 2 min_size 1 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
44740 lfor 0/0/31387 flags hashpspool stripe_width 0 application rgw
pool 12 'mathfs_data' replicated size 2 min_size 1 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
44740 lfor 0/31370/31368 flags hashpspool stripe_width 0 application cephfs
pool 13 'mathfs_metadata' replicated size 2 min_size 2 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
44740 lfor 0/27164/27162 flags hashpspool stripe_width 0 application cephfs
pool 15 'default.rgw.lc' replicated size 2 min_size 1 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
44740 lfor 0/0/31374 flags hashpspool stripe_width 0 application rgw
pool 21 'libvirt' replicated size 3 min_size 1 crush_rule 0 object_hash
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 56244 lfor
0/33144/33142 flags hashpspool,selfmanaged_snaps stripe_width 0
application rbd
pool 36 'monthly_archive_metadata' replicated size 2 min_size 1
crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on
last_change 45338 lfor 0/27845/27843 flags hashpspool stripe_width 0
application cephfs
pool 37 'monthly_archive_data' replicated size 2 min_size 1 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
45334 lfor 0/44535/44533 flags hashpspool stripe_width 0 application cephfs
pool 38 'device_health_metrics' replicated size 2 min_size 1 crush_rule
0 object_hash rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change
56507 flags hashpspool stripe_width 0 pg_num_min 1 application
mgr_devicehealth
pool 41 'lensfun_metadata' replicated size 2 min_size 1 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
54066 flags hashpspool stripe_width 0 pg_autoscale_bias 4 pg_num_min 16
recovery_priority 5 application cephfs
pool 42 'lensfun_data' replicated size 2 min_size 1 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change
54066 flags hashpspool stripe_width 0 application cephfs
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io