There are three factors that impact disk utilization of an OSD:
 1. number of PGs on the OSD (determined by CRUSH)
 2. number of objects with each PG (better to pick a 2 power PG number to make 
this one more even)
 3. object size deviation

with 'ceph osd reweight-by-pg', you can tune (1). And if you would like to get 
a better understanding of what is the root cause in your cluster, you can find 
more information from 'pg dump', from where you can get the raw data for 1 and 
2.

Once the cluster is filled, you properly go with 'ceph osd 
reweight-by-utilization', be careful of that since it could incur lots of data 
movement...

----------------------------------------
> To: ceph-users@lists.ceph.com
> From: vedran.fu...@gmail.com
> Date: Fri, 14 Aug 2015 00:15:17 +0200
> Subject: Re: [ceph-users] OSD space imbalance
>
> On 13.08.2015 18:01, GuangYang wrote:
>> Try 'ceph osd  <int>' right after creating the pools?
>
> Would it do any good now when pool is in use and nearly full as I can't
> re-create it now. Also, what's the integer argument in the command
> above? I failed to find proper explanation in the docs.
Please check it out here - 
https://github.com/ceph/ceph/blob/master/src/mon/OSDMonitor.cc#L469
>
>> What is the typical object size in the cluster?
>
> Around 50 MB.
>
>
> Thanks,
> Vedran
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
                                          
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to