Thanks for reply, >>ceph df
GLOBAL: SIZE AVAIL RAW USED %RAW USED 13910G 2472G 11437G 82.22 POOLS: NAME ID USED %USED MAX AVAIL OBJECTS rbd 0 3792G 27.26 615G 971526 How to free raw used space? From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Henrik Korkuc Sent: Tuesday, March 10, 2015 10:13 AM To: ceph-users@lists.ceph.com Subject: Re: [ceph-users] Ceph free space On 3/10/15 11:06, Mateusz Skała wrote: Hi, In my cluster is something wrong with free space. In cluster with 10OSD (5*1TB + 5*2TB) ‘ceph –s’ shows: 11425 GB used, 2485 GB / 13910 GB avail But I have only 2 rbd disks in one pool (‘rbd’): >>rados df pool name category KB objects clones degraded unfound rd rd KB wr wr KB rbd - 3976154023 971434 0 6474 0 11542224 1391869743 742847 385900453 total used 11988041672 971434 total avail 2598378648 total space 14586420320 >>rbd ls vm-100-disk-1 vm-100-disk-2 >>rbd info vm-100-disk-1 rbd image 'vm-100-disk-1': size 16384 MB in 4096 objects order 22 (4096 kB objects) block_name_prefix: rbd_data.14ef2ae8944a format: 2 features: layering >>rbd info vm-100-disk-2 rbd image 'vm-100-disk-2': size 4096 GB in 1048576 objects order 22 (4096 kB objects) block_name_prefix: rbd_data.15682ae8944a format: 2 features: layering So my rbd disks use only 4112GB. Default size of cluster is 2 so used space should be ca 8224GB, why ceph –s shows 11425 GB ? Can someone explain this situation? Thanks, Mateusz Hey, what does "ceph df" show? ceph -s shows raw disk usage so there will be some overhead from file system, also maybe you left some files there? _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com