Thanks Greg, No i am more into large scale RADOS system not filesystem . however for geographic distributed datacentres specially when network flactuate how to handle that as i read it seems CEPH need big pipe of network
/Zee On Fri, Jan 9, 2015 at 7:15 PM, Gregory Farnum <g...@gregs42.com> wrote: > On Thu, Jan 8, 2015 at 5:46 AM, Zeeshan Ali Shah <zas...@pdc.kth.se> > wrote: > > I just finished configuring ceph up to 100 TB with openstack ... Since we > > are also using Lustre in our HPC machines , just wondering what is the > > bottle neck in ceph going on Peta Scale like Lustre . > > > > any idea ? or someone tried it > > If you're talking about people building a petabyte Ceph system, there > are *many* who run clusters of that size. If you're talking about the > Ceph filesystem as a replacement for Lustre at that scale, the concern > is less about the raw amount of data and more about the resiliency of > the current code base at that size...but if you want to try it out and > tell us what problems you run into we will love you forever. ;) > (The scalable file system use case is what actually spawned the Ceph > project, so in theory there shouldn't be any serious scaling > bottlenecks. In practice it will depend on what kind of metadata > throughput you need because the multi-MDS stuff is improving but still > less stable.) > -Greg > -- Regards Zeeshan Ali Shah System Administrator - PDC HPC PhD researcher (IT security) Kungliga Tekniska Hogskolan +46 8 790 9115 http://www.pdc.kth.se/members/zashah
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com