Thanks Roy, we're planning to grow on this cluster if can get the performance that we need, the idea is to run non-relational databases here, so it would be high-io intensive. We are talking in grow terms of about 40-50 OSD servers with no more than 6 OSD daemons per server. If you got some hints or docs out there on how to compile ceph with accelio it would be awesome.
*German* 2015-09-01 15:31 GMT-03:00 Somnath Roy <somnath....@sandisk.com>: > Thanks ! > > I think you should try installing from the ceph mainstream..There are some > bug fixes went on after Hammer (not sure if it is backported).. > > I would say try with 1 drive -> 1 OSD first since presently we have seen > some stability issues (mainly due to resource constraint) with more OSDs in > a box. > > The another point is, installation itself is not straight forward. You > need to build all the components probably, not sure if it is added as git > submodule or not, Vu , could you please confirm ? > > > > Since we are working to make this solution work at scale, could you please > give us some idea what is the scale you are looking at for future > deployment ? > > > > Regards > > Somnath > > > > *From:* German Anders [mailto:gand...@despegar.com] > *Sent:* Tuesday, September 01, 2015 11:19 AM > *To:* Somnath Roy > *Cc:* Robert LeBlanc; ceph-users > > *Subject:* Re: [ceph-users] Accelio & Ceph > > > > Hi Roy, > > I understand, we are looking for using accelio with an starting small > cluster of 3 mon and 8 osd servers: > > 3x MON servers > > 2x Intel Xeon E5-2630v3 @2.40Ghz (32C with HT) > > 24x 16GB DIMM DDR3 1333Mhz (384GB) > > 2x 120GB Intel SSD DC S3500 (RAID-1 for OS) > > 1x ConnectX-3 VPI FDR 56Gb/s ADPT DP > > 4x OSD servers > > 2x Intel Xeon E5-2609v2 @2.50Ghz (8C) > > 8x 16GB DIMM DDR3 1333Mhz (128GB) > > 2x 120GB Intel SSD DC SC3500 (RAID-1 for OS) > > 3x 120GB Intel SSD DC SC3500 (Journals) > > 4x 800GB Intel SSD DC SC3510 (OSD-SSD-POOL) > > 5x 3TB SAS (OSD-SAS-POOL) > > 1x ConnectX-3 VPI FDR 56Gb/s ADPT DP > > 4x OSD servers > > 2x Intel Xeon E5-2650v2 @2.60Ghz (32C with HT) > > 8x 16GB DIMM DDR3 1866Mhz (128GB) > > 2x 200GB Intel SSD DC S3700 (RAID-1 for OS) > > 3x 200GB Intel SSD DC S3700 (Journals) > > 4x 800GB Intel SSD DC SC3510 (OSD-SSD-POOL) > > 5x 3TB SAS (OSD-SAS-POOL) > > 1x ConnectX-3 VPI FDR 56Gb/s ADPT DP > > and thinking of using *infernalis v.9.0.0* or *hammer* release? comments? > recommendations? > > > *German* > > > > 2015-09-01 14:46 GMT-03:00 Somnath Roy <somnath....@sandisk.com>: > > Hi German, > > We are working on to make it production ready ASAP. As you know RDMA is > very resource constrained and at the same time will outperform TCP as well. > There will be some definite tradeoff between cost Vs Performance. > > We are lacking on ideas on how big the RDMA deployment could be and it > will be really helpful if you can give some idea on how you are planning to > deploy that (i.e how many nodes/OSDs/SSD or HDDs/ EC or Replication etc. > etc.). > > > > Thanks & Regards > > Somnath > > > > *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On Behalf > Of *German Anders > *Sent:* Tuesday, September 01, 2015 10:39 AM > *To:* Robert LeBlanc > *Cc:* ceph-users > *Subject:* Re: [ceph-users] Accelio & Ceph > > > > Thanks a lot for the quick response Robert, any idea when it's going to be > ready for production? any alternative solution for similar-performance? > > Best regards, > > > *German * > > > > 2015-09-01 13:42 GMT-03:00 Robert LeBlanc <rob...@leblancnet.us>: > > -----BEGIN PGP SIGNED MESSAGE----- > > Hash: SHA256 > > > > Accelio and Ceph are still in heavy development and not ready for production. > > > > - ---------------- > > Robert LeBlanc > > PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 > > > > On Tue, Sep 1, 2015 at 10:31 AM, German Anders wrote: > > Hi cephers, > > > > I would like to know the status for production-ready of Accelio & Ceph, does > anyone had a home-made procedure implemented with Ubuntu? > > > > recommendations, comments? > > > > Thanks in advance, > > > > Best regards, > > > > German > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > -----BEGIN PGP SIGNATURE----- > > Version: Mailvelope v1.0.2 > > Comment: https://www.mailvelope.com > > > > wsFcBAEBCAAQBQJV5dWKCRDmVDuy+mK58QAAZWcQAKIRYhnlSzIQJ9PGaC1J > > FGYxZ9IOmXX89IbpZuM8Ns8Q1Y52SrYkez8jwtB/A1OWXH0uw2GT45shDfzX > > xFaqRVVHnjI7MiO+aijGkDZLrdE5fvGfTAOa1m2ovlx7BWRG6k0aSeqdMr92 > > OB/n2ona94ILvHW/Uq/o5YnoFsThUdTTRWckWeRMKIz9eA7v+bneukjXyLf/ > > VwFAk0V9LevzNZY83nARYThDfL20SYT05dAhJ6bbzYFowdymZcNWTEDkUY02 > > m76bhEQO4k3MypL+kv0YyFi3cDkMBa4CaCm3UwRWC5KG6MlQnFl+f3UQuOwV > > YhYkagw2qUP4rx+/5LIAU+WEzegZ+3mDgk0qIB6pa7TK5Gk4hvHZG884YpXA > > Fa6Lj9x7gQjszLI1esW1zuNhlTBUJfxygfdJQPV2w/9cjjFlXG8QgmZcgyJF > > XjtH/T1BK8t7x6IgerXBPEjPlU6tYI75HSSryarFH9ntKIIr6Yrcaaa8heLD > > /7S/S05yQ2TcfnkVPGapDzJ2Ko5h5gwO/29EIlOsYiHCwDYXDonRFFUrRa2Z > > SzSq9iiCywglYtqqzaDpqeU5soPIaijHn7ELSEq51Lc6D19pRdEMdmFnxcmt > > 8QAYEihGnckbcSLdwm1nOP0Nme5ixyGLxcEfxUYv6hTxhJt4RuAj83f2cFxh > > TiL2 > > =oSrX > > -----END PGP SIGNATURE----- > > > > > ------------------------------ > > > PLEASE NOTE: The information contained in this electronic mail message is > intended only for the use of the designated recipient(s) named above. If > the reader of this message is not the intended recipient, you are hereby > notified that you have received this message in error and that any review, > dissemination, distribution, or copying of this message is strictly > prohibited. If you have received this communication in error, please notify > the sender by telephone or e-mail (as shown above) immediately and destroy > any and all copies of this message in your possession (whether hard copies > or electronically stored copies). > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com