Mike - unless things have changed in the latest versions(s) of Ceph, I *not* believe CRUSH will be successful in creating a valid PG map if the ’n' value is 10 (k+m), your host count is 6, and your failure domain is set to host. You’ll need to increase your host count to match or exceed ’n', change the failure domain to OSD, or alter the k+m config to something that is more compatible to your host count…otherwise you’ll end up with incomplete PG’s. Also note that having more failure domains (i.e. - hosts) than your ’n’ value is recommended.
Beyond that, you’re likely to run operational challenges putting that many drives behind a single CPU-complex when the host count is quite low. My $.02. -- Paul On Nov 10, 2015, at 2:29 AM, Mike Almateia <mike.almat...@gmail.com<mailto:mike.almat...@gmail.com>> wrote: Hello. For our CCTV storing streams project we decided to use Ceph cluster with EC pool. Input requirements is not scary: max. 15Gbit/s input traffic from CCTV, 30 day storing, 99% write operations, a cluster must has grow up with out downtime. By now our vision of architecture it like: * 6 JBOD with 90 HDD 8Tb capacity each (540 HDD total) * 6 Ceph servers connected to it own JBOD (we will have 6 pairs: 1 Server + 1 JBOD). Ceph servers hardware details: * 2 x E5-2690v3 : 24 core (w/o HT), 2.6 Ghz each * 256 Gb RAM DDR4 * 4 x 10Gbit/s NIC port (2 for Client network and 2 for Cluster Network) * servers also have 4 (8) x 2.5" HDD SATA on board for Cache Tiering Feature (because ceph clients can't directly talk with EC pool) * Two HBA SAS controllers work with multipathing feature, for HA scenario. * For Ceph monitor functionality 3 servers have 2 SSD in Software RAID1 Some Ceph configuration rules: * EC pools with K=7 and M=3 * EC plugin - ISA * technique = reed_sol_van * ruleset-failure-domain = host * near full ratio = 0.75 * OSD journal partition on the same disk We think that first and second problems it will be CPU and RAM on Ceph servers. Any ideas? it is can fly? _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Paul Evans Principal Architect Daystrom Technology Group m: 707-479-1034 o: 800-656-3224 x511 f: 650-472-4005 e: paul.ev...@daystrom.com<mailto:paul.ev...@daystrom.com>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com