Thanks for posting your experiences John, very interesting read. I think the 
golden rule of around 1Ghz is still a realistic goal to aim for. It looks like 
you probably have around 16ghz for 60OSD's, or 0.26Ghz per OSD. Do you have any 
idea on how much CPU you think you would need to just be able to get away with 
it?

I have 24Ghz for 12 OSD's (2x2620v2) and I typically don't see CPU usage over 
about 20%, which indicates to me the bare minimum for a replicated pool is 
probably around 0.5Ghz per 7.2k rpm OSD. The next nodes we have will certainly 
have less CPU.

> -----Original Message-----
> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of
> John Hogenmiller
> Sent: 17 February 2016 03:04
> To: Василий Ангапов <anga...@gmail.com>; ceph-users@lists.ceph.com
> Subject: Re: [ceph-users] Recomendations for building 1PB RadosGW with
> Erasure Code
> 
> Turns out i didn't do reply-all.
> 
> On Tue, Feb 16, 2016 at 9:18 AM, John Hogenmiller <j...@hogenmiller.net>
> wrote:
> > And again - is dual Xeon's power enough for 60-disk node and Erasure
> Code?
> 
> 
> This is something I've been attempting to determine as well. I'm not yet
> getting
> I'm testing with some white-label hardware, but essentially supermicro
> 2twinu's with a pair of E5-2609 Xeons and 64GB of
> memory.  (http://www.supermicro.com/products/system/2U/6028/SYS-
> 6028TR-HTFR.cfm). This is attached to DAEs with 60 x 6TB drives, in JBOD.
> 
> Conversely, Supermicro sells a 72-disk OSD node, which Redhat considers a
> supported "reference architecture" device. The processors in those nodes
> are E5-269 12-core, vs what I have which is quad-
> core. http://www.supermicro.com/solutions/storage_ceph.cfm  (SSG-
> 6048R-OSD432). I would highly recommend reflecting on the supermicro
> hardware and using that as your reference as well. If you could get an eval
> unit, use that to compare with the hardware you're working with.
> 
> I currently have mine setup with 7 nodes, 60 OSDs each, radosgw running
> one each node, and 5 ceph monitors. I plan to move the monitors to their
> own dedicated hardware, and in reading, I may only need 3 to manage the
> 420 OSDs.   I am currently just setup for replication instead of EC, though I
> want to redo this cluster to use EC. Also, I am still trying to work out how
> much of an impact placement groups have on performance, and I may have a
> performance-hampering amount..
> 
> We test the system using locust speaking S3 to the radosgw. Transactions are
> distributed equally across all 7 nodes and we track the statistics. We started
> first emulating 1000 users and got over 4Gbps, but load average on all nodes
> was in the mid-100s, and after 15 minutes we started getting socket
> timeouts. We stopped the test, let load settle, and started back at 100
> users.  We've been running this test about 5 days now.  Load average on all
> nodes floats between 40 and 70. The nodes with ceph-mon running on them
> do not appear to be taxed any more than the ones without. The radosgw
> itself seems to take up a decent amount of cpu (running civetweb, no
> ssl).  iowait is non existent, everything appears to be cpu bound.
> 
> At 1000 users, we had 4.3Gbps of PUTs and 2.2Gbps of GETs. Did not capture
> the TPS on that short test.
> At 100 users, we're pushing 2Gbps  in PUTs and 1.24Gpbs in GETs. Averaging
> 115 TPS.
> 
> All in all, the speeds are not bad for a single rack, but the CPU utilization 
> is a
> big concern. We're currently using other (proprietary) object storage
> platforms on this hardware configuration. They have their own set of issues,
> but CPU utilization is typically not the problem, even at higher utilization.
> 
> 
> 
> root@ljb01:/home/ceph/rain-cluster# ceph status
>     cluster 4ebe7995-6a33-42be-bd4d-20f51d02ae45
>      health HEALTH_OK
>      monmap e5: 5 mons at {hail02-r01-06=172.29.4.153:6789/0,hail02-r01-
> 08=172.29.4.155:6789/0,rain02-r01-01=172.29.4.148:6789/0,rain02-r01-
> 03=172.29.4.150:6789/0,rain02-r01-04=172.29.4.151:6789/0}
>             election epoch 86, quorum 0,1,2,3,4 
> rain02-r01-01,rain02-r01-03,rain02-
> r01-04,hail02-r01-06,hail02-r01-08
>      osdmap e2543: 423 osds: 419 up, 419 in
>             flags sortbitwise
>       pgmap v676131: 33848 pgs, 14 pools, 50834 GB data, 29660 kobjects
>             149 TB used, 2134 TB / 2284 TB avail
>                33848 active+clean
>   client io 129 MB/s rd, 182 MB/s wr, 1562 op/s
> 
> 
> 
>  # ceph-osd + ceph-mon + radosgw
> top - 13:29:22 up 40 days, 22:05,  1 user,  load average: 47.76, 47.33, 47.08
> Tasks: 1001 total,   7 running, 994 sleeping,   0 stopped,   0 zombie
> %Cpu(s): 39.2 us, 44.7 sy,  0.0 ni,  9.9 id,  2.4 wa,  0.0 hi,  3.7 si,  0.0 
> st
> KiB Mem:  65873180 total, 64818176 used,  1055004 free,     9324 buffers
> KiB Swap:  8388604 total,  7801828 used,   586776 free. 17610868 cached Mem
> 
>     PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+
> COMMAND
>  178129 ceph      20   0 3066452 618060   5440 S  54.6  0.9   2678:49 ceph-
> osd
>  218049 ceph      20   0 6261880 179704   2872 S  33.4  0.3   1852:14 radosgw
>  165529 ceph      20   0 2915332 579064   4308 S  19.7  0.9 530:12.65 ceph-osd
>  185193 ceph      20   0 2932696 585724   4412 S  19.1  0.9 545:20.31 ceph-osd
>   52334 ceph      20   0 3030300 618868   4328 S  15.8  0.9 543:53.64 ceph-osd
>   23124 ceph      20   0 3037740 607088   4440 S  15.2  0.9 461:03.98 ceph-osd
>  154031 ceph      20   0 2982344 525428   4044 S  14.9  0.8 587:17.62 ceph-osd
>  191278 ceph      20   0 2835208 570100   4700 S  14.9  0.9 547:11.66 ceph-osd
> 
>  # ceph-osd + radosgw (no ceph-mon)
> 
>  top - 13:31:22 up 40 days, 22:06,  1 user,  load average: 64.25, 59.76, 58.17
> Tasks: 1015 total,   4 running, 1011 sleeping,   0 stopped,   0 zombie
> %Cpu0  : 24.2 us, 48.5 sy,  0.0 ni, 10.9 id,  1.2 wa,  0.0 hi, 15.2 si,  0.0 
> st
> %Cpu1  : 30.8 us, 49.7 sy,  0.0 ni, 13.5 id,  1.8 wa,  0.0 hi,  4.2 si,  0.0 
> st
> %Cpu2  : 33.9 us, 49.5 sy,  0.0 ni, 10.5 id,  2.1 wa,  0.0 hi,  3.9 si,  0.0 
> st
> %Cpu3  : 31.3 us, 52.4 sy,  0.0 ni, 10.5 id,  2.7 wa,  0.0 hi,  3.0 si,  0.0 
> st
> %Cpu4  : 34.7 us, 41.3 sy,  0.0 ni, 20.4 id,  3.3 wa,  0.0 hi,  0.3 si,  0.0 
> st
> %Cpu5  : 38.3 us, 36.7 sy,  0.0 ni, 21.0 id,  3.7 wa,  0.0 hi,  0.3 si,  0.0 
> st
> %Cpu6  : 36.6 us, 37.5 sy,  0.0 ni, 19.8 id,  6.1 wa,  0.0 hi,  0.0 si,  0.0 
> st
> %Cpu7  : 38.0 us, 38.0 sy,  0.0 ni, 19.5 id,  4.3 wa,  0.0 hi,  0.3 si,  0.0 
> st
> KiB Mem:  65873180 total, 61946688 used,  3926492 free,     1260 buffers
> KiB Swap:  8388604 total,  7080048 used,  1308556 free.  7910772 cached Mem
> 
>     PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+
> COMMAND
>  108861 ceph      20   0 6283364 279464   3024 S  27.6  0.4   1684:30 radosgw
>     546 root      20   0       0      0      0 R  23.4  0.0 579:55.28 kswapd0
>  184953 ceph      20   0 2971100 576784   4348 S  23.1  0.9   1265:58 ceph-osd
>  178967 ceph      20   0 2970500 594756   6000 S  18.9  0.9 505:27.13 ceph-osd
>  184105 ceph      20   0 3096276 627944   7096 S  18.0  1.0 581:12.28 ceph-osd
>   56073 ceph      20   0 2888244 542024   4836 S  13.5  0.8 530:41.86 ceph-osd
>   55083 ceph      20   0 2819060 518500   5052 S  13.2  0.8 513:21.96 ceph-osd
>  175578 ceph      20   0 3083096 630712   5564 S  13.2  1.0 591:10.91 ceph-osd
>  180725 ceph      20   0 2915828 553240   4836 S  12.9  0.8 518:22.47 ceph-
> osd
> 


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to