Guys,

Thanks for the responses it is appreciated.

On 8/28/12 5:28 PM, Bryan Whitehead wrote:
I'f found pricing for Infiniband switches / cards to be cheaper than
10G cards/switches with the addition of being 4X fast.

I will look into this but putting all of our compute on Infiniband may be cost prohibitive.


On Tue, Aug 28, 2012 at 11:44 AM, Joe Topjian <j...@topjian.net> wrote:
Hi Matt,

On Tue, Aug 28, 2012 at 9:29 AM, Matt Weil <mw...@genome.wustl.edu> wrote:

Since we are on the subject of hardware what would be the perfect fit for
a gluster brick. We where looking at a PowerEdge C2100 Rack Server.


Just a note: the c2100 has been superseded by the Dell r720xd. Although the
r720 is not part of the c-series, it's their official replacement.

I looked at these but they only hold 8 3.5" drives verses 12 and two on the inside on the 2100. I will ask our rep about this.

Do you typically run hot spares or just keep cold spares handy?



During testing I found it pretty easy to saturate 1 Gig network links.
This was also the case when multiple links where bonded together.  Are there
any cheap 10 gig switch alternatives that anyone would suggest?


While not necessarily cheap, I've had great luck with Arista 7050 switches.

also looking at dell's new force 10 switches. Wonder how they compare price wise.


We implement them in sets of two, linked together. We then use dual-port
10gb NICs and connect each NIC to each switch. It gives multiple layers of
redundancy + a theoretical 20gb throughput per server.

Thanks,
Joe

_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

Reply via email to