jp...@cam.ac.uk said: > I can't speak for this particular situation or solution, but I think in > principle you are wrong. Networks are fast. Hard drives are slow. Put a > 10G connection between your storage and your front ends and you'll have the > bandwidth[1]. Actually if you really were hitting 1000x8Mbits I'd put 2, > but that is just a question of scale. In a different situation I have boxes > which peak at around 7 Gb/s down a 10G link (in reality I don't need that > much because it is all about the IOPS for me). That is with just twelve 15k > disks. Your situation appears to be pretty ideal for storage hardware, so > perfectly achievable from an appliance.
Depending on usage, I disagree with your bandwidth and latency figures above. An X4540, or an X4170 with J4000 JBOD's, has more bandwidth to its disks than 10Gbit ethernet. You would need three 10GbE interfaces between your CPU and the storage appliance to equal the bandwidth of a single 8-port 3Gb/s SAS HBA (five of them for 6Gb/s SAS). It's also the case that the Unified Storage platform doesn't have enough bandwidth to drive more than four 10GbE ports at their full speed: http://dtrace.org/blogs/brendan/2009/09/22/7410-hardware-update-and-analyzing-t he-hypertransport/ We have a customer (internal to the university here) that does high throughput gene sequencing. They like a server which can hold the large amounts of data, do a first pass analysis on it, and then serve it up over the network to a compute cluster for further computation. Oracle has nothing in their product line (anymore) to meet that need. They ended up ordering an 8U chassis w/40x 2TB drives in it, and are willing to pay the $2k/yr retail ransom to Oracle to run Solaris (ZFS) on it, at least for the first year. Maybe OpenIndiana next year, we'll see. Bye Oracle.... Regards, Marion _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss