Jake,

Most of thenumbers you are showing are just the clocked speeds on busses and
cables. Certainly the components can clock at those speeds, but the biggest
issue with lower-end components is whether they can actually feed and
sustain data at that rate, and how well they handle contention for
resources. Generally this comes down to size of buffers, and whether have
fast or wide enough processors at those interface points.

But it sounds like for the business Backblaze is in they are building
something that is big rather than fast (or at least something that is fast
enough). And certainly if you can parallelize the system enough (and
maintain reliablity) then you probably even achieve cheap and fast.

Regards, Martin

martinvisse...@gmail.com


On Thu, Sep 3, 2009 at 12:40 PM, Jake Anderson <ya...@vapourforge.com>wrote:

> On 03/09/09 10:37, Mark Walkom wrote:
>
>> I was thinking the same, but I reckon because they are just backing
>> up/archiving data it wouldn't be too bad.
>> ie They aren't looking for huge performance, just huge, cheap storage.
>>
>>
>> 2009/9/3 Morgan Storey<m...@morganstorey.com>
>>
>>
>>
>>> I know I am a geek but that is hot.
>>> I am wondering if they see any throughput issues with the sata backplanes
>>> and pci sata cards.
>>>
>>>
>> The backplanes are probably fine, each sata cable is good for
> ~300mbytes/sec most physical disks couldn't hope to hit that.
> lesse what their maximum xfer rate is.
> each drive can hit 103mb/sec average (better than I thought)
> each sata channel will max out at 250mbyte/sec
>
> so they are going to be loosing some bandwidth there.
> their backplanes take 5 disks, so a potential bandwidth of well call it
> 500mbyte/second
> so 50% is out the window there actual bandwidth per 5 disks is going to be
> 250mbyte
>
> they have 9 of these channels for a total bandwith available of
> 2250mbyte/sec (2 gigabytes a second, that'll rip some dvds fast)
>
> standard PCI tops out at 133mbyte/sec so thats out ;->
>
> it looks like they are using PCI-E SATA cards
> the mbo they have and the cards they are using indicate they have 3x of
> something like this
> http://www.syba.com/index.php?controller=Product&action=Info&Id=861
> which maxes out at 250mbyte/sec per (1x PCI-E 1x lane)
> and one 4 port card which if it comes from that mob must be a PCI by the
> look of things.
> but I'll assume that its PCI-E and at least 4 lanes.
>
> so the total xfer rate is 1750mbyte/sec
> (or 883 if they are using the PCI card)
>
> Vs the total possible xfer rate of 4500
> they aren't doing *too* badly
>
> given that on a gigabit ethernet connection you are lucky to push
> 30mbyte/sec (or 60 if you tweak it), I think its not going to be a big issue
> ;->
>
>
> If they wanted more oomph their best bet would be to put 2x 16 port PCI-E
> 16x cards into a motherboard that supported it (most decent SLI motherboards
> will do that)
>
> better still one with 4x pci-E 16 slots so that you can put some 10gigE
> cards in as well something like
>
> http://www.aria.co.uk/Products/Components/Motherboards/Socket+AM3+(AMD)/MSI+790FX-GD70+AM3+Motherboard+with+4+x+PCIe+x+16+?productId=36604<http://www.aria.co.uk/Products/Components/Motherboards/Socket+AM3+%28AMD%29/MSI+790FX-GD70+AM3+Motherboard+with+4+x+PCIe+x+16+?productId=36604>
> say (but with intel of course ;->)
>
> That should net you (assuming you use dual port 10gig-E nics) an xfer rate
> out of the box of around 2400mbytes/sec
> almost fast enough to spy on teh entirez intarwebz!!
>
> --
> SLUG - Sydney Linux User's Group Mailing List - http://slug.org.au/
> Subscription info and FAQs: http://slug.org.au/faq/mailinglists.html
>
--
SLUG - Sydney Linux User's Group Mailing List - http://slug.org.au/
Subscription info and FAQs: http://slug.org.au/faq/mailinglists.html

Reply via email to