Re: [9fans] "FAWN: Fast array of wimpy nodes" (was: Plan 9 - the next 20 years)

2009-04-20 Thread Wes Kussmaul

ron minnich wrote:
RLX and Orion

multisystems showed there is not much of a market for lots of wimpy
nodes -- yet or never, is the real question. Either way, they did not
have enough buyers to stay in business. And RLX had to drop its wimpy
transmetas for P4s, and they could not keep up with the cheap
mainboards. It's a tough business.


All RLX showed was that they didn't know how to market benefits rather 
than nifty technology. Once again, the company with beautifully 
engineered products failed to understand that the decision makers who 
would buy the products were not engineers who loved them but people who 
needed education and handholding in order to understand them and 
overcome FUD from RLX competitors. A well-worn path.


Wes



Re: [9fans] "FAWN: Fast array of wimpy nodes" (was: Plan 9 - the next 20 years)

2009-04-20 Thread erik quanstrom
On Mon Apr 20 11:13:01 EDT 2009, jbar...@gmail.com wrote:
> > could you explain how raid 5 relates to sata vs sas?
> > i can't see now it's anything but a non-sequitor.
> 
> Here is the motivating real-world business case: You are in the movie
> post-production business and need > 50 TB of online storage at as low
> a price as possible with good performance and reliability.  7200 rpm
> SATA (currently ~15¢/GB on Newegg)

this example has nothing to do with raid.  if the object is to find the
lowest cost per gigabyte, enterprise sata drives are the cheeper option.
(it would make more sense to compare 7.2k sas and sata drives.  there
is also a premium on spindle speed.)

the original argument was that scsi is better than ata or sas is better
than sata (i'm not sure which); in my opinion, there are no facts
to justify either assertion.

> plus RAID narrows the performance
> and reliability of benefits of 15k rpm SAS (currently ~$1/GB) at a
> much lower cost.

without raid, such a configuration might be impossible to deal with.

most 15k drives are 73gb.  this means you would need 685 for 50tb.
the afr is probablly something like 0.15% - 0.25%.  this would mean you
will loose 1-2 drives/year.  (if you believe those rosy afr numbers.)

- erik



Re: [9fans] "FAWN: Fast array of wimpy nodes" (was: Plan 9 - the next 20 years)

2009-04-20 Thread ron minnich
On Sun, Apr 19, 2009 at 12:58 AM, John Barham  wrote:
> I certainly can't think ahead 20 years but I think it's safe to say
> that the next 5 (at least doing HPC and large-scale web type stuff)
> will increasingly look like this:
> http://www.technologyreview.com/computing/22504/?a=f, which talks
> about building a cluster from AMD Geode (!) nodes w/ compact flash
> storage.  Sure it's not super-fast, but it's very efficient per watt.
> If you had more cash you might substitute HE Opterons and SSD's but
> the principle is the same.

It's nice. We did that one a few years ago. Here is the 7 year old
version: http://eri.ca.sandia.gov/eri/howto.html

We've been doing these with the Geode stuff since about 2006. We are
certainly not the first. The RLX was doing what FAWN did about 8 years
ago; orion, about 3-4 years ago (both transmeta). RLX and Orion
multisystems showed there is not much of a market for lots of wimpy
nodes -- yet or never, is the real question. Either way, they did not
have enough buyers to stay in business. And RLX had to drop its wimpy
transmetas for P4s, and they could not keep up with the cheap
mainboards. It's a tough business.

ron



Re: [9fans] "FAWN: Fast array of wimpy nodes" (was: Plan 9 - the next 20 years)

2009-04-20 Thread John Barham
> could you explain how raid 5 relates to sata vs sas?
> i can't see now it's anything but a non-sequitor.

Here is the motivating real-world business case: You are in the movie
post-production business and need > 50 TB of online storage at as low
a price as possible with good performance and reliability.  7200 rpm
SATA (currently ~15¢/GB on Newegg) plus RAID narrows the performance
and reliability of benefits of 15k rpm SAS (currently ~$1/GB) at a
much lower cost.



Re: [9fans] "FAWN: Fast array of wimpy nodes" (was: Plan 9 - the next 20 years)

2009-04-19 Thread tlaronde
On Sun, Apr 19, 2009 at 09:27:43AM -0500, Eric Van Hensbergen wrote:
> 
> I'm not convinced that such ad-hoc DSM models are the way to go as a
> general principal.  Full blown DSM didn't fair very well in the past.
> Plan 9 distributed applications take a different approach and instead
> of sharing memory they share services in much more of a message
> passing model.  This isn't to say that all caches are bad -- I just
> don't believe in making them the foundation of your programing model
> as it will surely lead to trouble.
> 

FWIW, the more satisfying definition for me of a "computing unit" (an
"atom" OS based) is memory based: all the processing unit having direct
hardware access to a memory space/sharing the same directly hardware
accessible memory space. 

There seems to be 2 kinds of "NUMA" around there :

1) Cathedral model NUMA: a hierarchical association of memories, tightly
coupled but with different speeds (a lot of uniprocessor are NUMA these
days with cache1, cache2 and main memory). All directly known by the
cores.

2) Bazaar model NUMA, or software NUMA, or GPLNUMA: treating an
inorganized collection of storage as addressable memories since one can
always give a way to locate the ressource, including by URL, associating
high speed tightly connected memories with remote storage accessible via
IP packets sent by surface mail if the hardware drived whistle is heard
by the human writing the letters.

Curiously enough, I believe in 1. I don't believe in 2.
-- 
Thierry Laronde (Alceste) 
 http://www.kergis.com/
Key fingerprint = 0FF7 E906 FBAF FE95 FD89  250D 52B1 AE95 6006 F40C



Re: [9fans] "FAWN: Fast array of wimpy nodes" (was: Plan 9 - the next 20 years)

2009-04-19 Thread erik quanstrom
> To clarify, I meant that given X vs. Y, the cost benefits of X
> eventually overwhelm the initial technical benefits of Y.
> 
> With SATA vs. SCSI in particular, I wasn't so much thinking of command
> sets or physical connections but of providing cluster scale storage
> (i.e., 10's or 100's of TB) where it's fast enough and reliable enough
> but much cheaper to use commodity 7200 rpm SATA drives in RAID 5 than
> "server grade" 10k or 15k rpm SCSI or SAS drives.

this dated prejudice that scsi is for servers and ata is
for your dad's computer has just got to die.

could you explain how raid 5 relates to sata vs sas?
i can't see now it's anything but a non-sequitor.

you do realize that enterprise sata drives are available?
you do realize that many of said drives are built with the
same drive mechanism as sata hard drives?

as an example, the seagate es.2 drives are available with
a sas interface or a sata interface.

(by the way, enterprise drives are well worth it, as i discovered
on monday. :-(.)

while it's true there aren't any 15k sata drives currently
on the market, on the other hand if you want real performance,
you can beat sas by getting an intel ssd drive.  these are
not currently available in a sas.

- erik



Re: [9fans] "FAWN: Fast array of wimpy nodes" (was: Plan 9 - the next 20 years)

2009-04-19 Thread John Barham
>> Economics beats technology every time (e.g., x86/amd64 vs.
>> MIPS/Itanium, Ethernet vs. Infiniband, SATA vs. SCSI) so
>> don't try to fight it.
>
> if those examples prove your point, i'm not sure i agree.
>
> having just completed a combined-mode sata/sas driver,
> scsi vs ata is is fresh on my mind.  i'll use it as an example.

To clarify, I meant that given X vs. Y, the cost benefits of X
eventually overwhelm the initial technical benefits of Y.

With SATA vs. SCSI in particular, I wasn't so much thinking of command
sets or physical connections but of providing cluster scale storage
(i.e., 10's or 100's of TB) where it's fast enough and reliable enough
but much cheaper to use commodity 7200 rpm SATA drives in RAID 5 than
"server grade" 10k or 15k rpm SCSI or SAS drives.

  John



Re: [9fans] "FAWN: Fast array of wimpy nodes" (was: Plan 9 - the next 20 years)

2009-04-19 Thread Eric Van Hensbergen
On Sun, Apr 19, 2009 at 2:58 AM, John Barham  wrote:
> I certainly can't think ahead 20 years but I think it's safe to say
> that the next 5 (at least doing HPC and large-scale web type stuff)
> will increasingly look like this:
> http://www.technologyreview.com/computing/22504/?a=f, which talks
> about building a cluster from AMD Geode (!) nodes w/ compact flash
> storage.  Sure it's not super-fast, but it's very efficient per watt.
> If you had more cash you might substitute HE Opterons and SSD's but
> the principle is the same.
>

We thought this was the future several years ago
(http://bit.ly/16ZWjc), but couldn't convince the company that such an
approach would win out over big iron.  Of course, if you look at Blue
Gene, it's really just a massive realization of this model with
several really tightly coupled interconnects.

>
> Apparently they use the above cluster to implement some type of
> distributed memcached style cache.
>

I'm not convinced that such ad-hoc DSM models are the way to go as a
general principal.  Full blown DSM didn't fair very well in the past.
Plan 9 distributed applications take a different approach and instead
of sharing memory they share services in much more of a message
passing model.  This isn't to say that all caches are bad -- I just
don't believe in making them the foundation of your programing model
as it will surely lead to trouble.

-eric



Re: [9fans] "FAWN: Fast array of wimpy nodes" (was: Plan 9 - the next 20 years)

2009-04-19 Thread erik quanstrom
> I think the key to successfully being able to use Plan 9 commercially
> is to use its unique technical advantages to exploit disruptive
> economic changes.

works for coraid.

> Economics beats technology every time (e.g., x86/amd64 vs.
> MIPS/Itanium, Ethernet vs. Infiniband, SATA vs. SCSI) so
> don't try to fight it.

if those examples prove your point, i'm not sure i agree.

having just completed a combined-mode sata/sas driver,
scsi vs ata is is fresh on my mind.  i'll use it as an example.

sata and scsi can't be directly compared because sata is an
specific physical/data layer that supports the ata 7+ command set*,
while scsi is a set of command sets and an a set of physical
standards.

if you mean that the ata command set is not as good as the
scsi command set, i don't think i agree with this.  the ata
command set is simpler, and still gets the job done.  both
suffer from bad command formatting, but scsi is worse.  ata
has 28 bit and 48 bit (sic) commands.  scsi has 6, 10, 12, 16
and 32-byte commands).  one can find problems with both
command sets.

if you mean that sata is worse than sas, i think that's a hard
sell, too.  sata and sas use the same physical connections at
the same speeds.  there are some data-layer differences, but
they seem to me to be differences without distinction.  (as
evidenced by the aforementioned combined-mode hba.)

the main difference between sas and sata is that sas supports
dual-porting of drives so that if your hba fails, your drive can
keep working.  i don't see that as a real killer feature.  hard
drives fail so much more often than hbas.

- erik