On Sun, May 13, 2007 at 05:23:53PM -0700, Matt Mahoney wrote: > It is not that hard, really. Each of the 10^5 PCs simulates about 10 mm^3 of
You know, repeating assertions doesn't make them any more true. > brain tissue. Axon diameter varies but is typically 1-2 microns. This means Where have you pulled that number from? Why not um^3, or m^3, or a cubic lightyear? > each bit of brain tissue has at most on the order of 10^7 inputs and outputs, The network is only well-balanced when you're doing spatial problem decomposition across few-GByte 3d arrays, and use GBit Ethernet to exchange vicinal interfaces. This assumes reasonably expensive (cheap ones can't handle the load) switches, and at least a 3d lattice topology (crossbars don't exist on that scale, but fortunately you also don't need them). At some 20 GByte/s memory throughput, you don't get a lot of ops on your grid points, as well. > each carrying 10 bits per second of information, or 100 Mb/s. This was barely > within Google's network capacity in 2000, and probably well within it now. > http://en.wikipedia.org/wiki/Google_platform > > I think individuals and small groups trying to build AGI will have a hard time > competing with Google due to the cost of hardware. It costs $2 million/month > just for electricity for their server farms. Google is building a > supercomputer in Oregon that will have cooling towers 4 stories high. > http://en.wikipedia.org/wiki/Project_02 I recommend you read up on commodity cluster supercomputing (Beowulf). Not every cluster is a supercomputer. -- Eugen* Leitl <a href="http://leitl.org">leitl</a> http://leitl.org ______________________________________________________________ ICBM: 48.07100, 11.36820 http://www.ativel.com http://postbiota.org 8B29F6BE: 099D 78BA 2FD3 B014 B08A 7779 75B0 2443 8B29 F6BE ----- This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=4007604&user_secret=8eb45b07