I think Jeff underplays a couple of points, the main one being the speed
at which an AGI can learn. Yes, there is a natural limit to how much
experimentation in the real world can be done in a given amount of time.
But we humans are already going beyond this with, for example, protein
folding simulations, which speeds up the discovery of new drugs and such
by many orders of magnitude. Any sufficiently detailed simulation could
massively narrow down the amount of real world verification necessary,
such that new discoveries happen more and more quickly, possibly at some
point faster than we know the AGI is doing them. An intelligence
explosion is not a remote possibility. The major risk here is what
Eliezer Yudkowsky pointed out: not that the AGI is evil or something,
but that it is indifferent to humanity. No one yet goes out of their way
to make any form of AI care about us (because we don't yet know how).
What if an AI created self-replicating nanobots just to prove a hypothesis?
I think Nick Bostrom's book is what got Stephen, Elon, and Bill all
upset. I have to say it starts out merely interesting, but gets to a
dark place pretty quickly. But he goes too far in the other direction,
at the same time easily accepting that superinteligences have all manner
of cognitive skill, but at the same time can't fathom the how humans
might not like the idea of having our brain's pleasure centers
constantly poked, turning us all into smiling idiots (as i mentioned
here: http://blog.serotoninsoftware.com/so-smart-its-stupid).
On 5/25/2015 2:01 PM, Fergal Byrne wrote:
Just one last idea in this. One thing that crops up every now and
again in the Culture novels is the response of the Culture to Swarms,
which are self-replicating viral machines or organisms. Once these
things start consuming everything else, the AIs (mainly Ships and
Hubs) respond by treating the swarms as a threat to the diversity of
their Culture. They first try to negotiate, then they'll eradicate. If
they can contain them, they'll do that.
They do this even though they can themselves withdraw from real
spacetime. They don't have to worry about their own survival. They do
this simply because life is more interesting when it includes all the
rest of us.
Regards
Fergal Byrne
--
Fergal Byrne, Brenter IT
Author, Real Machine Intelligence with Clortex and NuPIC
https://leanpub.com/realsmartmachines
Speaking on Clortex and HTM/CLA at euroClojure Krakow, June 2014:
http://euroclojure.com/2014/
and at LambdaJam Chicago, July 2014: http://www.lambdajam.com
http://inbits.com - Better Living through Thoughtful Technology
http://ie.linkedin.com/in/fergbyrne/ - https://github.com/fergalbyrne
e:[email protected] t:+353 83 4214179
Join the quest for Machine Intelligence at http://numenta.org
Formerly of Adnet [email protected] http://www.adnet.ie
On Mon, May 25, 2015 at 5:04 PM, cogmission (David Ray)
<[email protected] <mailto:[email protected]>> wrote:
This was someone's response to Jeff's interview (see here:
https://www.facebook.com/fareedzakaria/posts/10152703985901330)
Please read and comment if you feel the need...
Cheers,
David
--
/With kind regards,/
David Ray
Java Solutions Architect
*Cortical.io <http://cortical.io/>*
Sponsor of: HTM.java <https://github.com/numenta/htm.java>
[email protected] <mailto:[email protected]>
http://cortical.io <http://cortical.io/>