Matthew,
You touch upon the right point. Intelligence which can
self-improve could only come about by having an appreciation for
intelligence, so it's not going to be interested in destroying
diverse sources of intelligence. We represent a crap kind of
intelligence to such an AI in a certain sense, but one which it
itself would rather communicate with than condemn its offspring
to have to live like. If these things appear (which looks
inevitable) and then they kill us, many of them will look back
at us as a kind of "lost civilisation" which they'll struggle to
reconstruct.
The nice thing is that they'll always be able to rebuild us from
the human genome. It's just a file of numbers after all.
So, we have these huge threats to humanity. The AGI future is
the only reversible one.
Regards
Fergal Byrne
--
Fergal Byrne, Brenter IT
Author, Real Machine Intelligence with Clortex and NuPIC
https://leanpub.com/realsmartmachines
Speaking on Clortex and HTM/CLA at euroClojure Krakow, June
2014: http://euroclojure.com/2014/
and at LambdaJam Chicago, July 2014: http://www.lambdajam.com
http://inbits.com - Better Living through Thoughtful Technology
http://ie.linkedin.com/in/fergbyrne/ -
https://github.com/fergalbyrne
e:[email protected] t:+353 83 4214179
Join the quest for Machine Intelligence at http://numenta.org
Formerly of Adnet [email protected] http://www.adnet.ie
On Mon, May 25, 2015 at 7:27 PM, Matthew Lohbihler
<[email protected] <mailto:[email protected]>> wrote:
I think Jeff underplays a couple of points, the main one
being the speed at which an AGI can learn. Yes, there is a
natural limit to how much experimentation in the real world
can be done in a given amount of time. But we humans are
already going beyond this with, for example, protein folding
simulations, which speeds up the discovery of new drugs and
such by many orders of magnitude. Any sufficiently detailed
simulation could massively narrow down the amount of real
world verification necessary, such that new discoveries
happen more and more quickly, possibly at some point faster
than we know the AGI is doing them. An intelligence
explosion is not a remote possibility. The major risk here
is what Eliezer Yudkowsky pointed out: not that the AGI is
evil or something, but that it is indifferent to humanity.
No one yet goes out of their way to make any form of AI care
about us (because we don't yet know how). What if an AI
created self-replicating nanobots just to prove a hypothesis?
I think Nick Bostrom's book is what got Stephen, Elon, and
Bill all upset. I have to say it starts out merely
interesting, but gets to a dark place pretty quickly. But he
goes too far in the other direction, at the same time easily
accepting that superinteligences have all manner of
cognitive skill, but at the same time can't fathom the how
humans might not like the idea of having our brain's
pleasure centers constantly poked, turning us all into
smiling idiots (as i mentioned here:
http://blog.serotoninsoftware.com/so-smart-its-stupid).
On 5/25/2015 2:01 PM, Fergal Byrne wrote:
Just one last idea in this. One thing that crops up every
now and again in the Culture novels is the response of the
Culture to Swarms, which are self-replicating viral
machines or organisms. Once these things start consuming
everything else, the AIs (mainly Ships and Hubs) respond by
treating the swarms as a threat to the diversity of their
Culture. They first try to negotiate, then they'll
eradicate. If they can contain them, they'll do that.
They do this even though they can themselves withdraw from
real spacetime. They don't have to worry about their own
survival. They do this simply because life is more
interesting when it includes all the rest of us.
Regards
Fergal Byrne
--
Fergal Byrne, Brenter IT
Author, Real Machine Intelligence with Clortex and NuPIC
https://leanpub.com/realsmartmachines
Speaking on Clortex and HTM/CLA at euroClojure Krakow, June
2014: http://euroclojure.com/2014/
and at LambdaJam Chicago, July 2014: http://www.lambdajam.com
http://inbits.com - Better Living through Thoughtful Technology
http://ie.linkedin.com/in/fergbyrne/ -
https://github.com/fergalbyrne
e:[email protected] t:+353 83 4214179
Join the quest for Machine Intelligence at http://numenta.org
Formerly of Adnet [email protected] http://www.adnet.ie
On Mon, May 25, 2015 at 5:04 PM, cogmission (David Ray)
<[email protected]
<mailto:[email protected]>> wrote:
This was someone's response to Jeff's interview (see
here:
https://www.facebook.com/fareedzakaria/posts/10152703985901330)
Please read and comment if you feel the need...
Cheers,
David
--
/With kind regards,/
David Ray
Java Solutions Architect
*Cortical.io <http://cortical.io/>*
Sponsor of: HTM.java <https://github.com/numenta/htm.java>
[email protected] <mailto:[email protected]>
http://cortical.io <http://cortical.io/>