Matthew,
You touch upon the right point. Intelligence which can self-improve
could only come about by having an appreciation for intelligence, so
it's not going to be interested in destroying diverse sources of
intelligence. We represent a crap kind of intelligence to such an AI
in a certain sense, but one which it itself would rather communicate
with than condemn its offspring to have to live like. If these
things appear (which looks inevitable) and then they kill us, many
of them will look back at us as a kind of "lost civilisation" which
they'll struggle to reconstruct.
The nice thing is that they'll always be able to rebuild us from the
human genome. It's just a file of numbers after all.
So, we have these huge threats to humanity. The AGI future is the
only reversible one.
Regards
Fergal Byrne
--
Fergal Byrne, Brenter IT
Author, Real Machine Intelligence with Clortex and NuPIC
https://leanpub.com/realsmartmachines
Speaking on Clortex and HTM/CLA at euroClojure Krakow, June 2014:
http://euroclojure.com/2014/
and at LambdaJam Chicago, July 2014: http://www.lambdajam.com
http://inbits.com - Better Living through Thoughtful Technology
http://ie.linkedin.com/in/fergbyrne/ - https://github.com/fergalbyrne
e:[email protected] t:+353 83 4214179
Join the quest for Machine Intelligence at http://numenta.org
Formerly of Adnet [email protected] http://www.adnet.ie
On Mon, May 25, 2015 at 7:27 PM, Matthew Lohbihler
<[email protected] <mailto:[email protected]>> wrote:
I think Jeff underplays a couple of points, the main one being
the speed at which an AGI can learn. Yes, there is a natural
limit to how much experimentation in the real world can be done
in a given amount of time. But we humans are already going
beyond this with, for example, protein folding simulations,
which speeds up the discovery of new drugs and such by many
orders of magnitude. Any sufficiently detailed simulation could
massively narrow down the amount of real world verification
necessary, such that new discoveries happen more and more
quickly, possibly at some point faster than we know the AGI is
doing them. An intelligence explosion is not a remote
possibility. The major risk here is what Eliezer Yudkowsky
pointed out: not that the AGI is evil or something, but that it
is indifferent to humanity. No one yet goes out of their way to
make any form of AI care about us (because we don't yet know
how). What if an AI created self-replicating nanobots just to
prove a hypothesis?
I think Nick Bostrom's book is what got Stephen, Elon, and Bill
all upset. I have to say it starts out merely interesting, but
gets to a dark place pretty quickly. But he goes too far in the
other direction, at the same time easily accepting that
superinteligences have all manner of cognitive skill, but at the
same time can't fathom the how humans might not like the idea of
having our brain's pleasure centers constantly poked, turning us
all into smiling idiots (as i mentioned here:
http://blog.serotoninsoftware.com/so-smart-its-stupid).
On 5/25/2015 2:01 PM, Fergal Byrne wrote:
Just one last idea in this. One thing that crops up every now
and again in the Culture novels is the response of the Culture
to Swarms, which are self-replicating viral machines or
organisms. Once these things start consuming everything else,
the AIs (mainly Ships and Hubs) respond by treating the swarms
as a threat to the diversity of their Culture. They first try
to negotiate, then they'll eradicate. If they can contain them,
they'll do that.
They do this even though they can themselves withdraw from real
spacetime. They don't have to worry about their own survival.
They do this simply because life is more interesting when it
includes all the rest of us.
Regards
Fergal Byrne
--
Fergal Byrne, Brenter IT
Author, Real Machine Intelligence with Clortex and NuPIC
https://leanpub.com/realsmartmachines
Speaking on Clortex and HTM/CLA at euroClojure Krakow, June
2014: http://euroclojure.com/2014/
and at LambdaJam Chicago, July 2014: http://www.lambdajam.com
http://inbits.com - Better Living through Thoughtful Technology
http://ie.linkedin.com/in/fergbyrne/ -
https://github.com/fergalbyrne
e:[email protected] t:+353 83 4214179
Join the quest for Machine Intelligence at http://numenta.org
Formerly of Adnet [email protected] http://www.adnet.ie
On Mon, May 25, 2015 at 5:04 PM, cogmission (David Ray)
<[email protected]
<mailto:[email protected]>> wrote:
This was someone's response to Jeff's interview (see here:
https://www.facebook.com/fareedzakaria/posts/10152703985901330)
Please read and comment if you feel the need...
Cheers,
David
--
/With kind regards,/
David Ray
Java Solutions Architect
*Cortical.io <http://cortical.io/>*
Sponsor of: HTM.java <https://github.com/numenta/htm.java>
[email protected] <mailto:[email protected]>
http://cortical.io <http://cortical.io/>