True, but it's more about the combination of power with intelligence. The 
entities with the most power are likely to be those with the most intelligence. 
We hope.



--

Fergal Byrne, Brenter IT

Author, Real Machine Intelligence with Clortex and NuPIC 
https://leanpub.com/realsmartmachines

Speaking on Clortex and HTM/CLA at euroClojure Krakow, June 2014: 
http://euroclojure.com/2014/
and at LambdaJam Chicago, July 2014: http://www.lambdajam.com

http://inbits.com - Better Living through Thoughtful Technology
http://ie.linkedin.com/in/fergbyrne/ - https://github.com/fergalbyrne

e:[email protected] t:+353 83 4214179
Join the quest for Machine Intelligence at http://numenta.org
Formerly of Adnet [email protected] http://www.adnet.ie

On Mon, May 25, 2015 at 7:49 PM, Matthew Lohbihler
<[email protected]> wrote:

> Good points Fergal. But do remember that you are assigning 
> sentimentality to the AGI, while claiming - like Jeff - that its way of 
> thinking will not be like ours. It could just as easily logically decide 
> that the lost civilization was non-optimal anyway, so no harm done.
> On 5/25/2015 2:40 PM, Fergal Byrne wrote:
>> Matthew,
>>
>> You touch upon the right point. Intelligence which can self-improve 
>> could only come about by having an appreciation for intelligence, so 
>> it's not going to be interested in destroying diverse sources of 
>> intelligence. We represent a crap kind of intelligence to such an AI 
>> in a certain sense, but one which it itself would rather communicate 
>> with than condemn its offspring to have to live like. If these things 
>> appear (which looks inevitable) and then they kill us, many of them 
>> will look back at us as a kind of "lost civilisation" which they'll 
>> struggle to reconstruct.
>>
>> The nice thing is that they'll always be able to rebuild us from the 
>> human genome. It's just a file of numbers after all.
>>
>> So, we have these huge threats to humanity. The AGI future is the only 
>> reversible one.
>>
>> Regards
>> Fergal Byrne
>>
>> --
>>
>> Fergal Byrne, Brenter IT
>>
>> Author, Real Machine Intelligence with Clortex and NuPIC
>> https://leanpub.com/realsmartmachines
>>
>> Speaking on Clortex and HTM/CLA at euroClojure Krakow, June 2014: 
>> http://euroclojure.com/2014/
>> and at LambdaJam Chicago, July 2014: http://www.lambdajam.com
>>
>> http://inbits.com - Better Living through Thoughtful Technology
>> http://ie.linkedin.com/in/fergbyrne/ - https://github.com/fergalbyrne
>>
>> e:[email protected] t:+353 83 4214179
>> Join the quest for Machine Intelligence at http://numenta.org
>> Formerly of Adnet [email protected] http://www.adnet.ie
>>
>>
>> On Mon, May 25, 2015 at 7:27 PM, Matthew Lohbihler 
>> <[email protected] <mailto:[email protected]>> wrote:
>>
>>     I think Jeff underplays a couple of points, the main one being the
>>     speed at which an AGI can learn. Yes, there is a natural limit to
>>     how much experimentation in the real world can be done in a given
>>     amount of time. But we humans are already going beyond this with,
>>     for example, protein folding simulations, which speeds up the
>>     discovery of new drugs and such by many orders of magnitude. Any
>>     sufficiently detailed simulation could massively narrow down the
>>     amount of real world verification necessary, such that new
>>     discoveries happen more and more quickly, possibly at some point
>>     faster than we know the AGI is doing them. An intelligence
>>     explosion is not a remote possibility. The major risk here is what
>>     Eliezer Yudkowsky pointed out: not that the AGI is evil or
>>     something, but that it is indifferent to humanity. No one yet goes
>>     out of their way to make any form of AI care about us (because we
>>     don't yet know how). What if an AI created self-replicating
>>     nanobots just to prove a hypothesis?
>>
>>     I think Nick Bostrom's book is what got Stephen, Elon, and Bill
>>     all upset. I have to say it starts out merely interesting, but
>>     gets to a dark place pretty quickly. But he goes too far in the
>>     other direction, at the same time easily accepting that
>>     superinteligences have all manner of cognitive skill, but at the
>>     same time can't fathom the how humans might not like the idea of
>>     having our brain's pleasure centers constantly poked, turning us
>>     all into smiling idiots (as i mentioned here:
>>     http://blog.serotoninsoftware.com/so-smart-its-stupid).
>>
>>
>>
>>     On 5/25/2015 2:01 PM, Fergal Byrne wrote:
>>>     Just one last idea in this. One thing that crops up every now and
>>>     again in the Culture novels is the response of the Culture to
>>>     Swarms, which are self-replicating viral machines or organisms.
>>>     Once these things start consuming everything else, the AIs
>>>     (mainly Ships and Hubs) respond by treating the swarms as a
>>>     threat to the diversity of their Culture. They first try to
>>>     negotiate, then they'll eradicate. If they can contain them,
>>>     they'll do that.
>>>
>>>     They do this even though they can themselves withdraw from real
>>>     spacetime. They don't have to worry about their own survival.
>>>     They do this simply because life is more interesting when it
>>>     includes all the rest of us.
>>>
>>>     Regards
>>>
>>>     Fergal Byrne
>>>
>>>     --
>>>
>>>     Fergal Byrne, Brenter IT
>>>
>>>     Author, Real Machine Intelligence with Clortex and NuPIC
>>>     https://leanpub.com/realsmartmachines
>>>
>>>     Speaking on Clortex and HTM/CLA at euroClojure Krakow, June 2014:
>>>     http://euroclojure.com/2014/
>>>     and at LambdaJam Chicago, July 2014: http://www.lambdajam.com
>>>
>>>     http://inbits.com - Better Living through Thoughtful Technology
>>>     http://ie.linkedin.com/in/fergbyrne/ - https://github.com/fergalbyrne
>>>
>>>     e:[email protected] t:+353 83 4214179
>>>     Join the quest for Machine Intelligence at http://numenta.org
>>>     Formerly of Adnet [email protected] http://www.adnet.ie
>>>
>>>
>>>     On Mon, May 25, 2015 at 5:04 PM, cogmission (David Ray)
>>>     <[email protected] <mailto:[email protected]>>
>>>     wrote:
>>>
>>>         This was someone's response to Jeff's interview (see here:
>>>         https://www.facebook.com/fareedzakaria/posts/10152703985901330)
>>>
>>>         Please read and comment if you feel the need...
>>>
>>>         Cheers,
>>>         David
>>>
>>>         -- 
>>>         /With kind regards,/
>>>         David Ray
>>>         Java Solutions Architect
>>>         *Cortical.io <http://cortical.io/>*
>>>         Sponsor of: HTM.java <https://github.com/numenta/htm.java>
>>>         [email protected] <mailto:[email protected]>
>>>         http://cortical.io <http://cortical.io/>
>>>
>>>
>>
>>

Reply via email to