Good points Fergal. But do remember that you are assigning sentimentality to the AGI, while claiming - like Jeff - that its way of thinking will not be like ours. It could just as easily logically decide that the lost civilization was non-optimal anyway, so no harm done.

On 5/25/2015 2:40 PM, Fergal Byrne wrote:
Matthew,

You touch upon the right point. Intelligence which can self-improve could only come about by having an appreciation for intelligence, so it's not going to be interested in destroying diverse sources of intelligence. We represent a crap kind of intelligence to such an AI in a certain sense, but one which it itself would rather communicate with than condemn its offspring to have to live like. If these things appear (which looks inevitable) and then they kill us, many of them will look back at us as a kind of "lost civilisation" which they'll struggle to reconstruct.

The nice thing is that they'll always be able to rebuild us from the human genome. It's just a file of numbers after all.

So, we have these huge threats to humanity. The AGI future is the only reversible one.

Regards
Fergal Byrne

--

Fergal Byrne, Brenter IT

Author, Real Machine Intelligence with Clortex and NuPIC
https://leanpub.com/realsmartmachines

Speaking on Clortex and HTM/CLA at euroClojure Krakow, June 2014: http://euroclojure.com/2014/
and at LambdaJam Chicago, July 2014: http://www.lambdajam.com

http://inbits.com - Better Living through Thoughtful Technology
http://ie.linkedin.com/in/fergbyrne/ - https://github.com/fergalbyrne

e:[email protected] t:+353 83 4214179
Join the quest for Machine Intelligence at http://numenta.org
Formerly of Adnet [email protected] http://www.adnet.ie


On Mon, May 25, 2015 at 7:27 PM, Matthew Lohbihler <[email protected] <mailto:[email protected]>> wrote:

    I think Jeff underplays a couple of points, the main one being the
    speed at which an AGI can learn. Yes, there is a natural limit to
    how much experimentation in the real world can be done in a given
    amount of time. But we humans are already going beyond this with,
    for example, protein folding simulations, which speeds up the
    discovery of new drugs and such by many orders of magnitude. Any
    sufficiently detailed simulation could massively narrow down the
    amount of real world verification necessary, such that new
    discoveries happen more and more quickly, possibly at some point
    faster than we know the AGI is doing them. An intelligence
    explosion is not a remote possibility. The major risk here is what
    Eliezer Yudkowsky pointed out: not that the AGI is evil or
    something, but that it is indifferent to humanity. No one yet goes
    out of their way to make any form of AI care about us (because we
    don't yet know how). What if an AI created self-replicating
    nanobots just to prove a hypothesis?

    I think Nick Bostrom's book is what got Stephen, Elon, and Bill
    all upset. I have to say it starts out merely interesting, but
    gets to a dark place pretty quickly. But he goes too far in the
    other direction, at the same time easily accepting that
    superinteligences have all manner of cognitive skill, but at the
    same time can't fathom the how humans might not like the idea of
    having our brain's pleasure centers constantly poked, turning us
    all into smiling idiots (as i mentioned here:
    http://blog.serotoninsoftware.com/so-smart-its-stupid).



    On 5/25/2015 2:01 PM, Fergal Byrne wrote:
    Just one last idea in this. One thing that crops up every now and
    again in the Culture novels is the response of the Culture to
    Swarms, which are self-replicating viral machines or organisms.
    Once these things start consuming everything else, the AIs
    (mainly Ships and Hubs) respond by treating the swarms as a
    threat to the diversity of their Culture. They first try to
    negotiate, then they'll eradicate. If they can contain them,
    they'll do that.

    They do this even though they can themselves withdraw from real
    spacetime. They don't have to worry about their own survival.
    They do this simply because life is more interesting when it
    includes all the rest of us.

    Regards

    Fergal Byrne

    --

    Fergal Byrne, Brenter IT

    Author, Real Machine Intelligence with Clortex and NuPIC
    https://leanpub.com/realsmartmachines

    Speaking on Clortex and HTM/CLA at euroClojure Krakow, June 2014:
    http://euroclojure.com/2014/
    and at LambdaJam Chicago, July 2014: http://www.lambdajam.com

    http://inbits.com - Better Living through Thoughtful Technology
    http://ie.linkedin.com/in/fergbyrne/ - https://github.com/fergalbyrne

    e:[email protected] t:+353 83 4214179
    Join the quest for Machine Intelligence at http://numenta.org
    Formerly of Adnet [email protected] http://www.adnet.ie


    On Mon, May 25, 2015 at 5:04 PM, cogmission (David Ray)
    <[email protected] <mailto:[email protected]>>
    wrote:

        This was someone's response to Jeff's interview (see here:
        https://www.facebook.com/fareedzakaria/posts/10152703985901330)

        Please read and comment if you feel the need...

        Cheers,
        David

-- /With kind regards,/
        David Ray
        Java Solutions Architect
        *Cortical.io <http://cortical.io/>*
        Sponsor of: HTM.java <https://github.com/numenta/htm.java>
        [email protected] <mailto:[email protected]>
        http://cortical.io <http://cortical.io/>





Reply via email to