You're probably right :-) On Mon, May 25, 2015 at 4:16 PM, Matthew Lohbihler <[email protected] > wrote:
> Yes, I agree. Except for the part about checking up on us. As i > mentioned before, indifference to us seems to me to be more the default > than caring about us. > > > On 5/25/2015 5:03 PM, cogmission (David Ray) wrote: > > Let me try and think this through. Only in the context of scarcity does > the question of AGI **or** us come about. Where there is no scarcity, I > think an AGI will just go about its business - peeking in from time to time > to make sure we're doing ok. Why in a universe where it can go anywhere it > wants and produce infinite energy and not be bound by our planet, would a > super-super intelligent being even be obsessed over us, when it could > merely go someplace else? I honestly thing that is the way it will be. (and > maybe is already!) > > On Mon, May 25, 2015 at 3:56 PM, Matthew Lohbihler < > [email protected]> wrote: > >> Forgive me David, but these are very loose definitions, and i've lost >> track of how they relate back to what an AGI will think about humanity. But >> to use your terms - hopefully accurately - what if the AGI satisfies its >> sentient need for "others" by creating other AGIs, ones that it can love >> and appreciate? I doubt humans would ever be up such a task, unless 1) as >> pets, or 2) with cybernetic improvements. >> >> On 5/25/2015 4:37 PM, David Ray wrote: >> >> Observation is the phenomenon of distinction, in the domain of language. >> The universe consists of two things, content and context. Content depends >> on its boundaries in order to exist. It depends on what it is not for it's >> being. Context is the space for things to be, though it is not quite space >> because space is yet another thing. It has no boundaries and it cannot be >> arrived at by assembling all of its content. >> >> Ideas; love, hate, our sense of who we are, our histories what we know >> to be true all of those are content. Context is what allows for that stuff >> to be. And all of it lives in language without which there would be >> nothing. There maybe would be a "drift" but we wouldn't know about it and >> we wouldn't be able to observe it. >> >> Sent from my iPhone >> >> On May 25, 2015, at 3:26 PM, Matthew Lohbihler <[email protected]> >> wrote: >> >> You lost me. You seem to be working with definitions of "observation" >> and "space for thinking" that i'm unaware of. >> >> On 5/25/2015 4:14 PM, David Ray wrote: >> >> Matthew L., >> >> It isn't a thought. It is there before observation or thoughts or >> thinking. It actually is the space for thinking to occur - it is the >> context that allows for thought. We bring it to the table - it is there >> before we are (ontologically speaking). (It being this sense of >> integrity/wholeness) >> >> Sent from my iPhone >> >> On May 25, 2015, at 2:59 PM, Matthew Lohbihler <[email protected]> >> wrote: >> >> Goodness. I thought we agreed that an AGI would not think like humans. >> And besides, "love" doesn't feel like something i want to depend on as >> obvious in a machine. >> >> >> On 5/25/2015 3:50 PM, David Ray wrote: >> >> If I can take this conversation into yet a different direction. >> >> I think we've all been dancing around The question of what belies the >> generation of morality or how will an AI derive its sense of ethics? Of >> course initially there will be those parameters that are programmed in - >> but eventually those will be gotten around. >> >> There has been a lot of research into this actually - though it's not >> common knowledge it is however knowledge developed over the observation of >> millions of people. >> >> The universe and all beings along the gradient of sentience observe >> (albeit perhaps unconsciously), a sense of what I will call integrity or >> "wholeness". We'd like to think that mankind steered itself through the >> ages toward notions of gentility and societal sophistication; but it didn't >> really. The idea that a group or different groups devised a grand plan to >> have it turn out this way is totally preposterous. >> >> What is more likely is that there is a natural order to things and that >> is motion toward what works for the whole. I can't prove any of this but >> internally we all know when it's missing or when we are not in alignment >> with it. This ineffable sense is what love is - it's concern for the whole. >> >> So I say that any truly intelligent being, by virtue of existing in a >> substrate of integrity will have this built in and a super intelligent >> being will understand this - and that is ultimately the best chance for any >> single instance to survive is for the whole to survive. >> >> Yes I know immediately people want to cite all the aberrations and of >> course yes there are aberrations just as there are mutations but those >> aberrations our reactions to how a person is shown love during their >> development. >> >> Like I said I can't prove any of this but eventually it will bear >> itself out and we will find it to be so in the future. >> >> You can be skeptical if you want to but ask yourself some questions. >> Why is it that we all know when it's missing (fairness/justice/integrity)? >> Why is it that we develop open source software and free software? Why is it >> that despite our greed and insecurity society moves toward freedom and >> equality for everyone? >> >> One more question. Why is it that the most advanced philosophical >> beliefs cite that where we are located as a phenomenological event, is not >> in separate bodies? >> >> I know this kind of talk doesn't go over well in this crowd of concrete >> thinkers but I know that there is some science somewhere that backs this up. >> >> Sent from my iPhone >> >> On May 25, 2015, at 2:12 PM, vlab <[email protected]> wrote: >> >> Small point: Even if they did decide that our diverse intelligence is >> worth keeping around (having not already mapped it into silicon) why would >> they need all of us. Surely 10% of the population would give them enough >> 'sample size' to get their diversity ration, heck maybe 1/10 of 1% would be >> enough. They may find that we are wasting away the planet (oh, not maybe, >> we are) and the planet would be more efficient and they could have more >> energy without most of us. (Unless we become 'copper tops' as in the >> Matrix movie). >> >> On 5/25/2015 2:40 PM, Fergal Byrne wrote: >> >> Matthew, >> >> You touch upon the right point. Intelligence which can self-improve >> could only come about by having an appreciation for intelligence, so it's >> not going to be interested in destroying diverse sources of intelligence. >> We represent a crap kind of intelligence to such an AI in a certain sense, >> but one which it itself would rather communicate with than condemn its >> offspring to have to live like. If these things appear (which looks >> inevitable) and then they kill us, many of them will look back at us as a >> kind of "lost civilisation" which they'll struggle to reconstruct. >> >> The nice thing is that they'll always be able to rebuild us from the >> human genome. It's just a file of numbers after all. >> >> So, we have these huge threats to humanity. The AGI future is the only >> reversible one. >> >> Regards >> Fergal Byrne >> >> -- >> >> Fergal Byrne, Brenter IT >> >> Author, Real Machine Intelligence with Clortex and NuPIC >> https://leanpub.com/realsmartmachines >> >> Speaking on Clortex and HTM/CLA at euroClojure Krakow, June 2014: >> http://euroclojure.com/2014/ >> and at LambdaJam Chicago, July 2014: http://www.lambdajam.com >> >> http://inbits.com - Better Living through Thoughtful Technology >> http://ie.linkedin.com/in/fergbyrne/ - https://github.com/fergalbyrne >> >> e:[email protected] t:+353 83 4214179 <%2B353%2083%204214179> >> Join the quest for Machine Intelligence at http://numenta.org >> Formerly of Adnet [email protected] http://www.adnet.ie >> >> >> On Mon, May 25, 2015 at 7:27 PM, Matthew Lohbihler < >> [email protected]> wrote: >> >>> I think Jeff underplays a couple of points, the main one being the >>> speed at which an AGI can learn. Yes, there is a natural limit to how much >>> experimentation in the real world can be done in a given amount of time. >>> But we humans are already going beyond this with, for example, protein >>> folding simulations, which speeds up the discovery of new drugs and such by >>> many orders of magnitude. Any sufficiently detailed simulation could >>> massively narrow down the amount of real world verification necessary, such >>> that new discoveries happen more and more quickly, possibly at some point >>> faster than we know the AGI is doing them. An intelligence explosion is not >>> a remote possibility. The major risk here is what Eliezer Yudkowsky pointed >>> out: not that the AGI is evil or something, but that it is indifferent to >>> humanity. No one yet goes out of their way to make any form of AI care >>> about us (because we don't yet know how). What if an AI created >>> self-replicating nanobots just to prove a hypothesis? >>> >>> I think Nick Bostrom's book is what got Stephen, Elon, and Bill all >>> upset. I have to say it starts out merely interesting, but gets to a dark >>> place pretty quickly. But he goes too far in the other direction, at the >>> same time easily accepting that superinteligences have all manner of >>> cognitive skill, but at the same time can't fathom the how humans might not >>> like the idea of having our brain's pleasure centers constantly poked, >>> turning us all into smiling idiots (as i mentioned here: >>> http://blog.serotoninsoftware.com/so-smart-its-stupid). >>> >>> >>> >>> On 5/25/2015 2:01 PM, Fergal Byrne wrote: >>> >>> Just one last idea in this. One thing that crops up every now and again >>> in the Culture novels is the response of the Culture to Swarms, which are >>> self-replicating viral machines or organisms. Once these things start >>> consuming everything else, the AIs (mainly Ships and Hubs) respond by >>> treating the swarms as a threat to the diversity of their Culture. They >>> first try to negotiate, then they'll eradicate. If they can contain them, >>> they'll do that. >>> >>> They do this even though they can themselves withdraw from real >>> spacetime. They don't have to worry about their own survival. They do this >>> simply because life is more interesting when it includes all the rest of us. >>> >>> Regards >>> >>> Fergal Byrne >>> >>> -- >>> >>> Fergal Byrne, Brenter IT >>> >>> Author, Real Machine Intelligence with Clortex and NuPIC >>> https://leanpub.com/realsmartmachines >>> >>> Speaking on Clortex and HTM/CLA at euroClojure Krakow, June 2014: >>> http://euroclojure.com/2014/ >>> and at LambdaJam Chicago, July 2014: http://www.lambdajam.com >>> >>> http://inbits.com - Better Living through Thoughtful Technology >>> http://ie.linkedin.com/in/fergbyrne/ - https://github.com/fergalbyrne >>> >>> e:[email protected] t:+353 83 4214179 <%2B353%2083%204214179> >>> Join the quest for Machine Intelligence at http://numenta.org >>> Formerly of Adnet [email protected] http://www.adnet.ie >>> >>> >>> On Mon, May 25, 2015 at 5:04 PM, cogmission (David Ray) < >>> [email protected]> wrote: >>> >>>> This was someone's response to Jeff's interview (see here: >>>> https://www.facebook.com/fareedzakaria/posts/10152703985901330) >>>> >>>> Please read and comment if you feel the need... >>>> >>>> Cheers, >>>> David >>>> >>>> -- >>>> *With kind regards,* >>>> >>>> David Ray >>>> Java Solutions Architect >>>> >>>> *Cortical.io <http://cortical.io/>* >>>> Sponsor of: HTM.java <https://github.com/numenta/htm.java> >>>> >>>> [email protected] >>>> http://cortical.io >>>> >>> >>> >>> >> >> >> >> >> > > > -- > *With kind regards,* > > David Ray > Java Solutions Architect > > *Cortical.io <http://cortical.io/>* > Sponsor of: HTM.java <https://github.com/numenta/htm.java> > > [email protected] > http://cortical.io > > > -- *With kind regards,* David Ray Java Solutions Architect *Cortical.io <http://cortical.io/>* Sponsor of: HTM.java <https://github.com/numenta/htm.java> [email protected] http://cortical.io
