No Richard, I meant...."Full Frontal Lobotomy", seeing as that has not been done before, ha, ha ! ...''Either that or this is a great title for a movie about a lab full of scientists trapped in a nudist colony'' God I shudder to think of the thought....mind you your AI's might have a thing or two to say about that....ha, ha. Right I've read what you've said about...Two completely separate mechanisms. I'm going to spend what little time I have this weekend reading Stefan's Jame5. I've started the first couple of pages and it reminds me of an Anime movie, I love the classics, 'Ghost in the shell 2' especially reminds me of the subject matter in this forum. Anyhow once I've digested that I may better understand yours and Stefan's side of the argument. Candice > Date: Fri, 26 Oct 2007 10:21:47 -0400> From: [EMAIL PROTECTED]> To: > singularity@v2.listbox.com> Subject: Re: [singularity] John Searle...> > > candice schuster wrote:> > Richard,> > > > Your responses to me seem to go in > round abouts. No insult intended > > however.> > > > You say the AI will in > fact reach full consciousness. How on earth > > would that ever be possible ? > > > I think I recently (last week or so) wrote out a reply to someone on the > > question of what a good explanation of "consciousness" might be (was it > > on this list?). I was implictly referring to that explanation of > > consciousness. It makes the definite prediction that consciousness > > (subjective awareness, qualia, etc. .... what Chalmer's called the Hard > > Problem of consciousness) is a direct result of an intelligent system > being > built with a sufficient level of complexity and self-reflection.> > Make no > mistake: the argument is long and tangled (I will write it up a > length when > I can) so I do not pretend to be trying to convince you of > its validity > here. All I am trying to do at this point is to state that > THAT is my > current understanding of what would happen.> > Let me rephrase that: we (a > subset of the AI community) believe that we > have discovered concrete > reasons to predict that a certain type of > organization in an intelligent > system produces consciousness.> > This is not meant to be one of those claims > that can be summarized in a > quick analogy, or quick demonstration, so there > is no way for me to > convince you quickly, all I can say is that we have > very string reasons > to believe that it emerges.> > > > You mentioned in > previous posts that the AI would only be programmed > > with 'Nice feelings' > and would only ever want to serve the good of > > mankind ? If the AI has > it's own ability to think etc, what is stopping > > it from developing > negative thoughts....the word 'feeling' in itself > > conjures up both good > and bad. For instance...I am an AI...I've > > witnessed an act of injustice, > seeing as I can feel and have > > consciousness my consciousness makes me > feel Sad / Angry ?> > Again, I have talked about this a few times before > (cannot remember the > most recent discussion) but basically there are two > parts to the mind: > the thinking part and the motivational part. If the AGI > has a > motivational that feels driven by empathy for humans, and if it does > not > possess any of the negative motivations that plague people, then it > > would not react in a negative (violent, vengeful, resentful.... etc) way.> > > Did I not talk about that in my reply to you? How there is a difference > > between having consciousness and feeling motivations? Two completely > > separate mechanisms/explanations?> > > > > Hold on...that would not be > possible seeing as my owner has an 'Off' > > button he can push to avoid me > feeling that way and hay I have only been > > programmed with 'Nice feelings' > even though my AI Creators have told the > > rest of the world I have a full > working conscious. It's starting to > > sound a bit like me presenting myself > to the world after my > > 'Hippocampus' has been removed or better yet I've > had a full frontal > > labotomy'.> > ["Full Frontal Labotomy"? :-) You mean > pre-frontal lobotomy, maybe. > Either that or this is a great title for a > movie about a lab full of > scientists trapped in a nudist colony].> > Not at > all like that. Did you ever have a good day, when you were so > relaxed that > nothing could disturb your feelings of generosity to the > world? Imagine a > creature that genuinely felt like that, and simple > never could have a bad > day.> > But to answer you more fully: all of this depends on exactly how the > > "motivation system" of humans and AGIs is designed. We can only really > > have that discussion in the context of a detailed knowledge of > specifics, > surely?> > > > And you say the AI will have thoughts and feelings about the > world > > around it ? I shudder to think what a newly born, pure AI had to > think > > about the world around us now. Or is that your ultimate goal in > this > > Utopia that you see Richard ? That the AI's will become like > Spiritual > > Masters to us and make everything 'all better' so to speak by > creating > > little 'ME' worlds for us very confused, 'life purpose' seeking > people ?> > No, they will just solve enough of the boring problems that we > can enjoy > the rest of life.> > Please also note the ideas in my parallel > discussion with Matt Mahoney: > do not be tempted to think of a Them and Us > situation: we would have > the ability to become just as knowledgeable as > they are, at any time. > We could choose our level of understanding on a day > to day basis, the > way we now choose our clothes. Same goes for them.> > We > would not be two species. Not master and servant. Just one species > with > more options than before.> > [I can see I am going to have to write this out > in more detail, just to > avoid the confusion caused by brief glimpses of the > larger picture].> > > > Richard Loosemore> > > > Candice> > > > > > > > > > Date: Thu, 25 Oct 2007 19:02:35 -0400> > > From: [EMAIL PROTECTED]> > > To: > singularity@v2.listbox.com> > > Subject: Re: [singularity] John Searle...> > > >> > > candice schuster wrote:> > > > Richard,> > > >> > > > Thank you for a > thought provoking response. I admire your ability to> > > > think with both > logic and reason. I think what Searle was trying to > > get> > > > at was > this...and I have read 'The Chinese Room Theory'...I think that> > > > what > he was trying to say was...if the human brain breaks down code > > like> > > > > a machine does, that does not make it understand the logic of the > > > code,> > > > it is afterall code. If you go back to basics, for example > binary > > code,> > > > it becomes almost sequence and you are (well some of > us are, like> > > > machines) able to understand how to put the puzzle > together again > > but we> > > > may not understand the logic behind that > code, ie: The Chinese > > Language> > > > as a whole.> > > >> > > > Although > for instance the AI has the ability to decifer the code> > > > and respond, > it does not understand the whole, which is funny in a way> > > > as you call > your cause 'Singularity'...which to me implies 'wholeness'> > > > for some > reason.> > > >> > > > Regarding your comment on....shock, horror, they made > an AI that has> > > > human cognitive thought processes, quite the contrary > Richard, if you> > > > and the rest of the AI community come up with the > goods I would be > > most> > > > intrigued to sit your AI down in front of me > and ask it.......'Do you> > > > understand the code 'SMILE' ?'> > >> > > A > general point about your reply.> > >> > > I think some people have a mental > picture of what a computer does when> > > it is running an AI program, in > which the computer does an extremely> > > simple bit of symbol manipulation, > and the very "simplicity" of what is> > > happening in their imagined > computer is what makes them think: this> > > machine is not really > understanding anything at all.> > >> > > So for example, if the computer is > set up SMILE subroutine that just> > > pulled a few muscles around, and this > SMILE subroutine was triggered,> > > say, when the audio detectors picked up > the sound of someone laughing,> > > then this piece of code would not be > understanding or feeling a smile.> > >> > > I agree: it would not. Most other > AI researchers would agree that such> > > a simple piece of code is not a > system that "understands" anything.> > > (Not all would agree, but let's > skirt that for the moment).> > >> > > But this where a simple mental image of > what goes in a computer can be a> > > very misleading thing. If you thought > that all AI programs were just> > > the same as this, then you might think > that it is just as easy to> > > dismiss all AI programs with the same "This > is not really understanding"> > > verdict.> > >> > > If Searle had only said > that he objected to simple programs being> > > described as "conscious" or > "self aware" then all power to him.> > >> > > So what happens in a real AI > program that actually has all the machinery> > > to be intelligent? ALL of > the machinery, mark you.> > >> > > Well, it is vastly more complex: a huge > amount of processing happens,> > > and the "smile" response comes out for the > right reasons.> > >> > > Why is that more than just a SMILE subroutine being > triggered by the> > > audio detectors measuring the sound of laughter?> > >> > > > Because this AI system is doing some very special things along with all> > > > the smiling: it is thinking about its own thoughts, among other things,> > > > and what we know (believe) is that when the system gets that complicated> > > > and has that particular mix of self-reflection in it, the net result is> > > > something that must talk about having an inner world of experience. It> > > > will talk about qualia, it will talk about feelings .... and not because> > > > it has been programmed to do that, but because when it tries to> > > > understand the world it really does genuinely find those things.> > >> > > > This is the step I mentioned in the last message I sent, and it is very> > > > very subtle: when you try to think about what is going on in the AI,> > > you > come to the inevitable conclusion that we are also "AI" systems, but> > > the > truth is that all AI systems (natural and artifical) possess some> > > > special properties: they have this thing that you describe as> > > subjective > consciousness.> > >> > > This is difficult to talk about in such a short > space, but the crude> > > summary is that if you make an AI extremely complex > (with> > > self-reflection, and with no direct connections between things > like a> > > smile and the causes of that smile) then that very complexity > gives rise> > > to something that was not there before: consciousness.> > >> > > >> > >> > > Richard Loosemore> > >> > > -----> > > This list is sponsored > by AGIRI: http://www.agiri.org/email> > > To unsubscribe or change your > options, please go to:> > > http://v2.listbox.com/member/?&> > > > > > > ------------------------------------------------------------------------> > > The next generation of MSN Hotmail has arrived - Windows Live Hotmail > > > <http://www.newhotmail.co.uk>> > > ------------------------------------------------------------------------> > > This list is sponsored by AGIRI: http://www.agiri.org/email> > To unsubscribe > or change your options, please go to:> > http://v2.listbox.com/member/?& > > > <http://v2.listbox.com/member/?&>> > -----> This list is sponsored by AGIRI: > http://www.agiri.org/email> To unsubscribe or change your options, please go > to:> http://v2.listbox.com/member/?& _________________________________________________________________ 100’s of Music vouchers to be won with MSN Music https://www.musicmashup.co.uk
----- This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=4007604&id_secret=57950185-dc7229