> Now about building a rational vs non-rational AGI, how would you go about 
> modeling a non-rational part of it?  Short of a random number generator?

Why would you want to build a non-rational AGI?  It seems like a *really* bad 
idea.  I think I'm missing your point here.

> For the most part we Do want a rational AGI, and it DOES need to explain 
> itself.  One fo the first tasks of AGI will be to replace all of the current 
> expert systems in fields like medicine.  

Yep.  That's my argument and you expand it well.

> Now for some tasks it will not be able to do this, or not within a small 
> amount of data and explanations.  The level that it is able to generalize 
> this information will reflect its usefullness and possibly intelligence.

Yep.  You're saying exactly what I'm thinking.

>  For many decisions I believe a small feature set is required, with the 
> larger possible features being so lowly weighted as to not have much impact.

This is where Ben and I are sort of having a debate.  I agree with him that the 
brain may well be using the larger number since it is massively parallel and it 
therefore can.  I think that we differ on whether or not the larger is required 
for AGI (Me = No, Ben = Yes) -- which reminds me . . . 

Hey Ben, if the larger number IS required for AGI, how do you intend to do this 
in a computationally feasible way in a non-massively-parallel system?





  ----- Original Message ----- 
  From: James Ratcliff 
  To: agi@v2.listbox.com 
  Sent: Tuesday, December 05, 2006 11:17 AM
  Subject: Re: Re: Re: Re: [agi] A question on the symbol-system hypothesis




  BillK <[EMAIL PROTECTED]> wrote:
    On 12/4/06, Mark Waser wrote:
    >
    > Explaining our actions is the reflective part of our minds evaluating the
    > reflexive part of our mind. The reflexive part of our minds, though,
    > operates analogously to a machine running on compiled code with the
    > compilation of code being largely *not* under the control of our conscious
    > mind (though some degree of this *can* be changed by our conscious minds).
    > The more we can correctly interpret and affect/program the reflexive part 
of
    > our mind with the reflective part, the more intelligent we are. And,
    > translating this back to the machine realm circles back to my initial 
point,
    > the better the machine can explain it's reasoning and use it's explanation
    > to improve it's future actions, the more intelligent the machine is (or, 
in
    > reverse, no explanation = no intelligence).
    >

    Your reasoning is getting surreal.

    As Ben tried to explain to you, 'explaining our actions' is our
    consciousness dreaming up excuses for what we want to do anyway. Are
    you saying that the more excuses we can think up, the more intelligent
    we are? (Actually there might be something in that!).

    You seem to have a real difficulty in admitting that humans behave
    irrationally for a lot (most?) of the time. Don't you read newspapers?
    You can redefine rationality if you like to say that all the crazy
    people are behaving rationally within their limited scope, but what's
    the point? Just admit their behaviour is not rational.

    Every time someone (subconsciously) decides to do something, their
    brain presents a list of reasons to go ahead. The reasons against are
    ignored, or weighted down to be less preferred. This applies to
    everything from deciding to get a new job to deciding to sleep with
    your best friend's wife. Sometimes a case arises when you really,
    really want to do something that you *know* is going to end in
    disaster, ruined lives, ruined career, etc. and it is impossible to
    think of good reasons to proceed. But you still go ahead anyway,
    saying that maybe it won't be so bad, maybe nobody will find out, it's
    not all my fault anyway, and so on.....

    Human decisions and activities are mostly emotional and irrational.
    That's the way life is. Because life is uncertain and unpredictable,
    human decisions are based on best guesses, gambles and basic
    subconscious desires.

    An AGI will have to cope with this mess. Basing an AGI on iron logic
    and 'rationality' alone will lead to what we call 'inhuman'
    ruthlessness.


    BillK


  You just rationlized the reasons for human choice in your above arguement 
yourself :}
  MOST humans act rationaly MOST of the time.  They may not make 'good' 
decisions, but they are rational ones, if you  decides to sleep with your best 
friends wife, you do so because you are attracted and you want her, and you 
rationlize you will probably not get caught.  You have stated the reasons, and 
you move ahead with that plan.
    Vague stuff you cant rationalize easily is why you like the appearance of 
someones face, or why you like this flavor of ice cream.  Those are hard to 
rationalize, but much of our behaviour is easier.
    Now about building a rational vs non-rational AGI, how would you go about 
modeling a non-rational part of it?  Short of a random number generator?

    For the most part we Do want a rational AGI, and it DOES need to explain 
itself.  One fo the first tasks of AGI will be to replace all of the current 
expert systems in fields like medicine.  
    For these it is not merely good enough to say, (as a Doctor AGI) I think he 
has this cancer, and you should treat him with this strange procedure.  There 
must be an accounting that it can present to other doctors and say, yes, I 
noticed a coorelation between these factors that lead me to believe this, with 
this certainty.  An early AI must also proove its merit by explaining what it 
is doing to build up a level of trust.
     Further, it is important in another fashion, in that we can turn around 
and use these smart AI's to further train other Doctors or specialists with the 
AGI's explainations.

  Now for some tasks it will not be able to do this, or not within a small 
amount of data and explanations.  The level that it is able to generalize this 
information will reflect its usefullness and possibly intelligence.

  In the Halo expirement for the Chemistry API, they were graded not only on 
correct answers but also in their explanations of how they got to those answers.
  Some of the explanations were short concise and well reasoned, some fo them 
though, went down to a very basic level of detail and lasted for a couple of 
pages.

  If you are flying to Austin, and asking a AGI to plan your route, and it 
chooses a Airline that sounds dodgy that you have never heard of, mainly 
because it was cheap or some other reasoning, you def want to know why it 
choose that, and tell it not to weight that feature as highly.
    For many decisions I believe a small feature set is required, with the 
larger possible features being so lowly weighted as to not have much impact.

  James Ratcliff



  _______________________________________
  James Ratcliff - http://falazar.com
  New Torrent Site, Has TV and Movie Downloads! 
http://www.falazar.com/projects/Torrents/tvtorrents_show.php


------------------------------------------------------------------------------
  Have a burning question? Go to Yahoo! Answers and get answers from real 
people who know.
------------------------------------------------------------------------------
   This list is sponsored by AGIRI: http://www.agiri.org/email
  To unsubscribe or change your options, please go to:
  http://v2.listbox.com/member/?list_id=303 

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303

Reply via email to