BillK <[EMAIL PROTECTED]> wrote: On 12/4/06, Mark Waser  wrote:
>
> Explaining our actions is the reflective part of our minds evaluating the
> reflexive part of our mind.  The reflexive part of our minds, though,
> operates analogously to a machine running on compiled code with the
> compilation of code being largely *not* under the control of our conscious
> mind (though some degree of this *can* be changed by our conscious minds).
> The more we can correctly interpret and affect/program the reflexive part of
> our mind with the reflective part, the more intelligent we are.  And,
> translating this back to the machine realm circles back to my initial point,
> the better the machine can explain it's reasoning and use it's explanation
> to improve it's future actions, the more intelligent the machine is (or, in
> reverse, no explanation = no intelligence).
>

Your reasoning is getting surreal.

As Ben tried to explain to you, 'explaining our actions' is our
consciousness dreaming up excuses for what we want to do anyway.  Are
you saying that the more excuses we can think up, the more intelligent
we are? (Actually there might be something in that!).

You seem to have a real difficulty in admitting that humans behave
irrationally for a lot (most?) of the time. Don't you read newspapers?
You can redefine rationality if you like to say that all the crazy
people are behaving rationally within their limited scope, but what's
the point? Just admit their behaviour is not rational.

Every time someone (subconsciously) decides to do something, their
brain presents a list of reasons to go ahead. The reasons against are
ignored, or weighted down to be less preferred. This applies to
everything from deciding to get a new job to deciding to sleep with
your best friend's wife. Sometimes a case arises when you really,
really want to do something that you *know* is going to end in
disaster, ruined lives, ruined career, etc. and it is impossible to
think of good reasons to proceed. But you still go ahead anyway,
saying that maybe it won't be so bad, maybe nobody will find out, it's
not all my fault anyway, and so on.....

Human decisions and activities are mostly emotional and irrational.
That's the way life is. Because life is uncertain and unpredictable,
human decisions are based on best guesses, gambles and basic
subconscious desires.

An AGI will have to cope with this mess. Basing an AGI on iron logic
and 'rationality' alone will lead to what we call 'inhuman'
ruthlessness.


BillK

You just rationlized the reasons for human choice in your above arguement 
yourself :}
MOST humans act rationaly MOST of the time.  They may not make 'good' 
decisions, but they are rational ones, if you  decides to sleep with your best 
friends wife, you do so because you are attracted and you want her, and you 
rationlize you will probably not get caught.  You have stated the reasons, and 
you move ahead with that plan.
  Vague stuff you cant rationalize easily is why you like the appearance of 
someones face, or why you like this flavor of ice cream.  Those are hard to 
rationalize, but much of our behaviour is easier.
  Now about building a rational vs non-rational AGI, how would you go about 
modeling a non-rational part of it?  Short of a random number generator?

  For the most part we Do want a rational AGI, and it DOES need to explain 
itself.  One fo the first tasks of AGI will be to replace all of the current 
expert systems in fields like medicine.  
  For these it is not merely good enough to say, (as a Doctor AGI) I think he 
has this cancer, and you should treat him with this strange procedure.  There 
must be an accounting that it can present to other doctors and say, yes, I 
noticed a coorelation between these factors that lead me to believe this, with 
this certainty.  An early AI must also proove its merit by explaining what it 
is doing to build up a level of trust.
   Further, it is important in another fashion, in that we can turn around and 
use these smart AI's to further train other Doctors or specialists with the 
AGI's explainations.

Now for some tasks it will not be able to do this, or not within a small amount 
of data and explanations.  The level that it is able to generalize this 
information will reflect its usefullness and possibly intelligence.

In the Halo expirement for the Chemistry API, they were graded not only on 
correct answers but also in their explanations of how they got to those answers.
Some of the explanations were short concise and well reasoned, some fo them 
though, went down to a very basic level of detail and lasted for a couple of 
pages.

If you are flying to Austin, and asking a AGI to plan your route, and it 
chooses a Airline that sounds dodgy that you have never heard of, mainly 
because it was cheap or some other reasoning, you def want to know why it 
choose that, and tell it not to weight that feature as highly.
  For many decisions I believe a small feature set is required, with the larger 
possible features being so lowly weighted as to not have much impact.

James Ratcliff



_______________________________________
James Ratcliff - http://falazar.com
New Torrent Site, Has TV and Movie Downloads! 
http://www.falazar.com/projects/Torrents/tvtorrents_show.php
 
---------------------------------
Have a burning question? Go to Yahoo! Answers and get answers from real people 
who know.

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303

Reply via email to