[EMAIL PROTECTED] wrote:
> 
> 
> On Oct 31, 3:28 pm, "Wei Dai" <[EMAIL PROTECTED]> wrote:
> 
>> 4. For someone on a practical mission to write an AI that makes sensible
>> decisions, perhaps the model can serve as a starting point and as
>> illustration of how far away we still are from that goal.
> 
> Heh.  Yes, very interesting indeed.  But a huge body of knowledge and
> a great deal of smartness is needed to even begin to grasp all that
> stuff ;)
> 
> As regards AI I gotta wonder whether that 'Decision Theory' stuff is
> really 'the heart of the matter'  - perhaps its the wrong level of
> abstraction for the problem.  That is it say, it would be great if the
> AI could work out all the decision theory for itself, rather than
> having us trying to program it in (and probably failing miserably).
> Certainly, I'm sure as hell not smart enough to come up with a working
> model of decisions.  So, rather than trying to do the impossible,
> better to search for a higher level of abstraction.  Look for the
> answers in communication theory/ontology, rather than decision
> theory.  Decision theory would be derivative of an effective ontology
> - that saves me the bother of trying to work it out ;)

Decisions require some value structure.  To get values from an ontology you'd 
have to get around the Naturalistic fallacy.

Brent Meeker

--~--~---------~--~----~------------~-------~--~----~
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To post to this group, send email to [EMAIL PROTECTED]
To unsubscribe from this group, send email to [EMAIL PROTECTED]
For more options, visit this group at 
http://groups.google.com/group/everything-list?hl=en
-~----------~----~----~----~------~----~------~--~---

Reply via email to