> Consider an AGI trying to discover world facts from textual inference and
> finding Dawkin's book "The River that Runs Uphill" (or things about the
> moon's color and finding the phrase "once in a blue moon" everywhere or the
> SF story "The Moon is Green").
>
> The problem seems to be that it's AI-complete to pick out the exceptions, the
> figures of speech, the metaphorical uses, and so forth -- so you could use
> textual inference if only you had an already working AGI :-)
>
> Bottom line: kids grow up in a confusing, ambiguous world (and there are
> plenty of funny stories obout how they get things wrong as they learn) -- yet
> they generally seem to be able to organize their experiences into usable
> knowledge corpora as they go. The human mind is autogenous. What's the trick?

The trick to understanding "once in a blue moon" is to either

-- look at the moon

or

-- ask someone

preferably both ... I don't see any deep cognitive trick here, just a
case where embodiment
and conversational interaction are much stronger than text mining...

-- Ben G

-------------------------------------------
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244&id_secret=95818715-a78a9b
Powered by Listbox: http://www.listbox.com

Reply via email to