> Consider an AGI trying to discover world facts from textual inference and > finding Dawkin's book "The River that Runs Uphill" (or things about the > moon's color and finding the phrase "once in a blue moon" everywhere or the > SF story "The Moon is Green"). > > The problem seems to be that it's AI-complete to pick out the exceptions, the > figures of speech, the metaphorical uses, and so forth -- so you could use > textual inference if only you had an already working AGI :-) > > Bottom line: kids grow up in a confusing, ambiguous world (and there are > plenty of funny stories obout how they get things wrong as they learn) -- yet > they generally seem to be able to organize their experiences into usable > knowledge corpora as they go. The human mind is autogenous. What's the trick?
The trick to understanding "once in a blue moon" is to either -- look at the moon or -- ask someone preferably both ... I don't see any deep cognitive trick here, just a case where embodiment and conversational interaction are much stronger than text mining... -- Ben G ------------------------------------------- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=95818715-a78a9b Powered by Listbox: http://www.listbox.com