Durk, Absolutely right about the need for what is essentially an imaginative level of mind. But wrong in thinking:
"Vision may be classified under "Narrow" AI" You seem to be treating this extra "audiovisual perception layer" as a purely passive layer. The latest psychology & philosophy recognize that this is in fact a level of v. active thought and intelligence. And our culture is only starting to understand imaginative thought generally. Just to begin reorienting your thinking here, I suggest you consider how much time people spend on audiovisual information (esp. tv) vs purely symbolic information (books). And allow for how much and how rapidly even academic thinking is going audiovisual. Know of anyone trying to give computers that extra layer? I saw some vague reference about this recently.of which I have only a confused memory. Durk:Although I symphathize with some of Hawkin's general ideas about unsupervised learning, his current HTM framework is unimpressive in comparison with state-of-the-art techniques such as Hinton's RBM's, LeCun's convolutional nets and the promising low-entropy coding variants. But it should be quite clear that such methods could eventually be very handy for AGI. For example, many of you would agree that a reliable, computationally affordable solution to Vision is a crucial factor for AGI: much of the world's information, even on the internet, is encoded in audiovisual information. Extracting (sub)symbolic semantics from these sources would open a world of learning data to symbolic systems. An audiovisual perception layer generates semantic interpretation on the (sub)symbolic level. How could a symbolic engine ever reason about the real world without access to such information? Vision may be classified under "Narrow" AI, but I reckon that an AGI can never understand our physical world without a reliable perceptual system. Therefore, perception is essential for any AGI reasoning about physical entities! Greets, Durk On Sun, Mar 30, 2008 at 4:34 PM, Derek Zahn <[EMAIL PROTECTED]> wrote: It seems like a reasonable and not uncommon idea that an AI could be built as a mostly-hierarchical autoassiciative memory. As you point out, it's not so different from Hawkins's ideas. Neighboring "pixels" will correlate in space and time; "features" such as edges should become principle components given enough data, and so on. There is a bunch of such work on self-organizing the early visual system like this. That overall concept doesn't get you very far though; the trick is to make it work past the first few rather obvious feature extraction stages of sensory data, and to account for things like episodic memory, language use, goal-directed behavior, and all other cognitive activity that is not just statistical categorization. I sympathize with your approach and wish you luck. If you think you have something that produce more than Hawkins has with his HTM, please explain it with enough precision that we can understand the details. ---------------------------------------------------------------------------- agi | Archives | Modify Your Subscription ------------------------------------------------------------------------------ agi | Archives | Modify Your Subscription ------------------------------------------------------------------------------ No virus found in this incoming message. Checked by AVG. Version: 7.5.519 / Virus Database: 269.22.1/1349 - Release Date: 3/29/2008 5:02 PM ------------------------------------------- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=98558129-0bdb63 Powered by Listbox: http://www.listbox.com