On 10/4/07, Edward W. Porter <[EMAIL PROTECTED]> wrote: > > > > Josh, > > (Talking of "breaking the small hardware mindset," thank god for the company > with the largest hardware mindset -- or at least the largest physical > embodiment of one-- Google. Without them I wouldn't have known what "FARG" > meant, and would have had to either (1) read your valuable response with > less than the understanding it deserves or (2) embarrassed myself by > admitting ignorance and asking for a clarification.) > > With regard to your answer, copied below, I thought the answer would be > something like that. > > So which of the below types of "representational problems" are the reasons > why their basic approach is not automatically extendable? > > > 1. They have no general purpose representation that can represent almost > anything in a sufficiently uniform representational scheme to let their > analogy net matching algorithm be universally applied without requiring > custom patches for each new type of thing to be represented. > > 2. They have no general purpose mechanism for determining what are relevant > similarities and generalities across which to allow slippage for purposes of > analogy. > > 3. They have no general purpose mechanism for automatically finding which > compositional patterns map to which lower level representations, and which > of those compositional patterns are similar to each other in a way > appropriate for slippages. > > 4. They have no general purpose mechanism for automatically determining what > would be appropriately coordinated slippages in semantic hyperspace. > > 5. Some reason not listed above. > > I don't know the answer. There is no reason why you should. But if you -- > or any other interested reader – do, or if you have any good thoughts on > the subject, please tell me.
I guess I do know more on this topic, but it is a long story for which I don't have the time to tell. Hopefully the following paper can answer some of the questions: A logic of categorization Pei Wang and Douglas Hofstadter Journal of Experimental & Theoretical Artificial Intelligence, Vol.18, No.2, Pages 193-213, 2006 Pei > I may be naïve. I may be overly big-hardware optimistic. But based on the > architecture I have in mind, I think a Novamente-type system, if it is not > already architected to do so, could be modified to handle all of these > problems (except perhaps 5, if there is a 5) and, thus, provide powerful > analogy drawing across virtually all domains. > > Edward W. Porter > Porter & Associates > 24 String Bridge S12 > Exeter, NH 03833 > (617) 494-1722 > Fax (617) 494-1822 > [EMAIL PROTECTED] > > > > -----Original Message----- > From: J Storrs Hall, PhD [mailto:[EMAIL PROTECTED] > Sent: Thursday, October 04, 2007 1:44 PM > To: agi@v2.listbox.com > Subject: Re: [agi] breaking the small hardware mindset > > > > On Thursday 04 October 2007 10:56:59 am, Edward W. Porter wrote: > > You appear to know more on the subject of current analogy drawing > > research than me. So could you please explain to me what are the major > > current problems people are having in trying figure out how to draw > > analogies using a structure mapping approach that has a mechanism for > > coordinating similarity slippage, an approach somewhat similar to > > Hofstadter approach in Copycat? > > > Lets say we want a system that could draw analogies in real time when > > generating natural language output at the level people can, assuming > > there is some roughly semantic-net like representation of world > > knowledge, and lets say we have roughly brain level hardware, what > > ever that is. What are the current major problems? > > The big problem is that structure mapping is brittlely dependent on > representation, as Hofstadter complains; but that the FARG school hasn't > really come up with a generative theory (every Copycat-like analogizer > requires a pile of human-written Codelets which increases linearly with the > knowledge base -- and thus there is a real problem building a Copycat that > can learn its concepts). > > In my humble opinion, of course. > > Josh > > ----- > This list is sponsored by AGIRI: http://www.agiri.org/email > To unsubscribe or change your options, please go to: > http://v2.listbox.com/member/?& > > ________________________________ > This list is sponsored by AGIRI: http://www.agiri.org/email > To unsubscribe or change your options, please go to: > http://v2.listbox.com/member/?& ----- This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=50073913-f918fd