My suggestion, criticized below (criticism can be valuable), was for just one of many possible uses of an open-source P2P AGI-at-home type system. I am totally willing to hear other proposals. Considering how little time I spent coming up with the one being criticized, I have a relatively low ego investment in it and I assume there will be better suggestions from others.
I think the hard part of AGI will be difficult to address on a P2P system with low interconnect bandwidth. I do because I believe the hard part of AGI will be learning appropriate dynamic controls for massively parallel systems computing over massive amounts of data, and the creation of automatically self organizing knowledge bases derived from computing from such massive amounts of knowledge in a highly non-localized way. For progress on these fronts at any reasonable speed you need massive bandwidth, which a current P2P system would lack, according to the previous communications on this thread. So a current P2P system on the web is not going to be a good test bed for anything approaching human-level AGI. But interesting things could be learned with P2P AGI-at-Home networks. In the NL example I proposed, the word senses and parsing were all to be learned with generalized AGI learning algorithms (although bootstrapped with some narrow AI tools) I think they could be a good test bed for AGI learning of self organizing gen-comp hierarchies because the training data is plentiful and easy to get, many of the gen-comp hierarchy of patterns that would be formed would be ones that we humans could understand, and the capabilities of the system would be ones we could compare to human level performance in a somewhat intuitive manner. With regard to the statement that "The proper order is: lexical rules first, then semantics, then grammar, and then the problem solving. The whole point of using massive parallel computation is to do the hard part of the problem" I have the following two comments: (1) As I have said before, the truly hard part of AGI is almost certainly going to be beyond a P2P network of PCs. And (2) with regard to the order of NL learning, I think a child actually learns semantics first (words associated with sets of experience), since most young children I have met start communicating first in single word statements. The word sense experts I proposed in the P2P system would be focusing on this level of knowledge. Unfortunately, they would be largely limited to experience in the form of a textual context, resulting in a quite limited form of experiential grounding. The type of generalized AGI learning algorithm I proposed would address lexical rules and grammar as part of both its study of grammar and word senses. I have only separated out different forms of expertise because each PC can only contain a relatively small amount of information, so there has to be some attempt to separate the P2P's AGI representation into regions with the highest locality of reference. In and ideal world this should be done automatically, but to do this well automatically would tend to require high bandwidth, which the P2P system wouldn't have. So at least initially it probably makes sense to have humans decide what the various fields of expertise are (although such decisions could be based on AGI derived data, such as that obtained from data access patterns on singe PC AGI prototypes, or even on an initial networked system). Also, I think we should take advantage of some of the narrow AI tools we have, such as parsers, WordNet, dictionaries, and word-sense quessers, to bootstrap the system so that we could get more deeply into the more interesting aspect of AGI such as semantic understanding faster. These narrow AI tools could be used in conjunction with AGI learning. For example, the output of a narrow AI parser or word sense labeler could be used to provide initial data used to train up AGI models, which could then replace or run in conjunction with the narrow AGI tools in a set of EM cycles, with the AGI models hopefully providing more consistent labeling at time progresses, and increasingly getting more weight relative to the narrow AI tools. Perhaps one aspect of the AGI-at-home project would be to develop a good generalized architecture for wedding various classes of narrow AI and AGI in such a learning environment. Narrow AI's are often very efficient, but they have very limitations which AGI can often overcome. Perhaps learning how to optimally wed the two could create systems that had the best features of both AGI and narrow AI, greatly increasing the efficiency of AGI. But there are all sorts of other interesting things that could be done with an AGI-at-home P2P system. I am claiming no special expertise as to what is the best use of it. For example, I think it would be interesting to see what sort of AGI's could be built on current PCs with up to 4G or RAM. It would be interesting to see just what capability they could have. If such systems were built, one could use multiples of them to more rapidly optimize the tuning of their various internal parameters for various tasks. Once such PC level AGI's were built, it would be interesting to see what communities of them would be able to do. It would also be interested to see what forms of useful AGI could be built using multiple machines on a P2P network, given the low bandwidth connecting them. It has been suggested that OpenCog be developed on Launchpad.com, which seems pretty cool. It is an open source website for cooperative software development. It has multiple different levels at which members of a community can collaborate, including at the what-do-we-want-to-do level. It least initially that will probably be one of the most interesting levels of the site. Hopefully over time people will come back with more and more working modules, and such feedback can better inform the high level discussions about what direction in which to go. The system is designed to let project have different branches that go in different directions, and to let people see which branches currently are being most used and attended to. So Richard Loosemore will be free to try to develop a group to explore complexity, another group will be able to focus on individual PC level AGI, some will be able to specialize on generalization, others on inference control, etc. It seems to me the key variables will be how much participation from talented people the project will be able to attract and how easy its initial code base is to work with. Ed Porter -----Original Message----- From: Matt Mahoney [mailto:[EMAIL PROTECTED] Sent: Monday, December 03, 2007 11:39 AM To: agi@v2.listbox.com Subject: Re: Hacker intelligence level [WAS Re: [agi] Funding AGI research] --- Richard Loosemore <[EMAIL PROTECTED]> wrote: > Menawhile, unfortunately, solving all those other issues like making > parsers and trying to do word-sense disambiguation would not help one > whit to get the real theoretical task done. I agree. AI has a long history of doing the easy part of the problem first: solving the mathematics or logic of a word problem, and deferring the hard part, which is extracting the right formal statement from the natural language input. This is the opposite order of how children learn. The proper order is: lexical rules first, then semantics, then grammar, and then the problem solving. The whole point of using massive parallel computation is to do the hard part of the problem. -- Matt Mahoney, [EMAIL PROTECTED] ----- This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?& ----- This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=71558467-cd9e0d
<<attachment: winmail.dat>>