Re: [agi] Second Life & the Gaza Conflict
On Sun, Jan 4, 2009 at 8:23 PM, Mike Tintner wrote: > I thought there might possibly be some interest in this (& perhaps > explanations) - the news item doesn't really explain how or why Second Life > is being used: > This has nothing to do with AGI. It is propaganda. Some Palestinians have set up a protest site in Second Life. There is a lot of propaganda activity (on both sides) if you do a google search. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b Powered by Listbox: http://www.listbox.com
Re: [agi] Creativity and Rationality (was: Re: Should I get a PhD?)
On Fri, Dec 19, 2008 at 6:40 PM, Ben Goertzel wrote: > > IMHO, Mike Tintner is not often rude, and is not exactly a "troll" because I > feel he is genuinely trying to understand the deeper issues related to AGI, > rather than mainly trying to stir up trouble or cause irritation > > However, I find conversing with him generally frustrating because he > combines > A) > extremely strong intuitive opinions about AGI topics > with > B) > almost utter ignorance of the details of AGI (or standard AI), or the > background knowledge needed to appreciate these details when compactly > communicated > > This means that discussions with Mike never seem to get anywhere... and, > frankly, I usually regret getting into them > In my opinion you are being too generous and your generosity is being taken advantage of. As well as trying to be nice to Mike, you have to bear list quality in mind and decide whether his ramblings are of some benefit to all the other list members. There are many types of trolls; some can be quite sophisticated. See: <http://ubuntuforums.org/showthread.php?p=1032102> The definitive guide to Trolls A classic troll tries to make us believe that he is a skeptic. He is divisive and argumentative with need-to-be-right attitude, "searching for the truth", flaming discussion, and sometimes insulting people or provoking people to insult him. A troll is usually an expert in reusing the same words of its opponents and in turning it against them. The Contrarian Troll. A sophisticated breed, Contrarian Trolls frequent boards whose predominant opinions are contrary to their own. A forum dominated by those who support firearms and knife rights, for example, will invariably be visited by Contrarian Trolls espousing their beliefs in the benefits of gun control. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b Powered by Listbox: http://www.listbox.com
Re: [agi] Creativity and Rationality (was: Re: Should I get a PhD?)
On Fri, Dec 19, 2008 at 3:55 PM, Mike Tintner wrote: > > (On the contrary, Pei, you can't get more narrow-minded than rational > thinking. That's its strength and its weakness). > Pei In case you haven't noticed, you won't gain anything from trying to engage with the troll. Mike does not discuss anything. He states his opinions in many different ways, pretending to respond to those that waste their time talking to him. But no matter what points are raised in discussion with him, they will only be used as an excuse to produce yet another variation of his unchanged opinions. He doesn't have any technical programming or AI background, so he can't understand that type of argument. He is against the whole basis of AGI research. He believes that rationality is a dead end, a dying culture, so deep-down, rational arguments mean little to him. Don't feed the troll! (Unless you really, really, think he might say something useful to you instead of just wasting your time). BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b Powered by Listbox: http://www.listbox.com
[agi] Re: [sl4] Join me on Bebo
On Thu, Dec 4, 2008 at 12:42 AM, Thomas McCabe wrote: > I'm not a moderator, but as a fellow list-user I do ask that you > please refrain from sending such things to SL4 in future. > This Bebo invite spam is more Bebo's fault than Ryan's. Bebo (like many social networking sites) by default assume that new users want to send an invite to everyone in their address book. They operate a block opt-in policy. If you just quickly click through the buttons you send an invite to everyone in your address book. Then sit and panic when you realize what you've done. :) Ryan isn't the only one that has been caught out by this system. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Entheogins, understainding the brain, and AGI
On Mon, Nov 24, 2008 at 7:51 PM, Eric Burton wrote: > This is a really good avenue of discussion for me. You'all probably should join rec.drugs.psychedelic <http://groups.google.com/group/rec.drugs.psychedelic/topics> People are still posting there, so the black helicopters haven't taken them all away yet. (Of course they might all be FBI agents. It's happened before) :) There are many similar interest groups for you to choose from. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Professor Asim Roy Finally Publishes Controversial Brain Theory
On Thu, Nov 20, 2008 at 3:52 PM, Ben Goertzel wrote: > > I skimmed over the paper at > http://wpcarey.asu.edu/pubs/index.cfm > and I have to say I agree with the skeptics. > > I don't doubt that this guy has made significant contributions in > other areas of science and engineering, but this paper displeases me a > great deal, due to making big claims of originality for ideas that are > actually very old hat, and bolstering these claims via attacking a > "straw man" of simplistic connectionism. > > > Double thumbs down: not for wrongheadedness, but for excessive claims > of originality plus egregious straw man arguments... > So, basically, you don't disagree with his paper to much. You just don't like his attitude.;) Danged AI researchers that think they know it all! ;) You don't think you could call it excessive PR where he is trying to dislodge an entrenched view? BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Professor Asim Roy Finally Publishes Controversial Brain Theory
On Thu, Nov 20, 2008 at 3:06 PM, Mark Waser <[EMAIL PROTECTED]> wrote: > Yeah. Great headline -- "Man beats dead horse beyond death!" > > I'm sure that there will be more details at 11. > > Though I am curious . . . . BillK, why did you think that this was worth > posting? > ??? Did you read the article? --- Quote: In the late '90s, Asim Roy, a professor of information systems at Arizona State University, began to write a paper on a new brain theory. Now, 10 years later and after several rejections and resubmissions, the paper "Connectionism, Controllers, and a Brain Theory" has finally been published in the November issue of IEEE Transactions on Systems, Man, and Cybernetics – Part A: Systems and Humans. Roy's theory undermines the roots of connectionism, and that's why his ideas have experienced a tremendous amount of resistance from the cognitive science community. For the past 15 years, Roy has engaged researchers in public debates, in which it's usually him arguing against a dozen or so connectionist researchers. Roy says he wasn't surprised at the resistance, though. "I was attempting to take down their whole body of science," he explained. "So I would probably have behaved the same way if I were in their shoes." No matter exactly where or what the brain controllers are, Roy hopes that his theory will enable research on new kinds of learning algorithms. Currently, restrictions such as local and memoryless learning have limited AI designers, but these concepts are derived directly from that idea that control is local, not high-level. Possibly, a controller-based theory could lead to the development of truly autonomous learning systems, and a next generation of intelligent robots. The sentiment that the "science is stuck" is becoming common to AI researchers. In July 2007, the National Science Foundation (NSF) hosted a workshop on the "Future Challenges for the Science and Engineering of Learning." The NSF's summary of the "Open Questions in Both Biological and Machine Learning" [see below] from the workshop emphasizes the limitations in current approaches to machine learning, especially when compared with biological learners' ability to learn autonomously under their own self-supervision: "Virtually all current approaches to machine learning typically require a human supervisor to design the learning architecture, select the training examples, design the form of the representation of the training examples, choose the learning algorithm, set the learning parameters, decide when to stop learning, and choose the way in which the performance of the learning algorithm is evaluated. This strong dependence on human supervision is greatly retarding the development and ubiquitous deployment of autonomous artificial learning systems. Although we are beginning to understand some of the learning systems used by brains, many aspects of autonomous learning have not yet been identified." Roy sees the NSF's call for a new science as an open door for a new theory, and he plans to work hard to ensure that his colleagues realize the potential of the controller model. Next April, he will present a four-hour workshop on autonomous machine learning, having been invited by the Program Committee of the International Joint Conference on Neural Networks (IJCNN). - Now his 'new' theory may be old hat to you personally, but apparently not to the majority of AI researchers, (according to the article). He must be saying something a bit unusual to have been fighting for ten years to get it published and accepted enough for him to now have been invited to do a workshop on his theory. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
[agi] Professor Asim Roy Finally Publishes Controversial Brain Theory
Nobody has mentioned this yet. <http://www.physorg.com/news146319784.html> Quotes: However, Roy's controversial ideas on how the brain works and learns probably won't immediately win over many of his colleagues, who have spent decades teaching robots and artificial intelligence (AI) systems how to think using the classic connectionist theory of the brain. Connectionists propose that the brain consists of an interacting network of neurons and cells, and that it solves problems based on how these components are connected. In this theory, there are no separate controllers for higher level brain functions, but all control is local and distributed fairly equally among all the parts. In his paper, Roy argues for a controller theory of the brain. In this view, there are some parts of the brain that control other parts, making it a hierarchical system. In the controller theory, which fits with the so-called computational theory, the brain learns lots of rules and uses them in a top-down processing method to operate. In his paper, Roy shows that the connectionist theory actually is controller-based, using a logical argument and neurological evidence. He explains that some of the simplest connectionist systems use controllers to execute operations, and, since more complex connectionist systems are based on simpler ones, these too use controllers. If Roy's logic correctly describes how the brain functions, it could help AI researchers overcome some inherent limitations in connectionist algorithms. "Connectionism can never create autonomous learning machines, and that's where its flaw is," Roy told PhysOrg.com. "Connectionism requires human babysitting of their learning algorithms, and that's not very brain-like. We don't guide and control the learning inside our head. etc BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] My prospective plan to neutralize AGI and other dangerous technologies...
On Tue, Nov 18, 2008 at 1:22 PM, Richard Loosemore wrote: > > I see how this would work: crazy people never tell lies, so you'd be able > to nail 'em when they gave the wrong answers. > Yup. That's how they pass lie detector tests as well. They sincerely believe the garbage they spread around. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: AW: AW: [agi] Language learning (was Re: Defining AGI)
On Thu, Oct 23, 2008 at 12:55 AM, Matt Mahoney wrote: > > > I suppose you are right. Instead of encoding mathematical rules as a grammar, > with enough training > data you can just code all possible instances that are likely to be > encountered. For example, instead > of a grammar rule to encode the commutative law of addition, > > 5 + 3 = a + b = b + a = 3 + 5 > a model with a much larger training data set could just encode instances with > no generalization: > > 12 + 7 = 7 + 12 > 92 + 0.5 = 0.5 + 92 > etc. > > I believe this is how Google gets away with brute force n-gram statistics > instead of more sophisticated > grammars. It's language model is probably > 10^5 times larger than a human model (10^14 bits vs > 10^9 bits). Shannon observed in 1949 that random strings generated by n-gram > models of English > (where n is the number of either letters or words) look like natural language > up to length 2n. For a > typical human sized model (1 GB text), n is about 3 words. To model strings > longer than 6 words we > would need more sophisticated grammar rules. Google can model 5-grams (see > http://googleresearch.blogspot.com/2006/08/all-our-n-gram-are-belong-to-you.html > ), so it is able to > generate and recognize (thus appear to understand) sentences up to about 10 > words. > Gigantic databases are indeed Google's secret sauce. See: <http://googleresearch.blogspot.com/2008/09/doubling-up.html> Quote: Monday, September 29, 2008 Posted by Franz Josef Och Machine translation is hard. Natural languages are so complex and have so many ambiguities and exceptions that teaching a computer to translate between them turned out to be a much harder problem than people thought when the field of machine translation was born over 50 years ago. At Google Research, our approach is to have the machines learn to translate by using learning algorithms on gigantic amounts of monolingual and translated data. Another knowledge source is user suggestions. This approach allows us to constantly improve the quality of machine translations as we mine more data and get more and more feedback from users. A nice property of the learning algorithms that we use is that they are largely language independent -- we use the same set of core algorithms for all languages. So this means if we find a lot of translated data for a new language, we can just run our algorithms and build a new translation system for that language. As a result, we were recently able to significantly increase the number of languages on translate.google.com. Last week, we launched eleven new languages: Catalan, Filipino, Hebrew, Indonesian, Latvian, Lithuanian, Serbian, Slovak, Slovenian, Ukrainian, Vietnamese. This increases the total number of languages from 23 to 34. Since we offer translation between any of those languages this increases the number of language pairs from 506 to 1122 (well, depending on how you count simplified and traditional Chinese you might get even larger numbers). - BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34 Powered by Listbox: http://www.listbox.com
Re: AW: AW: [agi] Re: Defining AGI
On Tue, Oct 21, 2008 at 10:31 PM, Ben Goertzel wrote: > > Incorrect things are wrapped up with correct things in peoples' minds > > However, pure slowness at learning is another part of the problem ... > Mark seems to be thinking of something like the checklist that the ISP technician walks through when you call with a problem. Even when you know what the problem is, the tech won't listen. He insists on working through his checklist, making you do all the irrelevant checks, eventually by a process of elimination, ending up with what you knew was wrong all along. Very little GI required. But Ben is saying that for evaluating science, there ain't no such checklist. The circumstances are too variable, you would need checklists to infinity. I go along with Ben. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34 Powered by Listbox: http://www.listbox.com
Re: [agi] Re: Meaning, communication and understanding
On Sun, Oct 19, 2008 at 1:23 PM, Dr. Matthias Heger wrote: > Natural communication and language understanding is completely comparable > with common processes in computer science. There is an internal data > representation. A subset of this data is translated into a linguistic string > and transferred to another agent which retranslates the message before it > possibly but not necessarily changes its database. > > The only reason why natural language understanding is so difficult is > because it needs a lot of knowledge to resolve ambiguities which humans > usually gain via own experience. > Google have just released a video showing their latest language research: The third video, "Applications of Human Language Technology," is a discussion of our enormous progress in large-scale automated translation of languages and speech recognition. Both of these technology domains are coming of age with capabilities that will truly impact what we expect of computers on a day-to-day basis. I discuss these technologies with human language technology experts Franz Josef Och, an expert in the automated translation of languages, and Mike Cohen, an expert in speech processing. <http://research.google.com/roundtable/HLT.html> BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34 Powered by Listbox: http://www.listbox.com
Re: [agi] Will Wright's Five Artificial Intelligence Prophecies
On Sat, Oct 18, 2008 at 8:28 PM, Bob Mottram wrote: > Some thoughts on this: > http://streebgreebling.blogspot.com/2008/10/will-wright-on-ai.html > I like his first point: MACHINES WILL NEVER ACHIEVE HUMAN INTELLIGENCE According to Wright, one of the main benefits of the quest for AI is a better definition of human intelligence. "Intelligence is whatever we can do that computers can't," says Wright. This reminds me of Mike Tintner. Even when these non-human intelligences are building space habitats, roaming the solar system and sending probes out to the stars, Mike will still be sitting there saying ' Ah, but they can't write poetry, so they are not really intelligent'. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34 Powered by Listbox: http://www.listbox.com
Re: [agi] META: A possible re-focusing of this list
On Wed, Oct 15, 2008 at 7:44 PM, John G. Rose wrote: > I'd go for 2 lists. Sometimes after working intensely on something concrete > and specific one wants to step back and theorize. And then particular AGI > approaches may be going down the wrong trail and need to step back and look > at things from a different perspective. > > Even so, with all this the messages in the one list still are grouped by > subject... I mean people can parse. But to simplify moderation and > organization, etc.. > I agree. I support more type 1 discussions. I have felt for some time that an awful lot of time-wasting has been going on here. I think this list should mostly be for computer tech discussion about methods of achieving specific results on the path(s) to AGI. I agree that there should be a place for philosophical discussion, either on a separate list, or uniquely identified in the Subject so that technicians can filter off such discussions. Some people may need to discuss philosophic alternative paths to AGI, to help clarify their thoughts. But if so, they are probably many years away from producing working code and might be hindering others who are further down the path of their own design. Two lists are probably best. Then if technicians want a break from coding, they can dip into the philosophy list, to offer advice or maybe find new ideas to play with. And, as John said. it would save on moderation time. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34 Powered by Listbox: http://www.listbox.com
Re: [agi] Advocacy Is no Excuse for Exaggeration
On Wed, Oct 15, 2008 at 9:46 AM, Eric Burton wrote: > My mistake I guess. I'm going to try harder to understand what you're > saying from now on. > Colin's profile on Nature says: I am a mature age PhD student with the sole intent of getting a novel chip technology and derivative products into commercial production. The chip technology facilitates natural learning of the kind biology uses to adapt to novelty. The artifacts will have an internal life. My mission is to create artificial (machines) that learn like biology learns and that have an internal life. Currently that goal requires lipid bilayer membrane molecular dynamics simulation. Publications Colin Hales. AI and Science's Lost Realm IEEE Intelligent Systems 21 , 76-81 (2006) Colin Hales. Physiology meets consciousness. A review of "The Primordial Emotions: The Dawning of Consciousness" by Derek Denton TRAFFIC EIGHT (2006) Hales, C. Qualia Ockham's Razor, Radio National, Australia 17 April (2005) Colin Hales. The 10 point framework and the altogether too hard basket Science and Consciousness Review (2003) --- BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34 Powered by Listbox: http://www.listbox.com
Re: [agi] Updated AGI proposal (CMR v2.1)
On Tue, Oct 14, 2008 at 2:41 PM, Matt Mahoney wrote: > But no matter. Whichever definition you accept, RSI is not a viable path to > AGI. An AI that is twice as smart as a > human can make no more progress than 2 humans. I can't say I've noticed two dogs being smarter than one dog. Admittedly, a pack of dogs can do hunting better, but they are not 'smarter'. Numbers just increase capabilities. Two humans can lift a heavier object than one human, but they are not twice as smart. As Ben says, I don't see a necessary connection between RSI and 'smarts'. It's a technique applicable from very basic levels. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34 Powered by Listbox: http://www.listbox.com
Re: [agi] Where the Future of AGI Lies
On Fri, Sep 19, 2008 at 10:05 PM, Matt Mahoney wrote: > From http://en.wikipedia.org/wiki/Yeltsin > > "Boris Yeltsin studied at Pushkin High School in Berezniki in Perm Krai. He > was fond of sports (in particular > skiing, gymnastics, volleyball, track and field, boxing and wrestling) > despite losing the thumb and index > finger of his left hand when he and some friends sneaked into a Red Army > supply depot, stole several > grenades, and tried to dissect them.[5]" > > But to be fair, Google didn't find it either. > I've had a play with this. I think you are asking the wrong question. See - It's your fault! :) The Yeltsin article doesn't say that he was a world leader. It says he was President of Russia. The article doesn't say he lost 2 fingers. It says he lost a thumb and index finger. So I think you are expecting quite a high level of understanding to match your query with these statements. If you ask "which president has lost a thumb and finger?". Then Powerset matches on the second page of results but Google matches on the first page of results. (Google is very good at keyword matching!) Cognition is still confused as it cannot find 'concepts' to match on. The Powerset FAQ says that it analyses your query and tries to extract a 'subject-relation-object' which it then tries to match. They give examples of the type of query they like. what was banned by the fda what caused the great depression The Cognition FAQ says that they try to find 'concepts' in your query and match on the 'concept' rather than actual words. i.e. The text "Did they adopt the bill?"; is known by Cognition to relate to information about "the approval of Proposition A", because "adopt" in the text means "to approve", and "bill" in the text means "a proposed law." So it looks like they don't have concepts for 'world leader = president' or 'thumb and index finger = 2 fingers' NLP isn't as easy as it looks! :) BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=114414975-3c8e69 Powered by Listbox: http://www.listbox.com
Re: [agi] Where the Future of AGI Lies
On Fri, Sep 19, 2008 at 6:13 PM, Matt Mahoney wrote: > Quick test. > > Q: What world leader lost 2 fingers playing with grenades as a boy? > powerset.com: doesn't know. > cognition.com (wiki): doesn't know. > > google.com: the second link leads to a scanned page of a book giving the > answer as Boris Yeltsin. > At present, they are not trying to compete with the Google search engine. :) They only search in Wikipedia, using this as a basis to test their Natural Language front-end. A better test would be to ask a complex question, where you know the answer is in Wikipedia, and see if they answer the question better than a Google search of Wikipedia only. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=114414975-3c8e69 Powered by Listbox: http://www.listbox.com
Re: [agi] Where the Future of AGI Lies
On Fri, Sep 19, 2008 at 3:15 PM, Jiri Jelinek wrote: > There is a difference between being good at > a) finding problem-related info/pages, and > b) finding functional solutions (through reasoning), especially when > all the needed data is available. > > Google cannot handle even trivial answer-embedded questions. > Last I heard Peter Norvig was saying that Google had no interest in putting a natural language front-end on Google. <http://slashdot.org/article.pl?sid=07/12/18/1530209> But other companies are interested. The main two are: Powerset <http://www.powerset.com/> and Cognition <http://www.cognition.com/> A new startup Eeggi is also interesting. <http://www.eeggi.com/> BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=114414975-3c8e69 Powered by Listbox: http://www.listbox.com
Re: [agi] Artificial humor
On Thu, Sep 11, 2008 at 2:28 PM, Jiri Jelinek wrote: > If you talk to a program about changing 3D scene and the program then > correctly answers questions about [basic] spatial relationships > between the objects then I would say it understands 3D. Of course the > program needs to work with a queriable 3D representation but it > doesn't need a "body". I mean it doesn't need to be a real-world > robot, it doesn't need to associate "self" with any particular 3D > object (real-world or simulated) and it doesn't need to be self-aware. > It just needs to be the 3D-scene-aware and the scene may contain just > a few basic 3D objects (e.g. the Shrdlu stuff). > Surely the DARPA autonomous vehicles driving themselves around the desert and in traffic show that computers can cope quite well with a 3D environment, including other objects moving around them as well? BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=111637683-c8fa51 Powered by Listbox: http://www.listbox.com
Re: AGI goals (was Re: Information theoretic approaches to AGI (was Re: [agi] The Necessity of Embodiment))
On Wed, Aug 27, 2008 at 8:43 PM, Abram Demski wrote: > By the way, where does this term "wireheading" come from? I assume > from context that it simply means self-stimulation. > Science Fiction novels. <http://en.wikipedia.org/wiki/Wirehead> In Larry Niven's Known Space stories, a wirehead is someone who has been fitted with an electronic brain implant (called a "droud" in the stories) to stimulate the pleasure centers of their brain. In 2006, The Guardian reported that trials of Deep brain stimulation with electric current, via wires inserted into the brain, had successfully lifted the mood of depression sufferers.[1] This is exactly the method used by wireheads in the earlier Niven stories (such as the 'Gil the Arm' story Death By Ectasy). In the Shaper/Mechanist stories of Bruce Sterling, "wirehead" is the Mechanist term for a human who has given up corporeal existence and become an infomorph. -- BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=111637683-c8fa51 Powered by Listbox: http://www.listbox.com
Re: [agi] AGI's Philosophy of Learning
On Tue, Aug 19, 2008 at 2:56 PM, Abram Demski <[EMAIL PROTECTED]> wrote: > Wow, sorry about that. I am using firefox and had no problems. The > site was just the first reference I was able to find using google. > > Wikipedia references the same fact: > > http://en.wikipedia.org/wiki/Feedforward_neural_network#Multi-layer_perceptron > I've done a bit more investigation. The web site is probably clean. These attacks are probably coming from a compromised ad server. ScanSafe Quote: "Online ads have become a primary target for malware authors because they offer a stealthy way to distribute malware to a wide audience. In many instances, the malware perpetrator can leverage the distributed nature of online advertising and the decentralization of website content to spread malware to hundreds of sites. So you might encounter these attacks at any site, because almost all sites serve up ads to you. And you're correct that FireFox with AdBlock Plus and NoScript is safe from these attacks. Using a Linux or Apple operating system is even safer. I dualboot to use Linux for browsing and only go into Windows when necessary. Nowadays you can also use virtualization to run several operating systems at once. Cooperative Linux also runs happily alongside Windows. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=111637683-c8fa51 Powered by Listbox: http://www.listbox.com
Re: [agi] AGI's Philosophy of Learning
On Tue, Aug 19, 2008 at 8:42 AM, Brad Paulsen wrote: > Abram, > > Just FYI... When I attempted to access the Web page in your message, > http://www.learnartificialneuralnetworks.com/ (that's without the > "backpropagation.html" part), my virus checker, AVG, blocked the attempt > with a message similar to the following: > > Threat detected! > Virus found: JS/Downloader.Agent > Detected on open > > Quarantined > > On a second attempt, I also got the IE 7.0 warning banner: > > "This website wants to run the following add-on: "Microsoft Data Access - > Remote Data Services Dat...' from 'Microsoft Corporation'. If you trust the > website and the add-on and want to allow it to run, click..." (of course, I > didn't click). > > This time, AVG gave me the option to "heal" the virus. I took this option. > > It may be nothing, but it also could be a "drive by" download attempt of > which the owners of that site may not be aware. > Yes, the possibility that the site has been hacked should always be considered as javascript injection attacks are becoming more and more common. Because of this, the latest version of AVG has been made to be very suspicious about javascript. This is causing some false detections when AVG encounters very complicated javascript as it errs on the side of safety. And looking at the source code for that page there is one large function near the top that might well have confused AVG (or it could be a hack, I'm not a javascript expert!). However, I scanned the site with Dr Web antivirus and it said the site was clean and the javascript was ok. This site has not yet been scanned by McAfee Site Advisor, but I have submitted it to them to be scanned soon. Of course, if you use the Mozilla FireFox browser you are protected from many drive by infections. Especially if you use the AdBlock Plus and NoScript addons. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=111637683-c8fa51 Powered by Listbox: http://www.listbox.com
Re: [agi] TOE -- US PATENT ISSUED for the TEN ETHICAL LAWS OF ROBOTICS
On Wed, Jul 23, 2008 at 1:13 AM, Mike Archbold wrote: > It seems to me like to be "real AGI" you have skipped over the parts of > Aristotle more applicable to AGI, like his metaphysics and logic. For > example in the metaphysics he talks about beginning and end, causes, > continuous/discrete, and this type of thing. At first glance it looks > like your invention starts with ethics; why not build atop a metaphysics > base? I'm not going to pass a judgement on your work but it seems like > it's not going over well here with the crowd that has dealt with patent > law. From my perspective I guess I don't like the idea of patenting some > automation of Aristotle unless it was in a kind of production-ready state > (ie., beyond mere concept stage). > His invention is ethics, because that's what his field of work is. See his list of books here: <http://www.allbookstores.com/author/John_E_Lamuth.html> * A Diagnostic Classification of the Emotions : A Three-digit Coding System for Affective Language by Jay D. Edwards (Illustrator), John E. Lamuth April 2005, Paperback List Price: $34.95 * Character Values : Promoting a Virtuous Lifestyle cover Character Values : Promoting a Virtuous Lifestyle by Jay D. Edwards (Illustrator), John E. Lamuth (Editor) April 2005, Paperback List Price: $28.95 * Communication Breakdown : Decoding The Riddle Of Mental Illness by Jay D. Edwards (Introduction by), John E. Lamuth (Editor) June 2004, Paperback List Price: $28.95 * A Revolution in Family Values : Tradition Vs. Technology by Jay D. Edwards (Illustrator), John E. Lamuth April 2002, Paperback List Price: $19.95 * A Revolution in Family Values : Spirituality for a New Millennium by John E. Lamuth March 2001, Hardcover List Price: $24.95 * The Ultimate Guide to Family Values : A Grand Unified Theory of Ethics and Morality by John E. Lamuth September 1999, Hardcover List Price: $19.95 and his author profile here: <http://www.angelfire.com/rnb/fairhaven/Contact_Fairhaven_Books.html> Biography John E. LaMuth is a 52 year-old counselor and author, native to the Southern California area. Credentials include a Bachelor of Science Degree in Biological Sciences from University of California, Irvine: followed by a Master of Science Degree in Counseling from California State University, Fullerton; with an emphasis in Marriage, Family, and Child Counseling. John is currently engaged in private practice in Divorce and Family Mediation Counseling in the San Bernardino County area - JLM Mediation Service - Lucerne Valley, CA 92356 USA. John also serves as an Adjunct Faculty Member at Victor Valley College, Victorville, CA. BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=108809214-a0d121 Powered by Listbox: http://www.listbox.com
Re: [agi] US PATENT ISSUED for the TEN ETHICAL LAWS OF ROBOTICS
On Mon, Jul 21, 2008 at 12:59 PM, Matt Mahoney wrote: > This is a real patent, unfortunately... > http://patft.uspto.gov/netacgi/nph-Parser?Sect2=PTO1&Sect2=HITOFF&p=1&u=%2Fnetahtml%2FPTO%2Fsearch-bool.html&r=1&f=G&l=50&d=PALL&RefSrch=yes&Query=PN%2F6587846 > > But I think it will expire before anyone has the technology to implement it. > :-) > I prefer Warren Ellis's angry, profane Three Laws of Robotics. (linked from BoingBoing) <http://www.warrenellis.com/?p=5426> BillK --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=108809214-a0d121 Powered by Listbox: http://www.listbox.com
Re: [agi] DARPA looking for 'Cutting Edge' AI projects
On Tue, Jun 24, 2008 at 11:52 AM, Stefan Pernar wrote: > "I'm a consultant with DARPA, and I'm working on an initiative to push the > boundaries of neuromorphic computing (i.e. artificial intelligence). The > project is designed to advance ideas all fronts, including measuring and > understanding biological brains, creating AI systems, and investigating the > fundamental nature of intelligence. I'm conducting a wide search of these > fields, but I wanted to know if any in this community know of neat projects > along those lines that I might overlook. Maybe you're working on a project > like that and want to talk it up? No promises (seriously), but interesting > work will be brought to the attention of the project manager I'm working > with. If you want to start up a dialog, send me an email, and we'll see > where it goes. I'll also be reading the comments for the story." > This sounds like the DARPA SyNAPSE program. Here is a blog post about the program with a link to the full specification document: <http://p9.hostingprod.com/@modha.org/blog/2008/04/> April 25, 2008 SyNAPSE: Systems of Neuromorphic Adaptive Plastic Scalable Electronics DARPA's Defense Sciences Office (DSO) has recently issued a Broad Agency Announcement entitled SyNAPSE. The program is led by Dr. Todd Hylton. --- This gives an idea of the people and projects already bidding. <http://www.eventmakeronline.com/dso/View/presenter.asp?MeetingID=561> DARPA SyNAPSE Bidders Workshop March 4, 2008 -- This is an overview of the meeting from one of the participants. <http://www.ine-news.org/view.php?article=rss-248&category=Workshops%3AGeneral> Report on the DARPA SyNAPSE Teaming Workshop Leslie Smith 14 March 2008 --- BillK --- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=106510220-47b225 Powered by Listbox: http://www.listbox.com
Re: [agi] More Info Please
On Tue, May 27, 2008 at 4:29 PM, Mark Waser wrote: >> Their overall figure of 73% for Apache and 19% for Microsoft IIS >> sounds reasonable to me. > > Yes, ladies and gentlemen, it's science in action . . . . pick the method > that produces the statistics that you want to hear. If surveying all the > web sites on the web doesn't produce the numbers you want/expect, select > some subset that does and then rationalize your choice. > > Remember the phrase about lies, damn lies, and statistics? I am appalled. > No Mark. It is partly the result of a deliberate MS policy to make their market share look bigger than it actually is. Always remember, the main thing that MS is good at is marketing. This article explains the two big changes in the Netcraft graph of websites (NOT web servers). <http://ostatic.com/158627-blog/web-servers-dont-count-apache-out> Quote: Microsoft picked up 5% of the market in one fell swoop in early 2006 when Go Daddy moved all of their millions of parked (i.e., not active by any sensible definition) domains from Apache servers to IIS. A win for Microsoft, to be sure, but not an indicator of any particular trend. And, Quote: In mid-2007, the 8 million or so Google-served sites started reporting themselves as "GFE" rather than Apache - even though GFE may well be a private fork of the Apache codebase. --- So MS has inflated it's Netcraft numbers by the simple tactic of persuading registrars to dump dead sites on IIS and Google has grabbed 6% by not reporting itself as Apache anymore. See - it's simple really. BillK --- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=103754539-40ed26 Powered by Listbox: http://www.listbox.com
Re: [agi] More Info Please
On Tue, May 27, 2008 at 3:53 PM, Mark Waser wrote: > No. You are not correct. Read their methodology > (http://www.securityspace.com/s_survey/faq.html?mondir=/200804&domdir=&domain=) > which I have copied and pasted below > >>>> We visit what we consider well-known sites. In our case, we define a >>>> well-known site as a site that had a link to it from at least one other >>>> site >>>> that we consider well-known. So, if we are visiting you, it means we know >>>> about you through a link from another site. > >>>> If a site stops responding to our request for 3 consecutive months, we >>>> automatically remove it from the survey. In this fashion, our list of known >>>> servers remains up to date. > >>>> Because of this technique, we find that we actually only visit about 10% >>>> of the web sites out on the web. This is because approximately 90% of all >>>> web sites are "fringe" sites, such as domain squatters, personal web sites, >>>> etc., that are considered unimportant by the rest of the web community >>>> (because no-one considers them important enough to link to.) > > That's fine by me. They are trying to survey the web servers that are actually *used* on the internet. Ignoring millions of parked domains on IIS servers run by some major registrars. Their overall figure of 73% for Apache and 19% for Microsoft IIS sounds reasonable to me. As J. Andrew Rogers said, Apache is probably a larger % than this in Silicon Valley. BillK --- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=103754539-40ed26 Powered by Listbox: http://www.listbox.com
Re: [agi] More Info Please
On Tue, May 27, 2008 at 2:20 PM, Mark Waser wrote: > Geez. What the heck is wrong with you people and your seriously bogus > stats? > > Try a real recognized neutral tracking service like Netcraft > (http://news.netcraft.com/archives/web_server_survey.html) > > > > March 2008 Percent April 2008 Percent Change > Apache 82,454,415 50.69% 83,554,638 50.42% -0.27 > Microsoft 57,698,503 35.47% 58,547,355 35.33% -0.14 > Google 9,012,004 5.54% 10,079,333 6.08% 0.54 > lighttpd 1,552,650 0.95% 1,495,308 0.90% -0.05 > Sun 546,581 0.34% 547,873 0.33% -0.01 > As I understand it, Netcraft's results are based on web sites, or more precisely, hostnames, rather than actual web servers. This introduces a bias because some servers run a large number of low-volume (or zero volume) web sites. This company attempts to survey web *servers* only (Note: Total is about 5% of Netcraft total) <http://www.securityspace.com/s_survey/data/200804/> More detail here: <http://www.securityspace.com/s_survey/data/200804/servers.html> This gives 73% for Apache and 19% for Microsoft. BillK --- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=103754539-40ed26 Powered by Listbox: http://www.listbox.com
Re: [agi] Porting MindForth AI into JavaScript Mind.html
On Sat, May 17, 2008 at 8:37 PM, John G. Rose wrote: >> It's hard to refute something vague or irrelevant, with 10 volumes of >> Astrology background behind it, and it's not necessary, as nonsense is >> usually self-evident. What's needed is merciless moderation... > > > I think it's overdue for someone to get a good Waseration. > That's a good one! Google has zero hits for Waseration. Explanation, please? BillK --- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=103754539-40ed26 Powered by Listbox: http://www.listbox.com
Re: [agi] How general can be and should be AGI?
On Sat, Apr 26, 2008 at 8:09 PM, Mike Tintner wrote: > So what you must tell me is how your or any geometrical system of analysis > is going to be able to take a rorschach and come up similarly with a > recognizable object or creature. Bear in mind, your system will be given no > initial clues as to what objects or creatures are suitable as potential > comparisons. It can by all means have a large set of visual images in > memory, as we do. But you must tell me how your system will connect the > rorschach with any of those images, such as a bat, - by *geometrical* > means. > This is called Content-based image retrieval (CBIR), also known as query by image content (QBIC) and content-based visual information retrieval (CBVIR) is the application of computer vision to the image retrieval problem, that is, the problem of searching for digital images in large databases. <http://en.wikipedia.org/wiki/CBIR> This is a hot area of computer research, with many test systems. (see article). Nothing to do with AGI, of course. Every post from Mike seems to be yet another different way of saying 'You're all wrong!' Are you sure you want to be on this list, Mike? BillK --- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=101455710-f059c4 Powered by Listbox: http://www.listbox.com
Re: [agi] Symbols
On Mon, Mar 31, 2008 at 10:56 AM, Mike Tintner wrote: > You guys probably think this is all rather peripheral and unimportant - they > don't teach this in AI courses, so it can't be important. > No. It means you're on the wrong list. > But if you can't see things whole, then you can't see or connect with the > real world. And, in case you haven't noticed, no AGI can connect with the > real world. In fact, there is no such thing as an AGI at the moment. And > there never will be if machines can't do what the brain does - which is, > first and last, and all the time, look at the world in images as wholes. > This list is not trying to duplicate a human brain. If you are the only person on the list who is correct, then you're wasting your time and our time here. It can't be much fun for you to spend all your time repeatedly telling everyone else that they've got it all wrong. Once or twice should be sufficient, else you turn into a troll. BillK --- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=98558129-0bdb63 Powered by Listbox: http://www.listbox.com
Re: [agi] Instead of an AGI Textbook
On Wed, Mar 26, 2008 at 12:47 PM, Ben Goertzel wrote: > Is there some kind of online software that lets a group of people > update a Mind Map diagram collaboratively, in the manner of a Wiki page? > > This would seem critical if a Mind Map is to really be useful for the purpose > you suggest... > Here is a recent review of online mind mapping software: <http://usableworld.terapad.com/index.cfm?fa=contentNews.newsDetails&newsID=41870&from=list&directoryId=14375> Online mindmap tools - Updated! By James Breeze in Mind Maps Published: Saturday, 08 March 08 BillK --- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=98558129-0bdb63 Powered by Listbox: http://www.listbox.com
Re: [agi] reasoning & knowledge
On Tue, Feb 26, 2008 at 8:29 PM, Ben Goertzel wrote: > > I don't think that formal logic is a suitably convenient language for > describing > motor movements or dealing with motor learning. > > But still, I strongly suspect one can produce software programs that do > handle > motor movement and learning effectively. They are symbolic at the level of > the programming language, but not symbolic at the level of the deliberative, > reflective component of the artificial mind doing the learning. > > A symbol is a symbol **to some system**. Just because a hunk of program > code contains symbols to the programmer, doesn't mean it contains symbols > to the mind it helps implement. Any more than a neuron being a symbol to a > neuroscientist, implies that neuron is a symbol to the mind it helps > implement. > > Anyway, I agree with you that formal logical rules and inference are not the > end-all of AGI and are not the right tool for handling visual imagination or > motor learning. But I do think they have an important role to play even so. > Asimo has a motor movement program. Obviously he didn't 'learn' it himself. But once written, it seems likely that similar sub-routines can be taken advantage of by later robots. BillK --- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=95818715-a78a9b Powered by Listbox: http://www.listbox.com
Re: [agi] Where are the women?
On Nov 30, 2007 2:37 PM, James Ratcliff wrote: > More Women: > > Kokoro (image attached) > So that's what a women is! I wondered.. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=70777441-ffcff3
Re: Hacker intelligence level [WAS Re: [agi] Funding AGI research]
On Nov 29, 2007 8:33 AM, Bob Mottram wrote: > My own opinion of all this, for what it's worth, is that the smart > hackers don't waste their time writing viruses/botnets. There are > many harder problems to which an intelligent mind can be applied. > > This discussion is a bit out of date. Nowadays no hackers (except for script kiddies) are interested in wiping hard disks or damaging your pc. Hackers want to *use* your pc and the data on it. Mostly the general public don't even notice their pc is working for someone else. When it slows down sufficiently, they either buy a new pc or take it to the shop to get several hundred infections cleaned off. But some infections (like rootkits) need a disk wipe to remove them completely. See: <http://blogs.zdnet.com/BTL/?p=7160&tag=nl.e589> Quote- On Wednesday, the SANS Institute released its top 20 security risks update for 2007. It's pretty bleak across the board. There are client vulnerabilities in browsers, Office software (especially the Microsoft variety), email clients and media players. On the server side, Web applications are a joke, Windows Services are a big target, Unix and Mac operating systems have holes, backup software is an issue as are databases and management servers. Even anti-virus software is a target. And assuming you button down all of those parts–good luck folks–you have policies to be implemented (rights, access, encrypted laptops etc.) just so people can elude them. Meanwhile, instant messaging, peer-to-peer programs and your VOIP system are vulnerable. The star of the security show is the infamous zero day attack. -- Original SANS report here - <http://www.sans.org/top20/?portal=bf37a5aa487a5aacf91e0785b7f739a4#c2> --- And, of course, all the old viruses are still floating around the net and have to be protected against. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=70081689-300ee8
Re: [agi] What best evidence for fast AI?
On 11/12/07, Mark Waser wrote: > :-) Then it was also translated into a Gantt chart because I can assure you > that I've seen it on far too many walls . . . . :-) > > but the point remains . . . . > > ON SECOND GLANCE -- the copyright on the Harris cartoon is 2007. I *know* > that I've seen the Gantt chart as far back as 1987 or 1986 because it was on > the wall on my Program Director's office -- so I suspect that the Gantt > chart is the original and Harris's the modification. > That's the site copyright message. The cartoon is old. I thought it was pre 1980, and I found this reference which said it was drawn in 1977. (Probably for the New Yorker magazine) <http://www.textbookleague.org/id-hx-1.htm> Like I said it is wildly popular. You get it on everything from T-shirts to coffee mugs. It is on the cover of his book of cartoons. It has been copied over into many domains, project planning, bio-science, etc. etc. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=64212365-c3e7f6
Re: [agi] What best evidence for fast AI?
On 11/12/07, Mark Waser wrote: > Hmm. Looks like my joke Gantt chart doesn't have the wide currency here > that I thought that it had. I assume that my previous explanation makes it > clear that we're in agreement. > It wasn't a Gantt chart. It was a famous (much-referenced) Sidney Harris cartoon: <http://www.sciencecartoonsplus.com/gallery.htm> BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=64179735-649824
Re: [agi] definition source?
On 11/6/07, Jiri Jelinek wrote: > Did you read the following definition somewhere? > "General intelligence is the ability to gain knowledge in one context > and correctly apply it in another." > I found it in notes I wrote for myself a few months ago and I'm not > sure now about the origin. > Might be one of my attempts, might not be. I ran a quick google search > (web and my emails) - no hit. > This article might be useful. <http://www.machineslikeus.com/cms/a-collection-of-definitions-of-intelligence.html> A Collection of Definitions of Intelligence Sat, 06/30/2007 By Shane Legg and Marcus Hutter This paper is a survey of a large number of informal definitions of "intelligence" that the authors have collected over the years. Naturally, compiling a complete list would be impossible as many definitions of intelligence are buried deep inside articles and books. Nevertheless, the 70-odd definitions presented here are, to the authors' knowledge, the largest and most well referenced collection there is. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=61544890-4b209a
Re: [agi] Nirvana? Manyana? Never!
On 11/3/07, Richard Hollerith wrote: > BillK writes: > >> Forget, for the moment, what you think is > >> possible - if you could have anything you wanted, is this the end you > >> would wish for yourself, more than anything else? > > > >Absolute Power over others and being worshipped as a God would be neat as > >well. > > > >Getting a dog is probably the nearest most humans can get to this. > > Is this a joke? > > Sometimes certain people can moralize too much or be too serious > all the time. Could it be that this is a reaction against > someone being too serious or too moralistic in the past and your > wanting to tweak their nose a bit and maybe get them to loosen up > a little? > On some levels. The best jokes have a lot of truth in them. Humor is a way of saying what it is forbidden to mention. Ever seen Bill Hicks in full flow? BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=60729119-674d0f
Re: [agi] Nirvana? Manyana? Never!
On 11/2/07, Eliezer S. Yudkowsky wrote: > I didn't ask whether it's possible. I'm quite aware that it's > possible. I'm asking if this is what you want for yourself. Not what > you think that you "ought to logically want", but what you really want. > > Is this what you lived for? Is this the most that Jiri Jelinek wants > to be, wants to aspire to? Forget, for the moment, what you think is > possible - if you could have anything you wanted, is this the end you > would wish for yourself, more than anything else? > Well, almost. Absolute Power over others and being worshipped as a God would be neat as well. Getting a dog is probably the nearest most humans can get to this. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=60258273-c65ec9
Re: Economic libertarianism [was Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On 10/6/07, a wrote: > I am skeptical that economies follow the self-organized criticality > behavior. > There aren't any examples. Some would cite the Great Depression, but it > was caused by the malinvestment created by Central Banks. e.g. The > Federal Reserve System. See the Austrian Business Cycle Theory for details. > In conclusion, economics is a bad analogy with complex systems. > My objection to economic libertarianism is that it's not a free market. A 'free' market is an impossibility. There will always be somebody who is bigger than me or cleverer than me or better educated than me, etc. A regulatory environment attempts to reduce the victimisation of the weaker members of the population and introduces another set of biases to the economy. A free market is just a nice intellectual theory that is of no use in the real world. (Unless you are in the Mafia, of course). BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=50792589-4d8a77
Re: The first-to-market effect [WAS Re: [agi] Religion-free technical content]
On 10/4/07, Bob Mottram <[EMAIL PROTECTED]> wrote: > To me this seems like elevating that status of nanotech to magic. > Even given RSI and the ability of the AGI to manufacture new computing > resources it doesn't seem clear to me how this would enable it to > prevent other AGIs from also reaching RSI capability. Presumably > "lesser techniques" means black hat activity, or traditional forms of > despotism. There seems to be a clarity gap in the theory here. > The first true AGI may be friendly, as suggested by Richard Loosemore. But if the military are working on developing an intelligent weapons system, then a sub-project will be a narrow AI project designed specifically to seek out and attack the competition *before* it becomes a true AGI. The Chinese are already constantly probing and attacking the western internet sites. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=49977621-104d4e
Re: [agi] Religion-free technical content
On 10/2/07, Mark Waser wrote: > A quick question for Richard and others -- Should adults be allowed to > drink, do drugs, wirehead themselves to death? > This is part of what I was pointing at in an earlier post. Richard's proposal was that humans would be asked in advance by the AGI what level of protection they required. So presumably Richard is thinking along the lines of a non-interfering AGI, unless specifically requested. There are obvious problems here. Humans don't know what is 'best' for them. Humans frequently ask for what they want, only later to discover that they really didn't want that. Humans change their mind all the time. Humans don't know 'in advance' what level of protection they would like. Death and/or mutilation comes very quickly at times. If I was intending to be evil, say, commit mass murder, I would request a lot of protection from the AGI, as other humans would be trying to stop me. -- I think the AGI will have great problems interfacing with these mixed-up argumentative humans. The AGI will probably have to do a lot of 'brain correction' to straighten out humanity. Let's hope that it knows what it is doing. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=48932774-c6db65
Re: [agi] Religion-free technical content
On 10/1/07, Richard Loosemore wrote: > > 3) The system would actually be driven by a very smart, flexible, subtle > sense of 'empathy' and would not force us to do painful things that were > "good" for us, for the simple reason that this kind of nannying would be > the antithesis of really intelligent empathy. > > If you want, give specific cases and we will try to see how it would behave. > This is heading straight for the eternal problem of good and evil. How does the AGI deal with bullying? Easy answer - It stops it. Then you have to get into the labyrinth of hard cases. Bullying covers everything from gang bosses committing murder and torture, to husband / wife psychological abuse, to office threats and allocation of unpleasant jobs, to kids verbal and physical abuse, to giving orders which you know will lead to much unpleasantness that the victim is unaware of. (Plausible denial comes in here, 'who me?'). In practice it rapidly becomes impossible to deal with these ornery humans. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=48545694-dbabc5
Re: [agi] Religion-free technical content
On 9/30/07, Edward W. Porter wrote: > > I think you, Don Detrich, and many others on this list believe that, for at > least a couple of years, it's still pretty safe to go full speed ahead on > AGI research and development. It appears from the below post that both you > and Don agree AGI can potentially present grave problems (which > distinguished Don from some on this list who make fun of anyone who even > considers such dangers). It appears the major distinction between the two > of you is whether, and how much, we should talk and think about the > potential dangers of AGI in the next few years. > Take the Internet, WWW and Usenet as an example. Nobody gave a thought to security while they were being developed. They were delighted and amazed that the thing worked at all. Now look at the swamp we have now. Botnets, viruses, trojans, phishing, DOS attacks, illegal software, illegal films, illegal music, pornography of every kind, etc. (Just wish I had a pornograph to play it on). BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=48269918-e87cb0
Re: [agi] Minimally ambiguous languages
On 6/5/07, Bob Mottram <[EMAIL PROTECTED]> wrote: I remember last year there was some talk about possibly using Lojban as a possible language use to teach an AGI in a minimally ambiguous way. Does anyone know if the same level of ambiguity found in ordinary English language also applies to sign language? I know very little about sign language, but it seems possible that the constraints applied by the relatively long time periods needed to produce gestures with arms/hands compared to the time required to produce vocalizations may mean that sign language communication is more compact and maybe less ambiguous. Also, comparing the way that the same concepts are represented using spoken and sign language might reveal something about how we normally parse sentences. <http://en.wikipedia.org/wiki/Basic_English> Ogden's rules of grammar for Basic English allows people to use the 850 words to talk about things and events in the normal English way. Ogden did not put any words into Basic English that could be paraphrased with other words, and he strove to make the words work for speakers of any other language. He put his set of words through a large number of tests and adjustments. He also simplified the grammar but tried to keep it normal for English users. More recently, it has influenced the creation of Simplified English, a standardized version of English intended for the writing of technical manuals. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=231415&user_secret=e9e40a7e
Re: [agi] My proposal for an AGI agenda
On 4/10/07, Eric Baum wrote: I'd commend to the LISP hackers' attention the compiler "Stalin" by Jeff Syskind, who last I knew was at Purdue. I'm uncertain the extent to which the compiler is available, but I imagine if you look around (for example find Syskind's home page) you will find papers or or pointers. My erstwhile collaborator Kevin Lang, initially a skeptic on the subject, ran extensive tests on Stalin and concluded the compiled code was substantially faster than compiled C and C++, even on problems where this was quite surprising. It's possible Kevin published something on these tests. <http://community.schemewiki.org/?Stalin> <http://en.wikipedia.org/wiki/Stalin_(Scheme_implementation)> <http://cobweb.ecn.purdue.edu/~qobi/software.html> Stalin is an aggressively optimizing Scheme compiler. It is the most highly optimizing Scheme compiler, and in fact one of the most highly optimizing compilers of any sort for any language. Stalin is publicly & freely available, licensed under the GNU GPL. It was written by Jeffrey M. Siskind. In detail, Stalin is a whole-program compiler that uses advanced flow analysis & closure conversion techniques. It compiles Scheme to highly optimized C. Stalin has a few very significant limitations, however: * it takes a *long* time to compile anything including the compiler * it is not designed for interactive use or debugging * it does not support R4RS/R5RS high-level macros * it does not support the full numeric tower The compiler itself does lifetime analysis and hence does not generate as much garbage as might be expected, but global reclamation of storage is done using the Boehm garbage collector. Scheme <http://cbbrowne.com/info/scheme.html> <http://en.wikipedia.org/wiki/Scheme_(programming_language)> Scheme is a LISP "dialect" that is relatively small, nicely supports tail recursion, provides block structure and lexical scoping, and gives a variety of object types "first-class" status (e.g. - first class objects are namable and can be passed around as function arguments, results, or as list elements). If Common LISP is considered analogous to C++ (which is not entirely unreasonable), Scheme would comparatively be analogous to C. Where Common LISP requires that the "rich functionality" (such as a wide variety of data structures and the Common Lisp Object System (CLOS)) come as an intrinsic part of the language, Scheme encourages the use of "libraries" to provide these sorts of additional functionality. The Scheme libraries are small enough that programmers commonly construct functionality using primitive functions, where a LISP system might have something already defined. This gives additional opportunity either to tune performance or shoot oneself in the foot by reimplementing it poorly... BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?list_id=303
Re: [agi] small code small hardware
On 3/29/07, kevin osborne wrote: You could argue that a lot of all this is the same kind of functions just operating in 'parrellel' with a lot of 'redundancy'. I'm not sure I buy that. Evolution is a miserly mistress. If thinking could have been achieved with less, it would have been, and any 'extra' would have no means of selection. The (also ridiculously large) amount of years involved in mammalian brain evolution all led towards what we bobble around with us today. I think there is an untold host of support functions necessary to take a Von Neumann machine to a tipping-point|critical-mass where it can truly think for itself. To even begin to equate top the generalised abilities of an imbecile. I think you have too high an opinion of Evolution. Evolution is kludge piled upon kludge. This is because evolution via natural selection cannot construct traits from scratch. New traits must be modifications of previously existing traits. This is called historical constraint. There are many examples available in nature of bad design. So it is not unlikely that a lot of the human brain processing is a redundant hangover from earlier designs. Of course, it is not a trivial problem to decide which functions are not required to create AGI. :) BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?list_id=303
[agi] Mind mapping software
I thought this free software opportunity might be of interest to some here. ConceptDraw MINDMAP 4 is a mind-mapping and team brainstorming tool with extended drawing capabilities. Use it to efficiently organize your ideas and tasks with the help of Mind Mapping technique. ConceptDraw MINDMAP 4 supports extra file formats, multi-page documents. It offers a rich collection of pre-drawn shapes. ConceptDraw MINDMAP 4 has extended capabilities for creating web sites and PowerPoint presentations. This software is temporarily available for free. * But you must download and install it within the next 19 hours. ** Restrictions for the free edition. 1. No free technical support 2. No free upgrades to future versions 3. Strictly non-commercial usage Normal price 119 USD. <http://www.giveawayoftheday.com/conceptdraw-mindmap-personal/> BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?list_id=303
Re: [agi] SOTA
On 1/6/07, Bob Mottram wrote: This is the way it's going to go in my opinion. In a house or office the robots would really be dumb actuators - puppets - being controlled from a central AI which integrates multiple systems together. That way you can keep the cost and maintenance requirements of the robot to a bare minimum. Such a system also future-proofs the robot in a rapidly changing software world, and allows intelligence to be provided as an internet based service. <http://www.pinktentacle.com/2006/12/top-10-robots-selected-for-robot-award-2006/> Robotic building cleaning system (Fuji Heavy Industries/ Sumitomo) - This autonomous robot roams the hallways of buildings, performing cleaning operations along the way. Capable of controlling elevators, the robot can move from floor to floor unsupervised, and it returns to its start location once it has finished cleaning. The robot is currently employed as a janitor at 10 high-rise buildings in Japan, including Harumi Triton Square and Roppongi Hills. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?list_id=303
Re: [agi] A question on the symbol-system hypothesis
On 12/5/06, Charles D Hixson wrote: BillK wrote: > ... > > Every time someone (subconsciously) decides to do something, their > brain presents a list of reasons to go ahead. The reasons against are > ignored, or weighted down to be less preferred. This applies to > everything from deciding to get a new job to deciding to sleep with > your best friend's wife. Sometimes a case arises when you really, > really want to do something that you *know* is going to end in > disaster, ruined lives, ruined career, etc. and it is impossible to > think of good reasons to proceed. But you still go ahead anyway, > saying that maybe it won't be so bad, maybe nobody will find out, it's > not all my fault anyway, and so on. > ... > > BillK I think you've got a time inversion here. The "list of reasons to go ahead" is frequently, or even usually, created AFTER the action has been done. If the list is being created BEFORE the decision, the list of reasons not to go ahead isn't ignored. Both "lists" are weighed, a decision is made, and AFTER the decision is made the reasons decided against have their weights reduced. If, OTOH, the decision is made BEFORE the "list of reasons" is created, then the list doesn't *get* created until one starts trying to justify the action, and for justification obviously reasons not to have done the thing are useless...except as a layer of whitewash to prove that "all eventualities were considered". For most decisions one never bothers to verbalize why it was, or was not, done. No time inversion intended. What I intended to say was that most (all?) decisions are made subconsciously before the conscious mind starts its reason / excuse generation process. The conscious mind pretending to weigh various reasons is just a human conceit. This feature was necessary in early evolution for survival. When danger threatened, immediate action was required. Flee or fight! No time to consider options with the new-fangled consciousness brain mechanism that evolution was developing. With the luxury of having plenty of time to reason about decisions, our consciousness can now play its reasoning games to justify what subconsciously has already been decided. NOTE: This is probably an exaggeration / simplification. ;) BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?list_id=303
Re: Marvin and The Emotion Machine [WAS Re: [agi] A question on the symbol-system hypothesis]
the effects of defects and bugs in other parts. This means that we will find it hard to guess both how our brains function as well as they do and why they evolved in the ways that they did, until we have had more experience at trying to build such systems ourselves, to learn which kinds of bugs are likely to appear and to find ways to keep them from disabling us. In the coming decades, many researchers will try to develop machines with Artificial Intelligence. And every system that they build will keep surprising us with their flaws (that is, until those machines become clever enough to conceal their faults from us). In some cases, we'll be able to diagnose specific errors in those designs and then be able to remedy them. But whenever we fail to find any such simple fix, we will have little choice except to add more checks and balances—for example, by adding increasingly elaborate Critics. And through all this, we can never expect to find any foolproof strategy to balance the advantage of immediate action against the benefit of more careful, reflective thought. Whatever we do, we can be sure that the road toward designing 'post-human minds' will be rough. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?list_id=303
Re: Re: Re: Re: [agi] A question on the symbol-system hypothesis
On 12/4/06, Mark Waser wrote: Explaining our actions is the reflective part of our minds evaluating the reflexive part of our mind. The reflexive part of our minds, though, operates analogously to a machine running on compiled code with the compilation of code being largely *not* under the control of our conscious mind (though some degree of this *can* be changed by our conscious minds). The more we can correctly interpret and affect/program the reflexive part of our mind with the reflective part, the more intelligent we are. And, translating this back to the machine realm circles back to my initial point, the better the machine can explain it's reasoning and use it's explanation to improve it's future actions, the more intelligent the machine is (or, in reverse, no explanation = no intelligence). Your reasoning is getting surreal. As Ben tried to explain to you, 'explaining our actions' is our consciousness dreaming up excuses for what we want to do anyway. Are you saying that the more excuses we can think up, the more intelligent we are? (Actually there might be something in that!). You seem to have a real difficulty in admitting that humans behave irrationally for a lot (most?) of the time. Don't you read newspapers? You can redefine rationality if you like to say that all the crazy people are behaving rationally within their limited scope, but what's the point? Just admit their behaviour is not rational. Every time someone (subconsciously) decides to do something, their brain presents a list of reasons to go ahead. The reasons against are ignored, or weighted down to be less preferred. This applies to everything from deciding to get a new job to deciding to sleep with your best friend's wife. Sometimes a case arises when you really, really want to do something that you *know* is going to end in disaster, ruined lives, ruined career, etc. and it is impossible to think of good reasons to proceed. But you still go ahead anyway, saying that maybe it won't be so bad, maybe nobody will find out, it's not all my fault anyway, and so on. Human decisions and activities are mostly emotional and irrational. That's the way life is. Because life is uncertain and unpredictable, human decisions are based on best guesses, gambles and basic subconscious desires. An AGI will have to cope with this mess. Basing an AGI on iron logic and 'rationality' alone will lead to what we call 'inhuman' ruthlessness. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?list_id=303
Re: [agi] A question on the symbol-system hypothesis
On 12/2/06, Mark Waser wrote: My contention is that the pattern that it found was simply not translated into terms you could understand and/or explained. Further, and more importantly, the pattern matcher *doesn't* understand it's results either and certainly could build upon them -- thus, it *fails* the test as far as being the central component of an RSIAI or being able to provide evidence as to the required behavior of such. Mark, I think you are making two very basic wrong assumptions. 1) That humans are able to understand everything if it is explained to them simply enough and they are given unlimited time. 2) That it is even possible to explain some very complex ideas in a simple enough fashion. Consider teaching the sub-normal. After much repetition they can be trained to do simple tasks. Not understanding 'why', but they can remember instructions eventually. Even high IQ humans have the same equipment, just a bit better. They still have limits to how much they can remember, how much information they can hold in their heads and access. If you can't remember all the factors at once, then you can't understand the result. You can write down the steps, all the different data that affect the result, but you can't assemble it in your brain to get a result. And I think the chess or Go examples are a good example. People who think that they can look through the game records and understand why they lost are just not trained chess or go players. They have a good reason to call some people 'Go masters' or 'chess masters'. I used to play competitive chess and I can assure you that when our top board player consistently beat us lesser mortals we could rarely point at move 23 and say 'we shouldn't have done that'. It is *far* more subtle than that. If you think you can do that, then you just don't understand the problem. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?list_id=303
Re: Re: [agi] A question on the symbol-system hypothesis
On 11/14/06, James Ratcliff wrote: If the "contents of a knowledge base for AGI will be beyond our ability to comprehend" then it is probably not human level AGI, it is something entirely new, and it will be alien and completely foriegn and unable to interact with us at all, correct? If you mean it will have more knowledge than we do, and do things somewhat differently, I agree on the point. "You can't look inside the box because it's 10^9 bits." Size is not a acceptable barrier to looking inside. Wiki, is huge and will get infineltly huge, yet I can look inside it, and see that "poison ivy causes rashes" or whatnot. The AGI will have enourmous complexity, I agree, but you should ALWAYS be able to look inside it. Not in the tradional sense of pages of code maybe or simple set of rules, but the AGI itself HAS to be able to generalize and tell what it is doing. So something like, I see these leafs that look like this, supply picture, can I pick them up safely, will generate a human readable output that can itself be debugged. Or asking about the process of doing something, will generate a possible plan that the AI would follow, and a human could say, no thats not right, and cause the AI to go back and reconsider with new possible information. We can always look inside the 'logic' of what the AGI is doing, we may not be able to directly change that ourselves easily. Doesn't that statement cease to apply as soon as the AGI starts optimizing it's own code? If the AGI is redesigning itself it will be changing before our eyes, faster than we can inspect it. You must be assuming a strictly controlled development system where the AGI proposes a change, humans inspect it for a week then tell the AGI to proceed with that change. I suspect you will only be able to do that in the very early development stages. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?list_id=303
Re: [agi] Natural versus formal AI interface languages
On 11/6/06, James Ratcliff wrote: In some form or another we are going to HAVE to have a natural language interface, either a translation program that can convert our english to the machine understandable form, or a simplified form of english that is trivial for a person to quickly understand and write. Humans use natural speech to communicate and to have an effective AGI that we can itneract with, it will have to have easy communication with us. That has been a critcal problem with all software since the beginning, a difficulty in the human computer interface. I go further to propose that as much knowledge information should be stored in easily recognizable natural language as well, only devolving into more complex forms where the cases warrant it, such as complex motor-sensor data sets, and some lower logic levels. Anybody remember short wave radio? The Voice of America does worldwide broadcasts in Special English. <http://www.voanews.com/specialenglish/about_special_english.cfm> Special English has a core vocabulary of 1500 words. Most are simple words that describe objects, actions or emotions. Some words are more difficult. They are used for reporting world events and describing discoveries in medicine and science. Special English writers use short, simple sentences that contain only one idea. They use active voice. They do not use idioms. -- There is also Basic English: <http://en.wikipedia.org/wiki/Basic_English> Basic English is a constructed language with a small number of words created by Charles Kay Ogden and described in his book Basic English: A General Introduction with Rules and Grammar (1930). The language is based on a simplified version of English, in essence a subset of it. Ogden said that it would take seven years to learn English, seven months for Esperanto, and seven weeks for Basic English, comparable with Ido. Thus Basic English is used by companies who need to make complex books for international use, and by language schools that need to give people some knowledge of English in a short time. Also see: <http://www.basic-english.org/> Basic English is a selection of 850 English words, used in simple structural patterns, which is both an international auxiliary language and a self-contained first stage for the teaching of any form of wider or Standard English. A subset, no unlearning. BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?list_id=303
Re: [agi] Natural versus formal AI interface languages
On 11/1/06, Charles D Hixson wrote: So. Lojban++ might be a good language for humans to communicate to an AI with, but it would be a lousy language in which to implement that same AI. But even for this purpose the language needs a "verifier" to insure that the correct forms are being followed. Ideally such a verifier would paraphrase the statement that it was parsing and emit back to the sender either an error message, or the paraphrased sentence. Then the sender would check that the received sentence matched in meaning the sentence that was sent. (N.B.: The verifier only checks the formal properties of the language to ensure that they are followed. It had no understanding, so it can't check the meaning.) This discussion reminds me of a story about the United Nations assembly meetings. Normally when a representative is speaking, all the translation staff are jabbering away in tandem with the speaker. But when the German representative starts speaking they all fall silent and sit staring at him. The reason is that they are waiting for the verb to come along. :) Billk - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/[EMAIL PROTECTED]
Re: [agi] SOTA
On 10/20/06, Richard Loosemore wrote: I would *love* to see those IBM folks put a couple of jabbering four-year-old children in front of that translation system, to see how it likes their 'low-intelligence' language. :-) Does anyone have any contacts on the team, so we could ask? I sent an email to Liang Gu on the IBM MASTOR project team (not really expecting a reply) :) and have just received this response. Sounds hopeful. BillK - Bill, Thanks for your interests on MASTOR. And your suggestion of MASTOR for Children is really great! It is definitely much more meaningful if MASTOR can not only help adults but also children communicate with each other around the world using different languages! Although recognizing Children's voice has been proved a very challenging task, the translation and text-to-speech techniques thus involved should be very similar to what we have now. We will seriously investigate the possibility of this approach and will send you a test link if we later developed a pilot system on the web. Regards and thanks again for your enthusiasm about MASTOR, Liang - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/[EMAIL PROTECTED]
Re: [agi] SOTA
On 10/20/06, Richard Loosemore <[EMAIL PROTECTED]> wrote: For you to blithely say "Most normal speaking requires relatively little 'intelligence'" is just mind-boggling. I am not trying to say that language skills don't require a human level of intelligence. That's obvious. That is what make humans human. But day-to-day chat can be mastered by children, even in a foreign language. Watch that video I referenced in my previous post, of an American chatting to a Chinese woman via a laptop running MASTOR software. <http://www.research.ibm.com/jam/speech_to_speech.mpg> Now tell me that that laptop is showing great intelligence to translate at the basic level of normal conversation. Simple subject object predicate stuff. Basic everyday vocabulary. No complex similes, metaphors, etc. There is a big difference between discussing philosophy and saying "Where is the toilet?" That is what I was trying to point out. Billk - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/[EMAIL PROTECTED]
Re: [agi] SOTA
On 10/19/06, Richard Loosemore <[EMAIL PROTECTED]> wrote: Sorry, but IMO large databases, fast hardware, and cheap memory ain't got nothing to do with it. Anyone who doubts this get a copy of Pim Levelt's "Speaking", read and digest the whole thing, and then meditate on the fact that that book is a mere scratch on the surface (IMO a scratch in the wrong direction, too, but that's neither here nor there). I saw a recent talk about an NLP system which left me stupified that so little progress has been made since 20 years ago. Having a clue about just what a complex thing intelligence is, has everything to do with it. Most normal speaking requires relatively little 'intelligence'. Adults who take young children on foreign holidays are amazed at how quickly the children appear to be chattering away to other children in a foreign language. They manage it for several reasons: 1) they don't have the other interests and priorities that adults have. 2) they use simple sentence structures and smallish vocabularies. 3) they discuss simple subjects of interest to children. The new IBM MASTOR system seems to be better than Babelfish. IBM are just starting on widespread commercial marketing of the system. Aiming at business travellers, apparently. MASTOR project description <http://domino.watson.ibm.com/comm/research.nsf/pages/r.uit.innovation.html> Here is a pdf file describing the MASTOR system in more detail <http://acl.ldc.upenn.edu/W/W06/W06-3711.pdf> Here is a 12MB mpg download of the system in use. Simple speech, but impressive. <http://www.research.ibm.com/jam/speech_to_speech.mpg> BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/[EMAIL PROTECTED]
Re: [agi] SOTA
On 10/19/06, Matt Mahoney wrote: - NLP components such as parsers, translators, grammar-checkers Parsing is unsolved. Translators like Babelfish have progressed little since the 1959 Russian-English project. Microsoft Word's grammar checker catches some mistakes but is clearly not AI. <http://www.charlotte.com/mld/charlotte/news/nation/15783022.htm> American soldiers bound for Iraq equipped with laptop translators Called the Two Way Speech-to-Speech Program, it's a translator that uses a computer to convert spoken English to Iraqi Arabic and vice versa. - If it is life-or-death, it must work pretty well. :) I believe this is based on the IBM MASTOR project. <http://domino.watson.ibm.com/comm/research.nsf/pages/r.uit.innovation.html> MASTOR's innovations include: methods that automatically extract the most likely meaning of the spoken utterance, store it in a tree structured set of concepts like "actions" and "needs", methods that take the tree-based output of a statistical semantic parser and transform the semantic concepts in the tree to express the same set of concepts in a way appropriate for another language; methods for statistical natural language generation that take the resultant set of transformed concepts and generate a sentence for the target language; generation of proper inflections by filtering hypotheses with an n-gram statistical language model; etc BillK - This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/[EMAIL PROTECTED]
Re: [agi] [META] Is there anything we can do to keep junk out of the AGI Forum?
On 7/26/06, Richard Loosemore wrote: I am beginning to wonder if this forum would be better off with a restricted membership policy. Richard Loosemore Davy Bartoloni - Minware S.r.l. wrote: > Which thing we want from a IA? , we want TRULY something? the doubt > rises me that nobody affidera' never to the words of a program, e' piu' easy > to entrust itself to a book. the book from the emergency, its words cannot > change, what there e' written e' nearly sure reliable. it is entrusted > to us of piu' to the handbook of the bimby in order making the mousse of It *might* sound a bit better in the original Italian??? But Babelfish / Google translate has made a real mess of it. BillK --- To unsubscribe, change your address, or temporarily deactivate your subscription, please go to http://v2.listbox.com/member/[EMAIL PROTECTED]
Re: [agi] Two draft papers: AI and existential risk; heuristics and biases
On 6/7/06, Eliezer S. Yudkowsky wrote: Because this is a young field, how much mileage you get out will be determined in large part by how much sweat you put in. That's the simple practical truth. The reasons why you do X are irrelevant given that you do X; they're "screened off", in Pearl's terminology. It doesn't matter how good your excuse is for putting off work on Friendly AI, or for not building emergency shutdown features, given that that's what you actually do. And this is the complaint of IT security professionals the world over; that people would rather not think about IT security, that they would rather do the minimum possible and just get it over with and go back to their day jobs. Who can blame them for such human frailty? But the result is poor IT security. This is the real world that you have to deal with. You cannot get the funding, or the time, to do the job properly, because there is always pressure to be the first to market. AGI is so tricky a problem that just getting it to work at all is regarded as a minor miracle. (Like the early days of computers and the internet). Implement first, we can always patch it afterwards. A much more worrying consideration, of course, is that the people with the most resources, DARPA (and the Chinese) want an AGI to help them kill their enemies. For defensive reasons only, naturally. When AGI is being developed as a weapon by massive government resources, AGI ethics and being Friendly doesn't even get into the specification. Following orders from the human owners does. BillK --- To unsubscribe, change your address, or temporarily deactivate your subscription, please go to http://v2.listbox.com/member/[EMAIL PROTECTED]
Re: [agi] Numenta: article on Jeff Hawkins' AGI approach
On 6/2/06, Ben Goertzel wrote: Mike You note that Numenta's approach seems oriented toward implementing an animal-level mind... I agree, and I do think this is a fascinating project, and an approach that can ultimately succeed... but I think that for it to succeed Hawkins will have to introduce a LOT of deep concepts that he is currently ignoring in his approach. Most critically he ignores the complex, chaotic dynamics of brain systems... I suppose part of the motivation for starting with animal mind is that the human mind is just a minor adjustment to the animal mind, which is sorta true genetically and evolutionarily But on the other hand, just because animal brains evolved into human brains, doesn't mean that every system with animal-brain functionality has similar evolve-into-human-brain potentiality Just from a computer systems design perspective, I think this project is admirable. I think it is safe to claim that all the big computer design disasters occurred because they tried to do too much all at once. ''We want it all, and we want it now!'. Ben may be correct in claiming that major elements are being omitted, but if they even get an animal level intelligence running, this will be a remarkable achievement. They will be world leaders and will learn a lot about designing such systems. Even if it cannot progress to higher levels of intelligence, the experience gained will set their technicians well on the road to the next generation design. BillK --- To unsubscribe, change your address, or temporarily deactivate your subscription, please go to http://v2.listbox.com/member/[EMAIL PROTECTED]