Hey Calum. Watch it! Pigs are very-very smart. They can sniff out truffles, and corpses from 6 feet under. :-)
I'm so itching to say it, but I won't. NOW can we get some ass into gear? Are we going to use this as a wakeup call, or not? Subject: Re: [agi] AI Protest in Texas From: a...@listbox.com Date: Wed, 18 Mar 2015 07:48:07 +0000 To: a...@listbox.com Ha! I like the Occupy Coke idea! You may not be a "PR guy" but you surely could be if you wanted to. I thought you believed AGI to be Good News, not a pig in need of better lipstick? Calum On 18 Mar 2015, at 02:21 AM, "Ben Goertzel via AGI" <a...@listbox.com> wrote: On Tue, Mar 17, 2015 at 11:13 PM, Robert Levy via AGI <a...@listbox.com> wrote: This is probably an important discussion, independent of the event that prompted it, but it turns out the protest at SXSW was a hoax/ viral marketing campaign staged to promote a new dating site. I wonder what's next.... "Was that a revolution in Russia we just saw, or just a large-scale advertisement for Kalashnikovs??" Will the border between political activity and marketing campaign blur even further? "Occupy your stomach with 'Occupy Cola' ,the true rebel's beverage of choice " ??? ben On Tue, Mar 17, 2015 at 8:06 AM, Calum Chace via AGI <a...@listbox.com> wrote: Very well said. (And Pandora's Brain does say it, as it happens...) But the pro-AGI community also needs to convince the public that the AGI we'll get will be a Friendly one. Calum On 17 March 2015 at 15:56, Ben Goertzel via AGI <a...@listbox.com> wrote: A problem is that careful, balanced discussions of difficult issues are boring and don't attract media attention Joel Pitt and I wrote a fairly thoughtful discussion of AGI safety issues a few years ago, http://jetpress.org/v22/goertzel-pitt.htm but of course our thoughts are more complex and nuanced, whereas a tweet from a billionaire comparing AI research to demon-summoning is a lot sexier... IMO, to get media attention sufficient to counteract the media's love of alarmism and doomsaying, the pro-AGI community would need to come forward very aggressively with the message that AGI is important for SAVING AND IMPROVING HUMAN LIVES ... for designing the next generation of medicines, for creating elder-care robots to make old age more livable, for extending healthspan for those who want it, for aiding the invention of new energy sources, for aiding in the fight against physical and cyber terrorism, and so forth.... "Don't worry too much, we'll be careful" is not a convincing counterargument -- a better counterargument to the Musks, Hawkings, Bostroms and Yudkowskys of the world is more like "Hey, I don't want your fear of science fiction bad guys to deny my grandma her life-extending, health-improving medicine and her robot friend, to eliminate my future of virtually unlimited energy and to put me at risk from terrorist attacks...." I.e. "DON'T LET THE LUDDITES KILL YOUR GRANDMA AND TAKE YOUR TOYS AWAY!! EMBRACE AI AND ROBOTS LIKE YOU'VE EMBRACED SMARTPHONES, AC POWER, THE INTERNET AND BIRTH CONTROL PILLS -- AND YOUR LIFE WILL BE BETTER -- " .... OK I'm semi-joking ;) ;p ... but unfortunately I think it's a mistake to overestimate the general public's appetite for rational, balanced discussion and thinking ;p ... When careful nuanced thinking on difficult issues it put out there, it tends to be vigorously ignored... -- Ben On Tue, Mar 17, 2015 at 8:01 PM, Benjamin Kapp via AGI <a...@listbox.com> wrote: If you think of governments as an artificial man (as was done by Aristotle and Hobbes amongst others) which is composed of humans who are the muscles (military, police), the intelligence (spys, scientists), the judging and planning (judges, politicians), etc.. In a way the state is a leviathan (a thing which has power to overawe any individual or group of individuals). And in this way AGI (or a super intelligence) already exists. On Tue, Mar 17, 2015 at 6:29 AM, Calum Chace via AGI <a...@listbox.com> wrote: Thanks Basile. I agree with Pitrat, although I might dial up the consideration of the downside possibility a touch. Hawking usually gets slightly mis-represented. He said that AGI could be either the best or the worst thing ever to happen to humanity. The "best" bit seems to get missed by both sides of the debate. So, my question is, what is the best way for people who think along these lines to try and steer the public debate on AGI? Alarmism is unhelpful, and hard to avoid. Secrecy won't work. Ben is tackling the issue head-on (as in the video he posted just now), but it's a hard debate to get right. Calum On 17 March 2015 at 11:17, Basile Starynkevitch <bas...@starynkevitch.net> wrote: On Tue, Mar 17, 2015 at 09:33:22AM +0100, Calum Chace via AGI wrote: > Steve > > I sympathise with your very understandable preference not to be targeted by > anti-AI crazies! > > What do you think is the best way to try and shape the growing public > debate about AGI? Following Bostrom's book, and the comments by Hawking, > Musk and Gates, a fair proportion of the general public is now aware that > AGI might arrive in the medium term, and that it will have a very big > impact. > > Some AI researchers seem to be responding by saying, "Don't worry, it can't > happen for centuries, if ever". No doubt some of them genuinely believe > that, but I wonder whether some are saying it in the (forlorn?) hope the > debate will go away. It won't. In fact I suspect that the new Avengers > movie will kick it up a level. > > Others are saying, "Don't worry, AGI cannot and will not harm humans." To > my mind (and I realise I may be in a small minority here on this) that is > hard to be certain about - as Bostrom demonstrated. > > Yet others are saying, "AI researcher will solve the problem long before > AGI arrives, and it's best not to worry everyone else in the meantime." > That seems a dangerous approach to me. If the public ever feels (rightly > or wrongly) that things have been hidden from them, they will react badly. > > But I do definitely sympathise with the desire not to be targeted by > crazies, or to be vilified by journalists who have half-understood the > situation! > [...] > >> > ------------------------------------------- > >> > AGI > >> > Archives: https://www.listbox.com/member/archive/303/=now [....] I would suggest reading J.Pitrat's december 2014 blog entry on that subject. J.Pitrat is probably not subscribing to that list, i so I am blind-carbon-copying him. http://bootstrappingartificialintelligence.fr/WordPress3/2014/12/not-developing-an-advanced-artificial-intelligence-could-spell-the-end-of-the-human-race/ He is explaining that "Not developing an advanced artificial intelligence could spell the end of the human race" and I believe he has a point. Of course AGI researchers should be careful. Regards -- Basile Starynkevitch http://starynkevitch.net/Basile/ -- Regards Calum AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription -- Ben Goertzel, PhD http://goertzel.org "The reasonable man adapts himself to the world: the unreasonable one persists in trying to adapt the world to himself. Therefore all progress depends on the unreasonable man." -- George Bernard Shaw AGI | Archives | Modify Your Subscription -- Regards Calum AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription -- Ben Goertzel, PhD http://goertzel.org "The reasonable man adapts himself to the world: the unreasonable one persists in trying to adapt the world to himself. Therefore all progress depends on the unreasonable man." -- George Bernard Shaw AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com