Great write up. My special interest is AI friendliness so I would like to comment on 11.
CEV is a concept that avoids answering the question of what friendliness is by letting an advanced AI figure out what good might be. Doing so makes endowing an AI implementation with friendliness not feasible. CEV is circular. See the following core sentence for example: "...if we knew more, thought faster, were more the people we wished we were, had grown up farther together; where the extrapolation converges rather than diverges, where our wishes cohere rather than interfere; extrapolated as we wish that extrapolated, interpreted as we wish that interpreted..." Simplified: "If we were better people we were better people." True - but not adding value as key concepts such as 'friendliness', 'good', 'better' and 'benevolence' remain undefined. In my recent book (see www.Jame5.com) I take the definition of friendliness further by grounding key terms such as 'good' and 'friendly'. If you rather not read my complete 45'000 word book I suggest focusing on the end of chapter 9 until 12. Those sum up the key concepts. Further I will post a 7 page paper (hopefully today) that further condenses the core ideas of what benevolence means and how hard goals for a friendly AI can be derived from those ideas. Kind regards, Stefan On 10/26/07, Kaj Sotala <[EMAIL PROTECTED]> wrote: > > Can be found at http://www.saunalahti.fi/~tspro1/objections.html . > > Answers the following objections: > > 1: There are limits to everything. You can't get infinite growth > 2: Extrapolation of graphs doesn't prove anything. It doesn't show > that we'll have AI in the future. > 3: A superintelligence could rewrite itself to remove human tampering. > Therefore we cannot build Friendly AI. > 4: What reason would a super-intelligent AI have to care about us? > 5: The idea of a hostile AI is anthropomorphic. > 6: Intelligence is not linear. > 7: There is no such thing as a human-equivalent AI. > 8: Intelligence isn't everything. An AI still wouldn't have the > resources of humanity. > 9: It's too early to start thinking about Friendly AI > 10: Development towards AI will be gradual. Methods will pop up to deal > with it. > 11: "Friendliness" is too vaguely defined. > 12: What if the AI misinterprets its goals? > 13: Couldn't AIs be built as pure advisors, so they wouldn't do > anything themselves? That way, we wouldn't need to worry about > Friendly AI. > 14: Machines will never be placed in positions of power. > > Constructive criticism welcome, as always. > > > -- > http://www.saunalahti.fi/~tspro1/ | http://xuenay.livejournal.com/ > > Organizations worth your time: > http://www.singinst.org/ | http://www.crnano.org/ | http://lifeboat.com/ > > ----- > This list is sponsored by AGIRI: http://www.agiri.org/email > To unsubscribe or change your options, please go to: > http://v2.listbox.com/member/?& > -- Stefan Pernar 3-E-101 Silver Maple Garden #6 Cai Hong Road, Da Shan Zi Chao Yang District 100015 Beijing P.R. CHINA Mobil: +86 1391 009 1931 Skype: Stefan.Pernar ----- This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=4007604&id_secret=57757116-051544
