[agi] Building a safe AI

2003-02-20 Thread Peter Voss


http://www.optimal.org/peter/siai_guidelines.htm

Peter




-Original Message-
From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED]]On
Behalf Of Ben Goertzel

I would recommend Eliezer's excellent writings on this topic if you don't
know them, chiefly www.singinst.org/CFAI.html .  Also, I have a brief
informal essay on the topic, www.goertzel.org/dynapsyc/2002/AIMorality.htm ,
although my thoughts on the topic have progressed a fair bit since I wrote
that.  Note that I don't fully agree with Eliezer on this stuff, but I do
think he's thought about it more thoroughly than anyone else (including me).

It's a matter of creating an initial condition so that the trajectory of the
evolving AI system (with a potentially evolving goal system) will have a
very high probability of staying in a favorable region of state space ;-)

---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]



Re: RE: [agi] Building a safe AI

2003-02-20 Thread peiwang
Thanks to Peter for starting this discussion and to Ben for following up. This seems 
to me a more constructive way to talk about Friendly AI.

Now it's my turn to comment on the 8 Guidelines, according to my NARS design (for 
people who have no idea what I'm talking about, see 
http://www.cogsci.indiana.edu/farg/peiwang/papers.html).

1. Friendliness-topped goal system

Though in NARS initial goals can be specified by the designer, they are not the 
supergoals described. Due to insufficient knowledge and resources, there is no way 
to guarantee the consistency (or even the relevance) of the subgoals and initial 
goals.  Therefore, to have friendliness as the only initial goal does not make the 
system friendly.

2. Cleanly causal goal system

Since all causal knowledge is empirical, and therefore maybe biased, the above problem 
is also here. Though NARS indeed has a “cleanly causal goal system”, where all 
goals are derived from the initial goals according to the current causal knowledge, 
this mechanism cannot guarantee friendliness.

3. Probabilistic supergoal content

In NARS all empirical knowledge has uncertainty, so this is not an issue for me.

4. Acquisition of Friendliness sources

As an adaptive system, NARS will learn various types of knowledge from human beings. I 
don’t see anything special about ethical knowledge on this issue.

5. Causal validity semantics.

Again, this is just causal inference, which can be applied to moral issues, but is not 
limited there.

6. Injunctions.

With insufficient knowledge and resources, it is impossible  “to prevent possible 
negative outcomes”.  Of course, the system will try to avoid them as much as 
possible.

7. Self-modeling of fallibility

as above.

8. Controlled ascent

I don’t think it is possible to implement an improvements counter. An adaptive 
system is changing all the time. I cannot tell when NARS is changing faster than 
usual, and if it is, whether this is an alarm signal. Furthermore, whether a change is 
an improvement or not is usually a subjective judgment based on feedback over a 
certain period. Something may look like an improvement now, but lead to a big trouble 
in the future. 

In summary, I agree with some of the “Guidelines”, though I think they are 
actually about AI, and are neutral to the friendliness issue. Of course, high 
intelligence can make a system friendlier, but it can also make the system more 
dangeours.  To select proper initial goals and to provide proper feedback will reduce 
the danger, but cannot eliminate it. 

I post this reply just to show that I fully agree that AI can be dangerous, and we 
should be ready to deal with the problems when designing an AI system. However, at the 
current time, moral discussions contribute little to the real work. I haven’t seen 
how the design of NARS should be changed to make it friendlier.

Pei

---Original Message---
From: Ben Goertzel [EMAIL PROTECTED]
Sent: 02/20/03 12:01 PM
To: [EMAIL PROTECTED]
Subject: RE: [agi] Building a safe AI

 
 
Peter,

At the end of the page you reference, you list

The Guidelines' eight design recommendations in the light of my theory of
mind/ intelligence:
1.  ...
2.  ...
...


All of your comments in that section apply to Novamente without
significant
modification.  Although in detail your design is of course quite different
from Novamente.

-- Ben


 -Original Message-
 From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED]]On
 Behalf Of Peter Voss
 Sent: Thursday, February 20, 2003 11:48 AM
 To: [EMAIL PROTECTED]
 Subject: [agi] Building a safe AI




 http://www.optimal.org/peter/siai_guidelines.htm

 Peter




 -Original Message-
 From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED]]On
 Behalf Of Ben Goertzel

 I would recommend Eliezer's excellent writings on this topic if you
don't
 know them, chiefly www.singinst.org/CFAI.html .  Also, I have a brief
 informal essay on the topic,
 www.goertzel.org/dynapsyc/2002/AIMorality.htm ,
 although my thoughts on the topic have progressed a fair bit since I
wrote
 that.  Note that I don't fully agree with Eliezer on this stuff, but I
do
 think he's thought about it more thoroughly than anyone else
 (including me).

 It's a matter of creating an initial condition so that the
 trajectory of the
 evolving AI system (with a potentially evolving goal system) will have a
 very high probability of staying in a favorable region of state space
;-)

 ---
 To unsubscribe, change your address, or temporarily deactivate
 your subscription,
 please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]


---
To unsubscribe, change your address, or temporarily deactivate your
subscription, 
please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]


---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]