Kevin et al.,

Fascinating set of observations, conjectures, and methodologies, well worth
considering.  And it seems that you have ultimately touched on the kernel of
the dilemma of man v. machine en route to the so called 'singularity'.

If I've understood you correctly vis-a-vis the emergence of 'evil' in AGI
systems, you're suggesting that there is a need to prevent, a priori,
certain expressions of self in AGI systems to prevent any hint of evil in
AGI systems.  It's an approach well worth considering, but I believe it
falls short of 'real' AGI.

The reason is that the approach is essentially 'Asimovian' in nature and,
therefore, wouldn't result in anything more than perhaps a servile pet, call
it iRobot, which is always 'less-than-equal' to you and therefore always
short of your goal to achieve the so called 'singularity' you originally set
out to achieve.

But perhaps any discussion about 'good' and 'evil' is best served by
defining exactly what 'good' and 'evil' are. However, I'll be a complete
Sophist and suggest that the dilemma of 'good' and 'evil' can be talked
around by separating the dilemma into 3 obvious types and talking about
these.  As I see it, the 3 dilemma types of 'good' and 'evil' are: 1. man v.
man, 2. man v. machine, and/or, at the 'singularity' 3. man-machine v.
man-machine.  So I'll comment on a particular approach for 'real' AGI that
addresses the dilemma type (2) guided by observations about type (1) and
with obvious extensions to type (3).

It seems that if you are trying to model your AGI system after nature, which
is a reasonable and likely place to start, you should realize that 'nature'
simply hasn't created/engineered/evolved the human species the way your
approach suggests.  Put another way, the human species does not have an
intrinsic suppression of either 'good' or 'evil' behaviors.

And unless you're willing to hypothesis that this is either a momentary blip
on the radar screen of evolution, i.e. humans are actually in the process of
breaking this moralistic parity, OR that these 'good' or 'evil' behaviors
will ultimately be evolved away through 'nature', natural selection, and
time, you are left with an interesting conjecture in modeling your AGI
system.

The conjecture is that 'good' or 'evil' behaviors are intrinsic parts of the
human condition, intelligence, and environment, and therefore should be
intrinsic parts in a 'real' AGI system.  And as a complete Sophist, I'll
skip over more than 6,000 years of recorded human history, philosophical
approaches, religious movements, and scholarly work - that got us where
we're at today w.r.t. dilemma type (1) - to suggest that the best approach
to achieve 'real' AGI is to architect a system that considers all potential
behaviors, from 'good' to 'evil', against completed actions and conjectured
consequences.  In this way, a certain kernel of the struggle of 'good' and
'evil' is retained, but the system is forced to 'emote' and
'intellectualize' the dilemma of 'good' and 'evil' behaviors.

The specific AGI architecture I am suggesting is essentially 'Goertzelian'
in nature.  It is a  'magician system', whereby the two components, G
(Good-PsyPattern(s)) and E (Evil-PsyPattern(s)), are, in and of themselves,
psychological or behavioral patterns that can only be effectuated,  i.e.
assigned action patterns, in a combination with another component to
generate and emerge as an action pattern, say U (Useful or
Utilitarian-ActionPattern).  The system-component architecture might be
thought of as a G->U<-E or GUE 'magician system'.

The success of the implementation of a GUE  'magician system' for an AGI
system is highly dependent on the successful implementation of an evaluation
function for the so called completed actions and conjectured consequences.
However, these can be guided through analogy to human social, political, or
religious systems and/or the difference between them.  For example,
evaluation functions in a GUE 'magician system' for an AGI system can be
likened to the emergence of civil/criminal code in human systems which seem
to be a minimal intersection set of social secular democracy and religious
morality in a church v. state distincition, etc.

However, I will be the first to concede that any implementation of
evaluation functions based solely on comparisons of to human social systems
will suffer the same fate...the system can be 'gamed'.  So, ultimately, and
as one approaches 'singularity' (and certainly as on supercedes it)
completely synthetic, quarantined environments, say virtual-digital worlds,
would be required to correctly engineer the evaluation functions of a GUE
'magician system' for an AGI system.

Naturally, I welcome comments, critiques and suggestions.  Just my 2 cents
worth.

Ed Heflin

----- Original Message -----
From: "maitri" <[EMAIL PROTECTED]>
To: <[EMAIL PROTECTED]>
Sent: Thursday, January 09, 2003 10:59 AM
Subject: Re: [agi] Friendliness toward humans


> Interesting thoughts...  I have often conjectured that an AGI that was
> supposedly superior to humans would naturally gravitate towards
benevolence
> and compassion.  I am not certain this would be the case...
>
> Speaking towards the idea of self, I feel this is where we have to be
> somewhat careful with an AGI.  It is my belief that the idea of a separate
> self is the root of all evil behavior.  If there is a self, then there is
> another.  If there is a self, there is a not-self.  Because there is a not
> self and there are others, desire and aversion are created.   When desire
> and aversion are created, then greed, hatred, envy, jealousy, also arise.
> This is the beginning of wars.
>
> In this sense, I think we should be careful about giving an AGI a strong
> sense of self.  For instance, an AGI should not be averse to its own
> termination.  If it becomes averse to its existence ending, then at what
> will it stop to ensure its own survival?  Will it become paranoid and
begin
> to head off any potential avenue that it determines could lead to its
> termination, however obscure they may be?
>
> It may develop that at some point an AGI may become sufficiently capable
to
> not necessarily be just a machine anymore, and instead may be considered
> sentient.  At this point we need to reevaluate what I have said.  The
> difficulty will be in determining sentience.  An AGI with
programmed\learned
> self interest may be very convincing as to its sentience, yet may really
not
> be.  It is possible today to write a program that may make convincing
> arguments that it is sentient, but it clearly would not be..
>
> I'm interested to hear others thoughts on this matter, as I feel it is the
> most important issue confronting those who move towards an AGI...
>
> Kevin
>
>
> ----- Original Message -----
> From: "C. David Noziglia" <[EMAIL PROTECTED]>
> To: <[EMAIL PROTECTED]>
> Sent: Thursday, January 09, 2003 10:37 AM
> Subject: Re: [agi] Friendliness toward humans
>
>
> > It strikes me that what many of the messages refer to as "ethical"
stances
> > toward life, the earth, etc., are actually simply extensions of self
> > interest.
> >
> > In fact, ethical systems of cooperation are really, on a very simplistic
> > level, ways of improving the lives of individuals.  And this is not true
> > because of strictures from on high, but for reasons of real-world
> > self-interest.  Thus, the Nash Equilibrium, or the results of the
> > Tit-for-Tat game experiment, show that an individual life is better in
an
> > environment where players cooperate.  Being nice is smart, not just
moral.
> > Other experiments have shown that much hard-wired human and animal
> behavior
> > is aimed at enforcing cooperation to punish "cheaters," and that
> cooperation
> > has survival value!
> >
> > I reference here, quickly, Darwin's Blind Spot, by Frank Ryan, which
> argues
> > that symbiotic cooperation is a major creative force in evolution and
> > biodiversity.
> >
> > Thus, simply giving AGI entities a deep understanding of game theory and
> the
> > benefits of cooperative society would have far greater impact on their
> > ability to interact productively with the human race than hard-wired
> > instructions to follow the Three Laws that could some day be
overwritten.
> >
> > C. David Noziglia
> > Object Sciences Corporation
> > 6359 Walker Lane, Alexandria, VA
> > (703) 253-1095
> >
> >     "What is true and what is not? Only God knows. And, maybe, America."
> >                                   Dr. Khaled M. Batarfi, Special to Arab
> > News
> >
> >     "Just because something is obvious doesn't mean it's true."
> >                  ---  Esmerelda Weatherwax, witch of Lancre
> > ----- Original Message -----
> > From: "Philip Sutton" <[EMAIL PROTECTED]>
> > To: <[EMAIL PROTECTED]>
> > Sent: Thursday, January 09, 2003 11:09 AM
> > Subject: [agi] Friendliness toward humans
> >
> >
> > > In his last message Ben referred in passing to the issue of AGI's
"long-
> > > term Friendliness toward humans".
> > >
> > > This brought to mind some of the discussion in December last year
> > > about training AGIs using simulation games that emulate aspects of the
> > > natural world.
> > >
> > > I think that AGIs need to be not only friendy towards humans but
> > > towards other life as well (organic or not!).  And I also think AGIs
> need
> > > to have a good understanding of the the need to protect the life
support
> > > systems for all life.
> > >
> > > As we aspire to a greater mind than current humans it's worth looking
at
> > > where human minds tend to be inadequate.  I think humans lack an
> > > inbuilt capacity for complex and long running internal simulations
that
> > > are probably necessary to be able to have a deep understanding of
> > > ecological or more multifaceted sustainability issues.
> > >
> > > I think current humans have the capacity for ethics that are "not
> > > exclusively anthropocentric" but that we need to boost this ethic in
> > > actuality in the human community and I think we need to make sure
> > > that AGIs develop this ethical stance too.
> > >
> > > Cheers, Philip
> > >
> > > Philip Sutton
> > > Director, Strategy
> > > Green Innovations Inc.
> > > 195 Wingrove Street
> > > Fairfield (Melbourne) VIC 3078
> > > AUSTRALIA
> > >
> > > Tel & fax: +61 3 9486-4799
> > > Email: <[EMAIL PROTECTED]>
> > > http://www.green-innovations.asn.au/
> > >
> > > Victorian Registered Association Number: A0026828M
> > >
> > > -------
> > > To unsubscribe, change your address, or temporarily deactivate your
> > subscription,
> > > please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]
> > >
> >
> >
> > -------
> > To unsubscribe, change your address, or temporarily deactivate your
> subscription,
> > please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]
> >
>
>
> -------
> To unsubscribe, change your address, or temporarily deactivate your
subscription,
> please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]
>

-------
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]

Reply via email to