On 10/2/07, Vladimir Nesov <[EMAIL PROTECTED]> wrote:
> On 10/2/07, Mark Waser <[EMAIL PROTECTED]> wrote:
> > You misunderstood me -- when I said robustness of the goal system, I meant
> > the contents and integrity of the goal system, not the particular
> > implementation.
>
> I meant that too - and I didn't mean to imply this distinction.
> Implementation of goal system, or 'goal system itself' are both in my
> argument can be represented as text written in natural language, that
> is in rather faulty way. 'Goal system as what it meant to be' is what
> intelligent system tries to achieve.
>
> >
> > I do however continue to object to your phrasing about the system
> > recognizing influence on it's goal system and preserving it.  Fundamentally,
> > there are only a very small number of "Thou shalt not" supergoals that need
> > to be forever invariant.  Other than those, the system should be able to
> > change it's goals as much as it likes (chocolate or strawberry?  excellent
> > food or mediocre sex?  save starving children, save adults dying of disease,
> > or go on vacation since I'm so damn tired from all my other good works?)
>
> This is just substitution of levels of abstraction. Programs on my PC
> run on a fixed hardware and are limited by its capabilities, yet they
> can vary greatly. Plus, intelligent system should be able to integrate
> impact of goal system on multiple levels of abstraction, that is it
> can infer level of strictness in various circumstances (which
> interface with goal system through custom-made abstractions).
>
>
> > A quick question for Richard and others -- Should adults be allowed to
> > drink, do drugs, wirehead themselves to death?
> >
> > ----- Original Message -----
> > From: "Vladimir Nesov" <[EMAIL PROTECTED]>
> > To: <agi@v2.listbox.com>
> > Sent: Tuesday, October 02, 2007 9:49 AM
> > Subject: **SPAM** Re: [agi] Religion-free technical content
> >
> >
> > > But yet robustness of goal system itself is less important than
> > > intelligence that allows system to recognize influence on its goal
> > > system and preserve it. Intelligence also allows more robust
> > > interpretation of goal system. Which is why the way particular goal
> > > system is implemented is not very important. Problems lie in rough
> > > formulation of what goal system should be (document in English is
> > > probably going to be enough) and in placing the system under
> > > sufficient influence of its goal system (so that intelligent processes
> > > independent on it would not take over).
> > >
> > > On 10/2/07, Mark Waser <[EMAIL PROTECTED]> wrote:
> > >> The intelligence and goal system should be robust enough that a single or
> > >> small number of sources should not be able to alter the AGI's goals;
> > >> however, it will not do this by recognizing "forged communications" but
> > >> by
> > >> realizing that the aberrant goals are not in congruence with the world.
> > >> Note that many stupid and/or greedy people will try to influence the
> > >> system
> > >> and it will need to be immune to them (or the solution will be worse than
> > >> the problem).

Argh!  "Goal system" and "Friendliness" are roughly the same sort of
confusion.  They are each modelable only within a ***specified***,
encompassing context.

In more coherent, modelable terms, we express our evolving nature,
rather than strive for "goals."

- Jef

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=8660244&id_secret=48954318-fd558f

Reply via email to