Interesting. I believe that we have a fundamental disagreement. I
would argue that the semantics *don't* have to be distributed. My
argument/proof would be that I believe that *anything* can be described
in
words -- and that I believe that previous narrow AI are brittle because
they
don't have both a) closure over the terms that they use and b) the
ability
to learn the meaning if *any* new term (traits that I believe that humans
have -- and I'm not sure at all that the "intelligent" part of humans
have
distributed semantics). Of course, I'm also pretty sure that my belief
is
in the minority on this list as well.
I believe that an English system with closure and learning *is* going
to
be a complex system and can be grounded (via the closure and interaction
with the real world). And scalable looks less problematic to me with
symbols than without.
We may be different enough in (hopefully educated) opinions that this
e-mail may not allow for a response other than "We shall see" but I would
be
interested, if you would, in hearing more as to why you believe that
semantics *must* be distributed (though I will immediately concede that
it
will make them less hackable).
Mark
----- Original Message -----
From: "Richard Loosemore" <[EMAIL PROTECTED]>
To: <agi@v2.listbox.com>
Sent: Monday, October 01, 2007 8:36 PM
Subject: **SPAM** Re: [agi] Religion-free technical content
> Mark Waser wrote:
>>> And apart from the global differences between the two types of AGI,
>>> it
>>> would be no good to try to guarantee friendliness using the kind of
>>> conventional AI system that is Novamente, because inasmuch as general
>>> goals would be encoded in such a system, they are explicitly coded as
>>> "statement" which are then interpreted by something else. To put it
>>> crudely (and oversimplify slightly) if the goal "Be empathic to the
>>> needs of human beings" were represented just like that, as some kind
>>> of
>>> proposition, and stored at a particular location, it wouldn't take
>>> much
>>> for a hacker to get inside and change the statement to "Make
>>> [hacker's
>>> name] rich and sacrifice as much of humanity as necessary". If that
>>> were to become the AGI's top level goal, we would then be in deep
>>> doodoo. In the system I propose, such events could not happen.
>>
>> I think that this focuses on the wrong aspect. It is not the fact
>> that
>> the goal is explicitly encoded as a statement that is a problem -- it
>> is
>> the fact that it is in only one place that is dangerous. My
>> assumption
>> is that your system basically build it's base constraints from a huge
>> number of examples and that it is distributed enough that that it
>> would
>> be difficult if not impossible to maliciously change enough to cause a
>> problem. The fact that you're envisioning your system as not having
>> easy-to-read statements is really orthogonal to your argument and a
>> system that explicitly codes all of it's constraints as readable
>> statements but still builds it's base constraints from a huge number
>> of
>> examples should be virtually as incorruptible as your system (with the
>> difference being security by obscurity -- which is not a good thing to
>> rely upon and also means that your system is less comprehensible).
>
> Mark,
>
> You have put your finger on one aspect of the proposal that came up, in
> a
> slightly different way, when Jef Allbright started talking about
> pragmatics: the "semantics" of the system. This is the hardest
> feature
> to explain in a short space.
>
> I really did consciously mean to have both things, not just distributed
> representation of the constraints, but also the fact that the semantics
> of
> the system is distributed. This distributed, semi-opaque semantics is
> what I meant by talking about the propositions not being explicitly
> encoded, above, and what I also was referring to in my comment to Jef.
>
> If the basic knowledge units ("atoms") of the system develop as a
> result
> of learning mechanisms + real world interaction (which together make
> them
> grounded), then the meaning of any given atom is encoded in the whole
> web
> of connections between it and the other atoms, and also by the
> mechanisms
> that browse on (/use, /modify) these atoms. It is not easy to point to
> an
> atom and say exactly what it does.
>
> This is not an optional part of the framework: it is crucial. It is
> the
> main reason why the system has some complexity. It is also the reason
> why
> the system can be properly grounded and is scalable (which is what,
> with
> an ordinary, conventional AI system, cannot be done because of the
> complex
> systems problem).
>
> In a sense the system is less comprehensible, but this is only a matter
> of
> degree. I don't think it makes any practical difference to our
> attempts
> to govern its behavior. It is going to be comprehensible enouigh that
> we
> can put hooks in for monitoring purposes.
>
> The great benefit of this way of doing things is that, once the system
> has
> matured to adulthood, it cannot be hacked: you cannot just write a
> worm
> to go around hunting for constraints and modifying them in a regular
> way
> (as you might be able to do with ordinary distributed constraints,
> where
> the semantics of each individual atom is well defined enough that you
> can
> make a clean edit), because if you tried to do this you would
> destabilize
> the whole thing and turn it into a gibbering wreck. It would stop
> working
> ... and the effect would be so dramatic that we (and it) could easily
> set
> up automatic shutdown mechanisms to intervene in such a case.
>
>
>
>
> Richard Loosemore
>
>
>
> -----
> This list is sponsored by AGIRI: http://www.agiri.org/email
> To unsubscribe or change your options, please go to:
> http://v2.listbox.com/member/?&
>
-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?&