Richard,
As I see it, in this long message you have given a conceptual sketch
of an AI design including a motivational subsystem and a cognitive
subsystem, connected via a complex network of continually adapting
connections. You've discussed the way such a system can potentially
build up a self-model involving empathy and a high level of awareness,
and stability, etc.
All this makes sense, conceptually; though as you point out, the story
you give is short on details, and I'm not so sure you really know how
to cash it out in terms of mechanisms that will actually function
with adequate intelligence ... but that's another story...
However, you have given no argument as to why the failure of this kind
of architecture to be stably Friendly is so ASTOUNDINGLY UNLIKELY as
you claimed in your original email. You have just argued why it's
plausible to believe such a system would probably have a stable goal
system. As I see it, you did not come close to proving your original
claim, that
The motivational system of some types of AI (the types you would
classify as tainted by complexity) can be made so reliable that the
likelihood of them becoming unfriendly would be similar to the
likelihood of the molecules of an Ideal Gas suddenly deciding to split
into two groups and head for opposite ends of their container.
I don't understand how this extreme level of reliability would be
achieved, in your design.
Rather, it seems to me that the reliance on complex, self-organizing
dynamics makes some degree of indeterminacy in the system almost
inevitable, thus making the system less than absolutely reliable.
Illustratng this point, humans (who are complex dynamical systems) are
certainly NOT reliable in terms of Friendliness or any other subtle
psychological property...
-- Ben G
On 10/25/06, Richard Loosemore [EMAIL PROTECTED] wrote:
Ben Goertzel wrote:
Loosemore wrote:
The motivational system of some types of AI (the types you would
classify as tainted by complexity) can be made so reliable that the
likelihood of them becoming unfriendly would be similar to the
likelihood of the molecules of an Ideal Gas suddenly deciding to split
into two groups and head for opposite ends of their container.
Wow! This is a vey strong hypothesis I really doubt this
kind of certainty is possible for any AI with radically increasing
intelligence ... let alone a complex-system-type AI with highly
indeterminate internals...
I don't expect you to have a proof for this assertion, but do you have
an argument at all?
ben
Ben,
You are being overdramatic here.
But since you ask, here is the argument/proof.
As usual, I am required to compress complex ideas into a terse piece of
text, but for anyone who can follow and fill in the gaps for themselves,
here it is. Oh, and btw, for anyone who is scarified by the
psychological-sounding terms, don't worry: these could all be cashed
out in mechanism-specific detail if I could be bothered -- it is just
that for a cognitive AI person like myself, it is such a PITB to have to
avoid such language just for the sake of political correctness.
You can build such a motivational system by controlling the system's
agenda by diffuse connections into the thinking component that controls
what it wants to do.
This set of diffuse connections will govern the ways that the system
gets 'pleasure' -- and what this means is, the thinking mechanism is
driven by dynamic relaxation, and the 'direction' of that relaxation
pressure is what defines the things that the system considers
'pleasurable'. There would likely be several sources of pleasure, not
just one, but the overall idea is that the system always tries to
maximize this pleasure, but the only way it can do this is to engage in
activities or thoughts that stimulate the diffuse channels that go back
from the thinking component to the motivational system.
[Here is a crude analogy: the thinking part of the system is like a
table ontaining a complicated model landscape, on which a ball bearing
is rolling around (the attentional focus). The motivational system
controls this situation, not be micromanaging the movements of the ball
bearing, but by tilting the table in one direction or another. Need to
pee right now? That's because the table is tilted in the direction of
thoughts about water, and urinary relief. You are being flooded with
images of the pleasure you would get if you went for a visit, and also
the thoughts and actions that normally give you pleasure are being
disrupted and associated with unpleasant thoughts of future increased
bladder-agony. You get the idea.]
The diffuse channels are set up in such a way that they grow from seed
concepts that are the basis of later concept building. One of those
seed concepts is social attachment, or empathy, or imprinting the
idea of wanting to be part of, and approved by, a 'family' group. By
the time the system is mature, it has