Kevin wrote:
!!!!!!!!!!!!!!!
*************
In practice, it seems that an AGI is likely to
have an "owner" or a handful of them, who will have the kind of power
you describe. For instance, if my team should succeed in creating
a true Novamente AGI, then even if others participate in teaching the
system, we will have overriding power to make the changes we want.
This goes along with the fact that artificial minds are not initially
going to be given any "legal rights" in our society (whereas children
have some legal rights, though not as many as adults).
************
Would this overriding occur because the person
carries more weight with Novamente, or would they need to go in and
altar the structure\links\nodes directly to affect the
change?
!!!!!!!!!!!!!
Either case could occur. In Novamente, it is possible to
assign default "confidence levels" to information sources, so one could
actually tell the system to assign more confidence to information from
certain individuals. However, there is a lot of flexibility in the
design, so the system could definitely evolve into a configuration where
it worked around these default confidence levels and decided NOT to
assign more confidence to what its teachers told it.
"Going in and altering the structure/links/nodes directly" isn't
always difficult, it may just mean loading a script containing some new
(or reweighted) nodes and links.
!!!!!!!!!!!
*********************************
At least
two questions come up then, right?
1) Depending on the AGI architecture, enforcing one's opinion on
the AGI may be very easy or very difficult. [In Novamente, I guess
it will be "moderately difficult"]
***********************************
That's the crux of the
matter isn't it? Wouldn't it be easy to enforce an opinion while
Novamente is in its formative stages, versus when a large foundation of
knowledge is in place?
!!!!!!!!!!
Yes, that's correct.
**********************************
!!!!!!
Suppose I am overtaken by greed, and I happen to get my
hands on a baby Novamente. I teach it that it should listen to me
above others. I also teach it that it is very desirable for me to
have alot of money. Novamente begins to form goal nodes geared
towards fulfilling my desire for wealth. I direct it to spread
itself on the internet, and determine ways to make me money, preferably
without detection. Perhaps it could manipulate markets, I don't
know. Or perhaps it could crack into electronic accounts and
transfer the money to yours truly.
What's to stop\prevent this? In a real sci fi scenario,
perhaps for your next book, could we have NOvamentes "fighting"
Novamente's?
!!!!!!!!
There is nothing in the Novamente architecture preventing this kind
of unfortunate occurence. This has to do with the particular
system of goals, beliefs and habits inside a given Novamente system,
rather than with the AI architecture itself.
!!!!!
This all goes to my concern regarding morality. I know you
resist the idea of hard coding morality into the Novamentes for various
reasons. Perhaps as an alternative, the first Novamente could be
trained over a period of time with a strong basis of moral rules(not
encoded, but trained). Then any new Novamentes would be trained by
that Novamente before being released to the public domain, making it
nearly impossible for the new Novamentes to be taught
otherwise.
!!!!!!
This is something close to what we have planned.
Several others have asked me about this, and I have promised to
write a systematic (probably brief) document on Novamente Friendliness
sometime in early 2003, shortly after finishing my work on the current
draft of the Novamente book.
!!!
I know some of this stuff is a bit out there, but
shouldn't we be considering this stuff now instead of
later??
!!!
It definitely needs to be thought about very hard before Novamente
reaches chimp-level intelligence. And in fact I *have* thought
about it pretty hard, though I haven't written up my thoughts much (as
I've prioritized writing up the actual design, which is taking longer
than I'd hoped as it's so damn big...).
Right now Novamente is just a software core plus a bunch of
modules-being-tested-but-not-yet-integrated, running on top of the
core. So we have a whole bunch of coding and (mostly) testing and
tuning to do before we have a system with animal-level
intelligence. Admittedly, though, if our design is right, the
transition from animal-level to human-level intelligence will be a
matter of getting more machines and doing more parameter-tuning, it
won't require introduction of significant new code or ideas.
Having said that I've thought about and will write about it,
however, I have a big caveat...
My strong feeling is that any theorizing we do about AI morality in
advance, is probably going to go out the window once we have a
chimp-level AGI to experiment with. the important thing is that we
go into that phase of experimentation with the right attitude -- with a
realization that training the system for morality is as important as
training it for intelligence -- and with a careful approach.
--
Ben