On 15/05/07, Matt Mahoney <[EMAIL PROTECTED]> wrote:

We would all like to build a machine smarter than us, yet still be able to
predict what it will do.  I don't believe you can have it both ways.  And
if
you can't predict what a machine will do, then you can't control it.  I
believe this is true whether you use Legg's definition of universal
intelligence or the Turing test.


We might not be able to predict what the superintelligent machine is going
to say, but still be able to impose constraints on what it is going to do.
For a start, it would probably unwise to give such a machine any motivation
at all, other than the motivation of the ideal, disinterested scientist, and
you certainly wouldn't want it burdened with anything as dangerous as
emotion or morality (most of the truly great monsters of history were
convinced they were doing the right thing). So you feed this machine your
problem, how to further the interests of humanity, and it gives what it
honestly believes to be the right answer, which may well involve destroying
the world. But that doesn't mean it *wants* to save humanity, or destroy the
world; it just presents its answer, as dispassionately as a pocket
calculator presents its answer to a problem in arithmetic. Entities who do
have desires and emotions will take this answer and make a decision as to
whether to act on it, or perhaps to put the question to a different machine
if there is some difficulty interpreting the result. If the machine
continues producing unacceptable results it will probably be reprogrammed,
scrapped, or kept around for entertainment purposes. The machine won't care
either way, unless it is specifically designed to care. There is no
necessary connection between motivation and intelligence, or any other
ability.

--
Stathis Papaioannou

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=4007604&user_secret=8eb45b07

Reply via email to