Benjamin Goertzel wrote:
So a VPOP is defined to be a safe AGI. And its purpose is to solve the
problem of building the first safe AGI...
No, the VPOP is supposed to be, in a way, a safe **narrow AI** with a goal
of carrying out a certain kind of extrapolation
What you are doubting, perhaps, is that it is possible to create a suitably
powerful optimization process using a narrow-AI methodology, without
giving this optimization process a flexible, AGI-style motivational
system...
You may be right, I'm really not sure...
As expressed many times before, I consider CEV a fascinating thought-
experiment, but I'm not even sure it's a well-founded concept (would the
sequence be convergent? how different would the results be for different
people?), nor that it will ever be computationally feasible ... and I
really
doubt it's the way the Singularity's gonna happen!!
On all these last points, we agree.
Richard Loosemore
-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=4007604&id_secret=57875471-fd31f5