Humans give subtler rewards to each other (not just one-dimensional
rewards) because we share a complex emotional/social system.
Potentially, AGIs could learn to accept complex, nuanced rewards from
humans via interacting with them in a sim world for a while, in a
variety of situations...
This i
>Right now, the only representationally explicit goal is "please the
>teacher." Learning/creating information is as of now left as an
>implicit goal. But once the system has reached Piaget's formal stage,
>it will be useful to make learning/creating information a reflectively
>(and possibly repre
I intend to start at a bit higher age level of teen / reduced knowledge
adult,
That is not possible in an approach that, like Novamente, is primarily
experiential-learning-based...
-- Ben
-
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, p
On 12/8/06, James Ratcliff <[EMAIL PROTECTED]> wrote:
What are the "meta-goal" properties there defined?
For example:
-- have as few distinct supergoals as possible
-- keep the supergoals as simple as possible
-- avoid logical contradiction between supergoals
-- minimize pragmatic, probabilisti