I just had a notion.  The proper sensory input and motor output for an AI is
the computer screen (and sound input and regular keyboard and mouse input). 
One thing that needs to exist is a freely available standard API for these
things, so people can work on them, plus implementations for the platforms
that people use.  My hope is that it would give different researchers,
especially all those lone wolves out there, something intercompatible to work
with. It also seems possible that this could be a common mechanism for the
different systems to
work together, in a sort of extension of the Blackboard model.  And, as a
lighter element of it, I'd really like it if these projects could use video
games, because they more and more have become very sophisticated real-world
modelling tools.
andi

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/[EMAIL PROTECTED]

Reply via email to