I just had a notion. The proper sensory input and motor output for an AI is the computer screen (and sound input and regular keyboard and mouse input). One thing that needs to exist is a freely available standard API for these things, so people can work on them, plus implementations for the platforms that people use. My hope is that it would give different researchers, especially all those lone wolves out there, something intercompatible to work with. It also seems possible that this could be a common mechanism for the different systems to work together, in a sort of extension of the Blackboard model. And, as a lighter element of it, I'd really like it if these projects could use video games, because they more and more have become very sophisticated real-world modelling tools. andi
----- This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/[EMAIL PROTECTED]