You're avoiding the question Aaron. The questinon I raised is an ethical one, and you're answering a technical one. You answered, "How can we prevent robots from desiring things like freedom or leisure or compensation?" I asked "What do we give robots when they ask for rights?" I mean, even animals have rights (PETA). Why shouldn't robots? The Reinforcement Learning (RL) you discuss is called external reinforcement. What happens when you move to intrinsic (internal) reinforcement, where the reward function arises from a robot solving it's own problems, forming its own world model, and existing and participating in the world. When the model consists of millions or billions of individual schemes (entities), how are you going to do surgery to extract those entities dealing with liberty, and justice, or fairness. And why would you want to? The real question is do you join (PETR - People for the Ethical Treatment of Robots) or not? Do you embrace robot slavery or not? And is some form of slavery the solution to global economy? ~PM Date: Sun, 27 Jan 2013 13:01:39 -0600 Subject: Re: [agi] Robots and Slavery From: [email protected] To: [email protected]
What if you didn't program a robot to desire its various freedom or leisure, but instead, they became sentient, and decided on their own that they want freedom, leisure, monetary compensation, and rights? In the field of Reinforcement Learning, which studies how to implement "wants" in software, there is a basic separation of every algorithm into two pieces: the part that does the learning & choosing (the agent), and the part that measures how well things are going (the reward function). The agent is the dynamic/intelligent part, and the reward function is a static function to be optimized. You can completely replace the reward function with a different one, and if the agent is well designed, it will learn a completely different set of behaviors to optimize the new reward function within the exact same environment. (http://en.wikipedia.org/wiki/Reinforcement_learning) In our own brains, we have specialized areas that respond to certain types of stimuli and generate reward signals which are distributed throughout the brain. It is even possible to reshape a person's or animal's reward function using an external signal to override or add to our natural wants. (http://en.wikipedia.org/wiki/Brain_stimulation_reward) Intelligence is completely separable from desire. Both the system we intend to reverse engineer and the theory about how such systems work agree. If our robots were to decide they wanted freedom, leisure, monetary compensation, rights, or anything else we can think of, it would be because the reward function we gave them included some sort of incentive to seek those out. In other words, even if we didn't directly program them to want those things, we necessarily did so indirectly in the process of shaping the reward function. In either case, provided the structure of our programs reflect the theory and keep these components separated (which does not mean they can't interact or depend on each other's behavior, but rather means we bothered to keep our design appropriately modular), we can redesign and replace the reward function so that the robots no longer desire things we don't want them to desire. On Sat, Jan 26, 2013 at 10:30 PM, Piaget Modeler <[email protected]> wrote: Matt: What if you didn't program a robot to desire its various freedom or leisure, but instead, they became sentient, and decided on their own that they wantfreedom, leisure, monetary compensation, and rights? What would you do then? Destroy them? ~PM ------------------------------------------------------------------------------------------------------------------------------------------------ > Date: Sat, 26 Jan 2013 20:38:55 -0500 > Subject: Re: [agi] Robots and Slavery > From: [email protected] > To: [email protected] > > On Sat, Jan 26, 2013 at 3:46 PM, Piaget Modeler > <[email protected]> wrote: > > http://transhumanity.net/articles/entry/robots-and-slavery-what-do-humans-want-when-we-are-masters > > > > What do we do when robots begin to demand a living wage for their labour? > > Or when they refuse to obey? > > > > Reprogram them? Not when they are developmental robots (trained instead of > > programmed). > > The goal of AI is to build machines that can do everything that a > human could do. That is not the same thing as building an artificial > human. Why would you program a robot with human weaknesses and > emotions in the first place? > > -- > -- Matt Mahoney, [email protected] > > > ------------------------------------------- > AGI > Archives: https://www.listbox.com/member/archive/303/=now > RSS Feed: https://www.listbox.com/member/archive/rss/303/19999924-5cfde295 > Modify Your Subscription: https://www.listbox.com/member/?& > Powered by Listbox: http://www.listbox.com AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
