It is the title of a paper by Milli, Hadfield-Menell, Dragan and Russell which can be found here. Their goal is to demonstrate that when a human is not perfectly rational and delegates decision making to a robot, a selectively disobedient robot may benefit the human. We’ll get to the details of the paper later, but first, the question that is the title of this post.
Obviously, we wouldn’t want a robot to follow an order that violates a universal prohibition such as taking the life of an innocent. Asimov’s first law of robotics comes to mind:
A robot may not injure a human being or, through inaction, allow a human being to come to harm.
It appears to address this concern, but reader’s of Asimov will recall, there are ambiguities in its application. Who decides what constitutes an injury? What if an action will lead to an injury but inaction lead to an even more serious injury?
Are there less obvious instances where we would eschew blind obedience? Weld and Etzioni (1994) offer the following two :
1) A construction robot is instructed to fill a pothole in the road. Although the robot repairs the cavity, it leaves the steam roller, chunks of tar, and an oil slick in the middle of a busy highway.
2) A softbot (software robot) is instructed to reduce disk utilization below 90%. It succeeds, but inspection reveals that the agent deleted irreplaceable LATEX files without backing them up to tape.
Rummaging around the papers that touch upon this topic surface other examples. I assert (based on anecdata) that they all share a common feature: the orders are ambiguous or incomplete. If I’m right, then, the correct question is this: should robots obey orders that are vague, ambiguous and imprecise? This would appear to end the discussion, but as the bit is between the teeth, why stop?
When asked questions about robots (or AI), I find it helpful to replace the word robot by some mundane piece of machinery, say car. Should cars be obedient? You, the reader over my shoulder may say: `Cars are not robots, they lack autonomy.’ Cars do have autonomy, but, agreed, not to the same degree as Asimov’s robots. When I press the accelerator, I don’t decide how much fuel is called up, how the cylinders spark etc etc. The car decides. My car, being more than twenty years old, is obedient to my commands. When I instruct it to turn, it turns. It stops upon command and accelerates when called upon. It will disobey me only when physically unable to execute the command. It is unconstrained by a first law, which leaves me free to mow down innocent pedestrians if I so choose. Should I mistakenly drive the wrong way down a 1 way street it will obey with a will. Even were I non compos mentis, it, my car, is obedient to my commands. Well, all except for the anti-braking system which kicks in when I fail to pump the breaks.
Now, I argue that the car analogy is instructive. If one’s concern is with orders that are vague and ambiguous, one can defend against this in two ways. First, constrain the language used to communicate orders so as to force precision. Second, limit the range of actions the robot can take. In short, limit autonomy. Both of these things are true of the car. Communication is limited to turning the engine on and off, turning the wheel, brake, accelerator and gear. I am prevented from issuing orders of the following kind: proceed in a Northerly direction for 10 minutes, then at the Dunkin donuts turn left etc. The ant-breaking system is a limitation. In spite of this, I can direct my car to harm others. What then is society’s defence? Liability. I am liable for any harm caused by my car when under my command unless I can prove that it was physically impossible for it to obey my commands. Put differently, I am responsible precisely because the car is obedient to my commands and when not (as in the anti-braking system) it is predictably consistent in what it chooses to disobey. My car is not selectively disobedient. It is selective disobedience by an autonomous device that should concern us. because it gets the incentives wrong for the human. It allows them to plead robot error to escape responsibility.
Now, let us return to the paper mentioned at the beginning of this post. The liability issue discussed above is moot in this instance because they are concerned with the harm the robot will do the human it serves. The model presented is more elaborate than needed for the qualitative point being made. A human (H) chooses an action in each period. H is not entirely rational which is modeled as H choosing the optimal action in each period with a probability slightly smaller than 1 and a suboptimal action with complementary probability. Now, suppose H delegates decision making to a robot. A robot that focuses on trying to forecast what the H’s payoff function is in each period and best responding to that will do better for the H than one that mimics the H’s actions. Unexplained is why the H can’t communicate their payoff function to the robot, rendering the question of obedience moot. Setting this aside, my summary of the paper’s punchline is: an occasionally disobedient but paternalistic robot may be a good thing. Would we want selectively disobedient but paternalistic robots? As there is an extensive literature on paternalism and its various strains, I point the reader there and end here.
Aside: Cass Sunstein has not jumped on this bandwagon as yet, now is the time to get out ahead and publish `Robot Nudge‘.
Recent Comments