Category Archives: Ethics

A Few Thoughts on “Killer Robots”

I’ve been following the discussion over on IEEE Automaton about the letter from the AI conference in Brazil, (for, against, rebuttal, etc.) , and since this is the kind of thing that normally gets me very engaged (or enraged), I was surprised to find myself ambivalent on the subject.

The original letter ends, “Starting a military AI arms race is a bad idea, and should be prevented by a ban on offensive autonomous weapons beyond meaningful human control.”  but this gives us several propositions to consider:

  • “Starting a military AI arms race is a bad idea”

I think we can agree that this is true.

  • “… a military AI arms race … should be prevented”

This is also something we can agree on.  But if we look at the underlying assumptions behind the complete statement, we get a few more propositions that are less easy to respond to.

  • A military AI arms race can be prevented by a ban on the development of specific types of weapons

I disagree with this.  It seems to be saying that the only research that would support a military AI arms race is research into specific types of weapons, specifically weaponized autonomous robots.  It seems eminently plausible to me that research into many technologies other than offensive autonomous weapons would contribute to a military AI arms race:  research into improved mind-machine interfaces for non-autonomous robots and other systems; research into more intelligent surveillance systems.  A military AI arms race would result in many terrifying things other than autonomous weapons.  (For example, anything involving public utilities or transportation networks.)  The underlying technologies that would be developed in this arms race are still likely to be developed for other applications.  It is entirely plausible that you could conduct a military AI arms race without ever reaching for a weapon.

The only aspect of the problem that requires a weapon is the part where the robot learns how to control the aiming and initiation of that weapon, and that isn’t a problem for artificial intelligence research, that’s a low level controller problem that’s already being solved for many types of systems.  The artificial intelligence question that contributes to a military AI arms race is “when should I fire”, and you don’t need a gun to study that one.  You don’t even need a problem space that includes guns.  You just need an analogous non-military problem and you can work on that technology without ever violating a ban on autonomous offensive weapons.

  • “offensive autonomous weapons” are at a sufficiently early stage of their development that banning them is practical

I’m not at all sure this is true, since there are numerous examples of armed military robots, and even individual non-experts that have already assembled terrifying combinations of robots and weaponry.  The step from this RC quadcopter to an autonomous Lego Mindstorms weapon requires skills and technology available to high school students.

  • “offensive autonomous weapons” are a sufficiently concrete subject that they can meaningfully be banned

Bear in mind that this letter is coming from the AI community, not the robotics community.   They are not suggesting a ban on the technology they are most concerned with (the decision-making and possibly the perceptual elements of the software), but are instead suggesting a ban on other peoples’ work (specific combinations of software and hardware).

  • “beyond meaningful human control” is a reasonable way of specifying the particularly troublesome weapons that we ought to ban

We don’t even have measures for how much human control is associated with “autonomous”.  How on Earth would we ever figure out whether a given autonomous robot is actually “beyond” “meaningful” human control?

The discussions so far seem to have missed a key distinction that we need to make in order to talk sensibly about this topic.  We are discussing at least three separate questions.

  1. Should the international community be doing anything to prevent their constituent militaries from exploring this as an option for future combat between countries?
  2. Should there be restrictions or laws in place in individual countries governing what individuals choose to do with equipment that they legally purchase or create?
  3. Are we discussing the development of specific technology, or are we discussing applications of technology that has been developed for other purposes?

The original letter is explicitly aimed at the first question – they say that this application of technology is both sufficiently dangerous and sufficiently unpredictable that the international community should be treating it like certain types of mines and banning its development entirely.  But it asks us to ban technology development, not specific applications of technology, and that is a core problem.

It is appropriate to discuss a ban on specific applications of autonomous robots, in the same way we have bans on specific applications of other technologies, both at the national and international level.  But we should not conflate that with banning development of any technology that might eventually be able to be used for those applications, and we should not assume that banning those applications will have any more effect than the “severe stigmatization” mentioned in the for article.  The whole purpose of robotics as a technology is its flexibility and adaptability and the ease with which we can alter its uses and goals.

Not putting a ban in place until we understand the potential benefits we are giving up, as suggested in Dr. Arkin’s article, is an eminently reasonable proposition, but his response ignores a core aspect of question 1.  It addresses only cases where the situations involving the autonomous robots are in some way comparable to the situations involving the use of soldiers.  This is unlikely, as two of the main reasons autonomy is used instead of automation or teleoperation are (1) to increase the safety of the system as it performs the task in an unpredictable environment (automation is ineffective when the obstacle positions are unknown; teleoperation is ineffective when communications latency is high or bandwidth is low and the task is urgent) and (2) to allow systems to perform tasks that were not previously possible.

We cannot assume that the robots will be performing the same kinds of missions as human soldiers, and we cannot assume that the kinds of missions they perform will reduce casualties compared to humans if those missions weren’t possible before.  Instead, it is just as likely that there will be additional casualties as the robots fail to figure out who the target is, and as the target humans become adept at disguising themselves as non-combatants.  I agree that we do not yet know what those trade-offs are, and we will make much better decisions about what needs to be banned once we know more, but I disagree that the only way to discover those trade-offs is by continuing research into explicitly dangerous robots.

We can explore perception and ethical robot operation without bringing in battlefield robots – there are plenty of other areas in which these would be beneficial capabilities (child care, search and rescue of stranded hikers, elder care, crowd control, even disrupting a bar fight).  Banning the application doesn’t prevent us from estimating what the benefits of the banned approach might be.

But even if estimating the benefits were only possible with continued development, preventing a ban on a technology because of the affect on our risk assessment of that technology seems like a bad idea.  Banning a technology because it might eventually be used to develop something bad is a worse one.

Which takes us to question 2.  Robotics isn’t like chemistry, where bans on specific things have created sanitized chemistry sets until they have very little of interest left in them and have criminalized human curiosity (and I worry that this is happening in robotics), or like nuclear weapons, where one specific element of the weapon (in this case, radioactive elements) is tightly regulated.

Evan Ackerman makes an excellent point early in his article against banning killer robots.  He points out that the barrier to entry for this particular technology is so low that international bans cannot be a valid answer, any more than international bans on land mines prevent individuals from creating improvised explosive devices and putting them on roads to blow up their enemies’ trucks.

But I would take that argument a step further – it’s not just that the barrier to entry is unreasonably low.  It’s that the only element we could ban that would be associated uniquely with autonomous killer robots is the weapons, and that banning the weapons is insufficient to prevent the development of autonomous killer robots (note: edited on 2/5/16 to clarify argument).  I’m not arguing that an autonomous robot with a gun isn’t a fundamentally scarier and potentially a more dangerous entity than a gun by itself (or even a gun in the hands of a human).  I’m saying that the thing that makes the robot scary is the gun, but that banning the gun won’t prevent scary robots from existing.  Robots with knives could be scary, but right now they’re chopping vegetables.  Robots with legs, even when they stop falling over, even if they can run, are still at the mercy of the software that defines their goals.

Autonomous robots don’t have a physical property that is the difference between dangerous and safe. The only concrete difference between an “autonomous killer robot” and a perfectly normal, safe-for-everyday-purposes autonomous robot is a sign error in the goal definition.  A perfectly safe consumer good (an autonomous car) becomes an autonomous killer robot when the acceleration associated with the “don’t hit pedestrians” goal goes from negative to positive.

To address the third question:  there are only two things we can regulate on the consumer end:  the weapons themselves (which is not going to prevent people from making killer robots) and the software they run (even excluding learning and adaptation, we do not have tools that would enable us to determine definitively whether a given robot was capable of harming a human intentionally – in almost all cases, the robot wouldn’t know either).

Even assuming that we could overcome the deep attachment that the public and industry has for the many peaceable uses of robots, there is no way to ban an autonomous robot that can kill without also banning robots.  An autonomous car running you over is just as fatal as a robot soldier shooting you with a gun.

The difference between an “offensive autonomous weapon”  and an autonomous non-killer robot is that the non-killer robot doesn’t have a goal that says “kill”.

Are “Animal Robots” Robots At All?

IEEE Spectrum had an article from a Chinese lab on “animal robots”. When I was back in grad school, we worked on a proposal for a similar project. We were going to connect moth antennae to a robot and use them as sensors to drive the robot around. Various people have worked on similar ideas over the years, and they’re a little disturbing but generally very informative.

This paper had a diametrically opposed approach. Instead of replacing the animal’s actuators with a robotic interface, they mounted a controller on the back of a rat and connected it to the rat’s motor neurons. Instead of creating a robot with biological sensors, they created a remote control animal.

I am profoundly uncomfortable with this, especially as the research community moves away from insects and towards mammals. While I have no problem with doing horrible things to mosquitoes (largely in retaliation for the severe discomfort they have caused me over the years), rats are intelligent and interesting creatures. The thought of an intelligent animal being forced into actions via the equivalent of a muscular tic is repugnant.

However, that is not the larger question we need to address. This is:

Is a biological animal with a controller attached to its motor neurons a robot?

Setting aside my discomfort (to the best of my ability), I think that these should not count as robots. In essence, a robot is a machine that does a task for a human. We do not consider police horses or working dogs to be robots, because they are not machines. Just because we have traded reins for a computer and a bridle for wires into the brain and as a result have reduced their ability to act independently does not mean that the animal has suddenly become a robot.

We have a word for organisms that have been merged with mechanisms until it is difficult to tell where the animal ends and the machines begin: we call them “cyborgs”.

Cyborg research is most closely related to robotics, so I expect that researchers developing cyborgs are going to be part of the robotics community for some time to come. But we shouldn’t expect them to stay part of our community forever, and we certainly shouldn’t accept this hijacking of our terminology. We have just managed to pry the word “robot” out of the hands of the computer science community (yes, robots have to have bodies); I’d hate to have to have that argument all over again…

Enough Inappropriate Ethical Analysis Already!

I’ve reached my threshold for uninformed articles about the supposed ethical conundrums that roboticists should be taking into account when they design autonomous cars.  At the end, I’ll be adding more questions to the list of potential research topics, but first, I want to address why these articles are so infuriating.

Grrrr.  Begin rant.

The articles don’t seem to have been vetted at all.  Exhortations to follow Asmiov’s three laws, or (worse yet) implement additional laws and address philosophical dilemmas about which group of people to kill in some avoidable concatenation of tragic events – they all incorporate assumptions about how robots are built and designed that are (at least currently) untrue.

These situations all involve an unrealistic view of a robot’s ability to perceive its environment.  Anyone who designs the software that robots use to perform tasks knows what a big problem perception is.  Sometimes advances in sensing make a big difference, but often improved sensors simply result in more data available to the robot, not in an improved understanding of its surroundings.  And all of these ethical and philosophical dilemmas fundamentally rest on perception.  The decision about which group of people to kill is only an ethical dilemma if, first, it is possible for the car to find itself in that situation, and if, second, it is actually possible for the car to differentiate between the two groups in some way.

Asimov’s Laws (often taken as an example of where to start) rest on the assumption that the robot can project the ramifications of its actions and determine whether an action it takes is likely to result in harm to a human or to itself.  The whole reason we have autonomous systems instead of automated systems is because the real world is chaotic and unpredictable.  We are just barely able to identify humans in well-defined environments with adequate lighting; predicting that a human will fall over if pushed is at least within the realm of possibility (assuming we ignore the complexities of the surface they’re standing on, the composition of their shoes, an at best inaccurate estimate of their weight, their physical strength, and any other objects that may be in the vicinity).  Determining whether a human is far enough away from an explosion to bound the probability that they will be injured would require a good understanding of the exact placement of the bomb, the masses, materials, and robustness to stress of every object in the environment, the properties of the human, including its clothing, and the precise composition of the bomb.  On top of this, the robot needs to compute the probable impact of any actions it takes, which implies awareness of its own composition and any time constraints.  Humans, with all our millenia of successful life or death decisions and extremely effective sensory post-processing, are incapable of deciding what a “correct” set of rules would look like – if we knew what the “correct” set was and could apply it consistently, we wouldn’t need judges and juries.  Even if we could implement them, how would we know what to implement?

We can’t even build a robot that would be able to perceive the world sufficiently to allow us to develop a robot that complied with Asimov’s Laws.  We might, theoretically, eventually be able to build a robot that could face the trolley problem (see the background reading below) in a meaningful way, if both the sensors and the perceptual software were mature enough, but why would we?

Fundamentally, the job of a car is to get its occupants safely from one place to another while leaving its environment in substantially the same condition it was when it started.  Sensors for these systems need to operate very, very quickly, which means efficient sensing and minimal processing.

In order to make sensing efficient, as a designer, you derive the smallest set of things you need to identify as important elements of the environment.  For a car, this means you pay attention to objects that are in the road and objects whose movements indicate that they are likely to end up in the road before you’re past them.  You also pay attention to signs and features that are there to help you decide what to do:  stop signs, yield signs, traffic lights, turn signals, brake lights and reverse lights, and so forth.

But the net result of this is that telling the difference between a human, an old human, a child, a dog, or a cow becomes simply a matter of answering “how fast is it moving,” “is it going to be in the road,” and “am I moving slowly enough to avoid it”.  There is no reason to give a car the ability to tell the difference between one object and another, except to determine whether the road is permanently blocked or whether the object is likely to move out of the way.  In no case should the car be expected to hit anything.

A human on a bicycle is faster than a toddler, but a toddler is more maneuverable and less predictable.  The simplest solution is simply to assume that all moving objects have the potential to unpredictably swerve around like a toddler as fast as someone on a bicycle and behave accordingly.  There is no incentive driving a car manufacturer to build in the kind of awareness of the environment that would enable a car to distinguish between an old man and a young boy beyond recognizing that they are both capable of movement.  Just as human drivers are sublimely unaware of whether the people they drive past are saints or murderers, an autonomous car would be unlikely to know or care about the age of the humans in its vicinity.

These ethical dilemmas aren’t actual dilemmas the robots are likely to face.  Instead of providing commentary that will help policy-makers understand the actual problems associated with robots, the writers use arguments based on incorrect assumptions to conclude that roboticists should be implementing ethical constraints that are fundamentally impossible to enunciate in a form that the robots will be able to understand.

And in the meantime, roboticists have actual ethical dilemmas that need to be faced.

What the media can do about it:

Before you run a story about scary things a robot might do and how people who design robots should really be thinking about this stuff, just call a robotics professor at your local university.  There will be at least one.  Just go to the main page, search for “robot”, and look for the professor associated with the articles.  Call them up and ask if they can look it over and tell you whether the article makes sense.  They’re probably busy, but if you ask and they don’t have time, they may have a grad student that can help you.  You might even consider offering to put the name of whoever helps in the paper for helping you vet the article (e.g. “reviewed by X at University Y” in a footnote).

What we should be worrying about:

We have bigger ethical problems than made-up stories about hypotheticals that, in most cases, aren’t even likely to be possible.

For instance, we know just how difficult it is to figure out whether a given robot is going to behave the way we want.  Most of the time, we haven’t even figured out what we want (see the “correct” rules problem, above)  We build it to do something that seems obvious (“don’t hit things”), and then we run the robot somewhere new and discover that we didn’t just want it to not hit things, we wanted it to also avoid things it could fall off of (“avoid stairs”) and to avoid the fringe on a rug because it gets caught in the wheels (“avoid fringe”) and to somehow get back to its charger before it runs out of power so we don’t have to rescue it from the middle of the floor.  Even when we think we have finally done enough experiments and figured out what we really want it to do, how sure are we that it is ready to operate in someone else’s house?  If we have built a robot to help take care of someone elderly who has difficulty opening pill bottles and keeping track of their medication, how certain do we have to be that we have covered the typical or even rare use cases?

Where is the ethical boundary around providing the elderly with telepresence robots when they become too infirm to participate in their childrens’ and grandchildrens’ lives?  Are we encouraging families to sacrifice physical contact for virtual participation?  Will they be more connected to their lives, or less?  (I’m guessing it will depend largely on the person.)

What are the ethics of robot-assisted childcare?  If we are talking about robots that allow institutionalized children with communicative and movement disorders to finally interact with their peers through robot avatars, surely that is a good use of robots.  Allowing sick children to attend school through a robot avatar, probably a good thing (although an argument could be made for allowing sick children to rest and recuperate instead, catching up once they’re better, or for customizing each child’s school experience so they learn at a pace that fits their circumstances, whatever those may be).

But what about perfectly healthy children whose parents create a worldview where the safest thing for their child is to stay home and only ever attend school virtually?  Already children rarely play in their front yards, or wander through their own neighborhoods unsupervised – both things that were common when I was a child.  When I was 7, I was expected to walk 2 blocks to school along my residential street, whether my neighbor (also 7) was there to keep me company or not.  By the time I was 10, I was expected to walk or bike to and from school on my own and encouraged to go to the park and play or take my bike to the nearest playground.   By the time I was 14, I was allowed to bike anywhere I could get to. And this was in an era where being able to call home meant getting to a pay phone and using quarters.  I never actually needed those quarters in my pocket, and mostly I didn’t even bother carrying them, since if I couldn’t walk, I wouldn’t be able to use them, and if I could walk, I could always get where I needed to go eventually.

When my son was in elementary school, almost none of the kids walked to and from school.  I walked an 11-year-old home because her parent hadn’t come to pick her up and she was too uneasy to walk home alone because of the potential for predatory strangers.  It was 6 blocks, through a leafy, reasonably prosperous, well kept up residential neighborhood with almost no traffic and no history of predatory strangers.

If we build a robot that a sick or disruptive child can use to attend school without endangering his or her classmates, there is nothing stopping a parent from using that robot to prevent the classmates from endangering their child, and it seems to be becoming more and more likely that some parents will choose to do so.

What kinds of constraints is it ethical to implement in the robot?  What are the ethics of socializing children through actual physical interaction versus socializing them entirely through virtual means, especially when (if the robot is designed for use by disruptive children) the robot disallows certain kinds of behavior?  And potentially even more worrying, how are we warping these childrens’ ability to explore new ways of thinking by constraining them in this way?  Are children who interact with others through an artificially constrained platform less likely to innovate as adults, or are they likely to have inbuilt blind spots where certain kinds of actions simply don’t register as options?

What about the ethics of medical robots?  A few years ago, a company put up a billboard advertisement asking “Would you swallow a robot?” whose subhead said that they didn’t make the hardware, they made the software that made it possible.  At a conference workshop, about half those present (primarily robotics researchers) agreed that they would swallow a robot given appropriate circumstances, such as a doctor’s recommendation, and proper certification procedures.  When it was specified that the robot included on-board software, that number dropped to about ten percent, even with certification authority and medial authority safeguards.

And this doesn’t even get into the ethics of military robots.

Some (many?) of these questions are questions for ethicists and policy-makers, but as roboticists we could at least be trying to address the elements that relate to the design of our systems.

Should we be building in hooks to support specific types of safety caging on our systems, so that when the policy-makers decide the ethical boundaries of acceptable action, our tools are ready to accept them?  Shoud roboticists be researching the ethical options that are even available to robots?  Should we be contemplating what additional sensors our systems should have to enable them to be sufficiently aware of their context to act ethically?

And, bringing it back to the hypothetical ethical decisions that cars are unlikely to ever have to make, should we be developing a language that enables us to describe the areas in which it is possible for a given system to operate ethically, and the areas in which it is, fundamentally, just a machine, doing exactly what we told it to do?

 

Background reading:

The Trolley Problem:

http://www.theatlantic.com/technology/archive/2013/10/the-ethics-of-autonomous-cars/280360/ –   “Programmers still will need to instruct an automated car on how to act for the entire range of foreseeable scenarios, as well as lay down guiding principles for unforeseen scenarios. So programmers will need to confront this decision, even if we human drivers never have to in the real world.”

http://www.wired.com/2014/08/heres-a-terrible-idea-robot-cars-with-adjustable-ethics-settings/ – “one customer may set the car (which he paid for) to jealously value his life over all others; another user may prefer that the car values all lives the same and minimizes harm overall; yet another may want to minimize legal liability and costs for herself; and other settings are possible”

http://www.popsci.com/blog-network/zero-moment/mathematics-murder-should-robot-sacrifice-your-life-save-two

http://www.forbes.com/sites/timworstall/2014/06/18/when-should-your-driverless-car-from-google-be-allowed-to-kill-you/ – “It’s seems clear and obvious that the technology is going to get sorted pretty soon….The basic Trolly Problem is easy enough, kill fewer people by preference.”

http://www.nature.com/news/intelligent-robots-must-uphold-human-rights-1.17167 – “We should extend Asimov’s Three Laws of Robotics to support work on AIonAI interaction. I suggest a fourth law: all robots endowed with comparable human reason and conscience should act towards one another in a spirit of brotherhood and sisterhood.”

And a couple that actually seem to be taking a more sensible approach to the problem:

http://www.wired.com/2014/05/the-robot-car-of-tomorrow-might-just-be-programmed-to-hit-you/

http://robohub.org/an-ethical-dilemma-when-robot-cars-must-kill-who-should-pick-the-victim/, and the slightly more disturbing followup:  http://phys.org/news/2014-08-ethics-driverless-cars.html

Situations in real life where children actually get hit:

Bystanders apparently get run over by potentially negligent parents (link), and occasionally a child gets hit by a police car responding to an emergency (link) or run over while chasing an ice cream truck on a bicycle (link), and sometimes children get run over because they’re close to a car when it starts moving and the driver doesn’t know they’re there (link, link, link), especially in a parking lot (link).  Crosswalks are apparently dangerous (link), as are apartments (link).

This is close to the runaway toddler scenario (link), in a parking lot with a distracted driver and an environment with poor visibility, but an autonomous vehicle wouldn’t be distracted and would be going slowly because of the uncertainty in its surroundings.  Instances of toddlers running abruptly into traffic on roads, rather than in parking lots, seem to be vanishingly infrequent.

Furthermore, parents today who simply let their children walk home from the park alone can end up on probation with child protective services (link).  Culturally, we’ve been shifting away from responsibility and towards safety for a long time.