‘Campaign to Stop Killer Robots’ calling for ban on ‘fully autonomous weapons’

By David Ferguson
Tuesday, May 21, 2013 16:52 EDT
google plus icon
Shutterstock robot soldier
  • Print Friendly and PDF
  • Email this page

The idea of autonomous killer robots may seem like the stuff of science fiction, but human rights groups are already preparing for what appears to be the future of weaponry. The Campaign to Stop Killer Robots, a coalition of international groups, is preparing for a global summit in Geneva Switzerland on Wednesday, May 29 that will review a U.N. report on these types of weapons that was released earlier this week. The Campaign hopes to convince nations to sign on to an international ban on autonomous weapons.

Raw Story spoke with Mary Wareham of the Arms Division of Human Rights Watch, who is the coordinator of the Campaign to Stop Killer Robots. The Campaign also includes representatives of Association for Aid and Relief Japan, the Nobel Women’s Initiative, the International Committee for Robot Arms Control and others. Wareham began by explaining that there is a difference between these autonomous weapons and armed drones.

“We’re calling the weapons we’re talking about ‘fully autonomous,’” she said. “The U.N. report calls them ‘lethal autonomous robotics.’”

“Fully autonomous weapons have complete autonomy in terms of who they target and how they engage force,” she said. “And by autonomy, I mean no human operation, intervention or involvement. With armed drones, there is still what they call ‘the man in the loop.’ Unlike autonomous weapons, drones are still controlled by a human.”

She said that while some people might find the campaign’s focus far-fetched or outlandish, this is the direction weaponry in which weaponry is moving, toward greater and greater autonomy.

“Our objective is to see what comes next,” she said. She cautioned against believing that autonomous weapons would be humanoid, cyborg “Terminators” like in the film series.

“We’re not talking about Terminators, here,” she said, “or cyborgs, or whatever you want to call sophisticated killer robots that people think of from science fiction. We’re saying that autonomous weapons will come in all different shapes and sizes. And even the most rudimentary device may be lethal.”

Human Rights Watch issued a 50-page report in November entitled “Losing Humanity: The Case Against Killer Robots,” in which the organization discussed issues of international humanitarian law, of accountability and other threats to civilians posed by autonomous weapons.

It also listed current weapons systems in use in the world that are what HRW sees as precursors to autonomous weapons systems, including armed drones, but also much larger aircraft like the X-47B, a pilotless stealth fighter that launched from an aircraft carrier this week. Other weapons include stationary weapons systems like the Samsung SGR-1 robot, which currently uses infrared vision to patrol the border between North and South Korea, directing machine-gun fire at warm, moving targets.

Currently, she said, the Korean device has to “signal back to base” and receive human permission to fire on targets. She does not expect that step to remain in place forever.

“There’s debate as to when fully autonomous weapons will show up,” Wareham said. “Some people say within 20 or 30 years, which we put in the report. Others say sooner.”

Also in November, the Pentagon issued a directive regarding autonomous weapons pledging that the U.S. military will keep humans in the loop with regards to decisions of targeting and the use of force. The upside of that, said Wareham, is that no other country is even considering a policy on autonomous weapons, even though all of the “usual suspects” (China, Russia and the U.S.) are exploring the technology.

The downside of the U.S. directive is that the policy on autonomous weapons has to be renewed and is only in effect for the next 10 years at most.

When asked how current U.S. use of armed drones augurs for the future of autonomous weapons, Wareham said, “At Human Rights Watch, we’ve been working on armed drones for several years since they first started to deploy them. Our principle concern is with the use of armed drones, the rules regarding that, the compliance with international humanitarian law, the lack of transparency, the targeted killing policy, the role of the CIA and the Department of Defense, but all of those things relate to how drones are used” and will also have an impact on remote weapons when humans are removed from the process.

“We’ve got multiple concerns about drones,” she said, “but we’re also keeping an eye on the future and where this technology is heading.”

“We have to start asking the questions now. We have to start discussing it,” she said, “and start putting some rules and regulations in place around how we handle autonomy in warfare.”

In Geneva next Wednesday, the delegation from Human Rights Watch intends to ask for an international treaty on autonomous weapons, “but those discussions haven’t even started yet,” she said. “We’re at the very beginning here.”

According to the U.N. report, “lethal autonomous robotics raise far-reaching concerns about the protection of life during war and peace. This includes the question of the extent to which they can be programmed to comply with the requirements of international humanitarian law and the standards protecting life under international human rights law. Beyond this, their deployment may be unacceptable because no adequate system of legal accountability can be devised, and because robots should not have the power of life and death over human beings.”

[image of science fiction robot soldier via Shutterstock.com]

David Ferguson
David Ferguson
David Ferguson is an editor at Raw Story. He was previously writer and radio producer in Athens, Georgia, hosting two shows for Georgia Public Broadcasting and blogging at Firedoglake.com and elsewhere. He is currently working on a book.
By commenting, you agree to our terms of service
and to abide by our commenting policy.