This week, the International Committee for Robot Arms Control (ICRAC) is holding a conference in Berlin exploring options for the regulation of “killer” robots.
Andrew Gibson is a freelance journalist interested in military robotics, arms control (particularly nuclear), civil wars and politics
This week, the International Committee for Robot Arms Control (ICRAC) is holding a conference in Berlin to explore options for the regulation of unmanned and fully autonomous combat vehicles. ICRAC, a coalition of roboticists, physicists and philosophers, have invited lawyers, members of the military and arms control wonks to offer perspectives on the feasibility of an international regulatory regime.
ICRAC have stated their goals as follows:
• The prohibition of the development, deployment and use of armed autonomous unmanned systems. Machines should not be allowed to make the decision to kill people;
• Limitations on the range and weapons carried by “man in the loop” unmanned systems and on their deployment in postures threatening to other states;
• A ban on arming unmanned systems with nuclear weapons;
• The prohibition of the development, deployment and use of robot space weapons.
Professor Noel Sharkey, a roboticist from Sheffield University and founding member of ICRAC, told Left Foot Forward:
“A number of us were dissatisfied at the lack of international discussion about the role of robots in war. I’m not an expert in ethics and law but I bring technological input.
“My colleague Jürgen Altmann has a lot of experience in the regulation of nuclear and nanotechnology and other committee members are philosophers within the Just War Theory tradition.
“We decided the best way to get some kind of arms control was to engage the international community and have invited a lot of knowledgeable people to the conference to get the ball rolling.”
Whilst ICRAC focuses on both teleoperated (i.e. ‘man-in-the-loop’) and fully autonomous combat vehicles, Professor Sharkey believes the two issues will become inseparable in the future:
“It will be unclear and nobody will make it clear whether UAVs and UGVs are remotely piloted, whether there is just someone working the weapons or whether there’s nobody working the weapons at all.”
He cites Carnegie Mellon’s Crusher, BAE’s Taranis and South Korea’s deployment of SGR-A1 sentry robots as examples of machines capable of autonomy, in which the public have no idea whether there is a human-in-the-loop or not.
Last Saturday, Oxford-based NGO Fellowship of Reconciliation launched their own anti-drone campaign at a one-day workshop called ‘Drone Wars’. The conference considered the history, legality and domestic implications of autonomous and semi-autonomous UAVs, bringing together academics and campaigners from across the country.
Both nationally and internationally, the humans are organising!Like this article? Sign up to Left Foot Forward's weekday email for the latest progressive news and comment - and support campaigning journalism by making a donation today.