Drones will soon decide who to kill

Peter Lee, University of Portsmouth
Algorithms will soon be able to decide who to target. Photo courtesy of U.S. Air Force
Algorithms will soon be able to decide who to target. Photo courtesy of U.S. Air Force

April 11 (UPI) -- The U.S. Army recently announced that it is developing the first drones that can spot and target vehicles and people using artificial intelligence. This is a big step forward. Whereas current military drones are still controlled by people, this new technology will decide who to kill with little human involvement.


Once complete, these drones will represent the ultimate militarization of AI and trigger vast legal and ethical implications for wider society. There is a chance that warfare will move from fighting to extermination, losing any semblance of humanity in the process. At the same time, it could widen the sphere of warfare so that the companies, engineers and scientists building AI become valid military targets.

Existing lethal military drones like the MQ-9 Reaper are carefully controlled and piloted via satellite. If a pilot drops a bomb or fires a missile, a human sensor operator actively guides it onto the chosen target using a laser.

RELATED Employees protest Google work for Pentagon drone strike program

Ultimately, the crew has the final ethical, legal and operational responsibility for killing designated human targets. As one Reaper operator states: "I am very much of the mindset that I would allow an insurgent, however important a target, to get away rather than take a risky shot that might kill civilians."


Even with these drone killings, human emotions, judgments and ethics have always remained at the center of war. The existence of mental trauma and post-traumatic stress disorder among drone operators shows the psychological impact of remote killing.

And this actually points to one possible military and ethical argument by Ronald Arkin, in support of autonomous killing drones. Perhaps if these drones drop the bombs, psychological problems among crew members can be avoided. The weakness in this argument is that you don't have to be responsible for killing to be traumatized by it. Intelligence specialists and other military personnel regularly analyze graphic footage from drone strikes. Research shows that it is possible to suffer psychological harm by frequently viewing images of extreme violence.

RELATED Air Force contracts with SRC for drone supplies, services

When I interviewed over 100 Reaper crew members for an upcoming book, every person I spoke to who conducted lethal drone strikes believed that, ultimately, it should be a human who pulls the final trigger. Take out the human and you also take out the humanity of the decision to kill.

Grave consequences

The prospect of totally autonomous drones would radically alter the complex processes and decisions behind military killings. But legal and ethical responsibility does not somehow just disappear if you remove human oversight. Instead, responsibility will increasingly fall on other people, including artificial intelligence scientists.

RELATED Lockheed Martin to upgrade weapons systems on naval destroyers

The legal implications of these developments are becoming evident. Under current international humanitarian law, "dual-use" facilities -- those that develop products for both civilian and military application -- can be attacked in the right circumstances. For example, in the 1999 Kosovo War, the Pancevo oil refinery was attacked because it could fuel Yugoslav tanks as well as fuel civilian cars.

With an autonomous drone weapon system, certain lines of computer code would probably be classed as dual-use. Companies like Google, its employees or its systems, could become liable to attack from an enemy state. For example, if Google's Project Maven image recognition AI software is incorporated into an American military autonomous drone, Google could find itself implicated in the drone "killing" business, as might every other civilian contributor to such lethal autonomous systems.

Ethically, there are even darker issues. The whole point of the self-learning algorithms -- programs that independently learn from whatever data they can collect -- that technology uses is that they become better at whatever task they are given. If a lethal autonomous drone is to get better at its job through self-learning, someone will need to decide on an acceptable stage of development -- how much it still has to learn -- at which it can be deployed. In militarized machine learning, that means political, military and industry leaders will have to specify how many civilian deaths will count as acceptable as the technology is refined.


Recent experiences of autonomous AI in society should serve as a warning. Uber's fatal experiment with a self-driving Tesla pretty much guarantees that there will be unintended autonomous drone deaths as computer bugs are ironed out.

The ConversationIf machines are left to decide who dies, especially on a grand scale, then what we are witnessing is extermination. Any government or military that unleashed such forces would violate whatever values it claimed to be defending. In comparison, a drone pilot wrestling with a "kill or no kill" decision becomes the last vestige of humanity in the often inhuman business of war.

Peter Lee is d director of security and risk and a reader in politics and ethics at the University of Portsmouth.

This article was originally published on The Conversation. Read the original article.

Latest Headlines