Considerations for the Study of Combat-Capable Robots. William Shelby

Similar documents
The challenges raised by increasingly autonomous weapons

RUNNING HEAD: Drones and the War on Terror 1. Drones and the War on Terror. Ibraheem Bashshiti. George Mason University

The use of armed drones must comply with laws

Preface to "What Principles Should Guide America's Conduct of War?" on Opposing Viewpoints,

Conference panels considered the implications of robotics on ethical, legal, operational, institutional, and force generation functioning of the Army

Preventing harm from the use of explosive weapons in populated areas

International Humanitarian Law and New Weapon Technologies

Academic Year

Princeton University Jan. 23, 2015 Dr. Maryann Cusimano Love

Autonomous weapons systems as WMD vectors a new threat and a potential for terrorism?

Drones and the Threshold for Waging War

ODUMUNC 39. Disarmament and International Security Committee. The Challenge of Lethal Autonomous Weapons Systems

Tren ds i n Nuclear Security Assessm ents

Drones and the Threshold for Waging War Ezio Di Nucci (University of Copenhagen)

FUTURE WAR WAR OF THE ROBOTS?

ENGINEERING A TRAITOR

NATIONAL DEFENSE AND SECURITY ECONOMICS

Mad, Mad Killer Robots By Lieutenant Colonel David W. Szelowski, USMCR (Ret.)

Jürgen Altmann: Uninhabited Systems and Arms Control

Selective obscenity : US checkered record on chemical weapons RT News

Duck & Cover: School Drills During the Cold War By Jessica McBirney 2016

Adapting for Unmanned Systems

(20%) account for an exorbitant. during this time period. Certainly,

Conflict Horizon Dallas Walker Conflict Horizon

Armin Krishnan: Ethical and Legal Challenges

Strategic Studies Seminar

AI for Global Good Summit. Plenary 1: State of Play. Ms. Izumi Nakamitsu. High Representative for Disarmament Affairs United Nations

CBC Learning authorizes the reproduction of material contained in this resource guide for educational purposes. Please identify the source.

PUBLICATIONS Books: (2018) Why Paramilitary Operations Fail, New York: Palgrave Macmillan (forthcoming).

Stars War: Peace, War, and the Legal (and Practical) Limits on Armed Conflict in Space

P U R D U E U N I V E R S I T Y. POL 237: MODERN WEAPONS AND INTERNATIONAL RELATIONS Spring 2015

Airplane. Estimated Casualty Statistics for the Battle of Tannenberg Allied Powers: 267,000 Central Powers: 80,000

ALMOST NUCLEAR: INTRODUCING THE NUCLEAR LATENCY DATASET

PROFILE. Jonathan Sherer 9/30/15 1

CONTENTS INTRODUCTION Compass Games, LLC. Don t fire unless fired upon, but if they mean to have a war, let it begin here.

Economic Aspects of the Military Robotics Introduction and Development in the War Theatre

PHIL 20628/STV 20228/IIPS 20912/HESB Ethics of Emerging Weapons Technologies Spring 2017

POSC 236 Global, National and Human Security --- Spring Class time: Thursday, Thursday 1:15-3:00. Weitz Center 233. Prof Greg Marfleet

Future of Financing. For more information visit ifrc.org/s2030

General Claudio GRAZIANO

When it comes to generic 25mm Science Fiction skirmish games, there are really only two choices.

Keynote address by Dr Jakob Kellenberger, ICRC President, and Conclusions by Dr Philip Spoerri, ICRC Director for International Law and Cooperation

Key elements of meaningful human control

A 12-part LRC series, featuring text and iphone Hipstamatic photography by Rita Leistner

PROFILE. Jonathan Sherer 9/10/2015 1

Why Record War Casualties?

Airplane. Estimated Casualty Statistics for the Battle of Tannenberg Allied Powers: 267,000 Central Powers: 80,000. Artillery

DoD Research and Engineering Enterprise

Mission Capability Packages

Table of Contents. Preface Acknowledgments

Big Picture for Autonomy Research in DoD

Science and Technology for Naval Warfare,

Report for the Drones and Forever War conference Saturday July 11 th at Friends House, London

High School Social Studies Grades 9 12

THE GAME RULES. This megagame is about the emergence of civil war in a fictional African country.

Convention on Certain Conventional Weapons (CCW) Meeting of Experts on Lethal Autonomous Weapons Systems (LAWS) April 2016, Geneva

DoD Research and Engineering Enterprise

Heidi Robinson Today, I m going to talk to you about resiliency. Resiliency is not a term that is easily defined nor is it easily achievable. As I con

ROBOTICS IN FUTURE WARFARE

HISTORY of AIR WARFARE

Disarmament and Arms Control An overview of issues and an assessment of the future

Governing Lethal Behavior: Embedding Ethics in a Hybrid Reactive Deliberative Architecture

Primo Victoria. A fantasy tabletop miniatures game Expanding upon Age of Sigmar Rules Compatible with Azyr Composition Points

Building an Ethical Framework for the Implementation of AI into the UN's Armed Conflicts Prevention Agenda

HOW TO PLAY This megagame is about the emergence of civil war in a fictional African country.

CalsMUN 2019 Future Technology. General Assembly 1. Research Report. The use of autonomous weapons in combat. Marije van de Wall and Annelieve Ruyters

Ground Robotics Market Analysis

THE NATO C3 AGENCY SUPPORT TO THE IED FIGHT: A COMPREHENSIVE APPROACH

LATE 19 th CENTURY WARGAMES RULES Based on and developed by Bob Cordery from an original set of wargames rules written by Joseph Morschauser

The Manhattan Project (NCSS8)

Whole of Society Conflict Prevention and Peacebuilding

Strategic Partner of the Report

Electronic Warfare Training in the Pacific Northwest

Challenging the Future with Ubiquitous Distributed Control

EMBEDDING THE WARGAMES IN BROADER ANALYSIS

Stanford Center for AI Safety

Lead Screws for Guided Munitions

What Is a Global Superpower? One World Student Ambassador: Nathan Araya. Unit Resource #2: Understanding the Reflection

Explosive Ordnance Disposal/ Low-Intensity Conflict. Improvised Explosive Device Defeat

A Thunderbolt + Apache Leader TDA

Biometrics redefining the phrase 'don't shoot until you see the whites of their eyes'

The Imperative for Change

Artificial Intelligence: Implications for Autonomous Weapons. Stuart Russell University of California, Berkeley

MEDIA KIT

Steven P. Andreasen Bruce G. Blair Matthew Bunn Sidney D. Drell

oids: Towards An Ethical Basis for Autonomous System Deployment

A NEW SIMULATION FRAMEWORK OF OPERATIONAL EFFECTIVENESS ANALYSIS FOR UNMANNED GROUND VEHICLE

BAPTISM OF FIRE III WW2 SKIRMISH. Sample file. Written by Gordon Roach, 2009 V.1.0. Copyright 2009 Gordon Roach, TFP Games v1.0

LATE 19 th CENTURY WARGAMES RULES Based on and developed by Bob Cordery from an original set of wargames rules written by Joseph Morschauser

Accelerating innovations in science and technology (S&T) are having profound effects on global civilization These developments will have strategic

The SMArt 155 SFW. Is it reasonable to refer to it as a cluster munition?

PLS 302 Syllabus. Dr. Aspin (aspin at bradley.edu) World Security 488 Bradley ( )

Deadly Cargo. Explosive Weapons in Populated Areas. By Steve Goose and Ole Solvang

SACT remarks at. Atlantic Council SFA Washington DC, George Washington University, Elliott School of International Affairs

What s wrong with drones?

Artificial Intelligence. Robert Karapetyan, Benjamin Valadez, Gabriel Garcia, Jose Ambrosio, Ben Jair

Countering Capability A Model Driven Approach

Autonomous Weapons Potential advantages for the respect of international humanitarian law

How is this novel about the threat of the lead character s demise? Choose at least one. Ethan s in the military. Frequent battle is life and death.

MODERN AMERICA BUSH, CLINTON, BUSH, OBAMA, TRUMP

Transcription:

Considerations for the Study of Combat-Capable Robots William Shelby Advances in robotics over the past decade have brought the militaries of science fiction to life. Robots are relatively cheap to produce and operate, their strikes are precise, and they involve no risk to the people who control them. Because robot operators are not at risk, these machines create a safer environment for soldiers. Still, scholars question whether the use of robots will incentivize governments to use force rather than diplomacy to achieve their goals. Since there are fewer risks to the lives of soldiers when force is executed by robots than when force is executed directly by soldiers, it is possible that government actors may resort to using force rather than negotiation because their ends are cheaply and quickly achieved by force. The implications for national security are clear: if robots make actors more willing to use force, then government actors will be less likely to use soft power like diplomacy to influence international relations. Defining and Describing Robots Robots can broadly be defined as a machine that can take in information, process it, and act on it (Singer 2009 67). This includes, but is not limited to, machines controlled by humans. Robots come in a variety of shapes, sizes, and forms; they can resemble humans or look like machines. Robots are used to complete jobs that are dirty, dull, or dangerous (Singer 2009 63, Schornig 2010 2). Although robots can be employed to execute any number of these jobs, this paper focuses on the importance and expanded use of robots for military operations.

Robots are useful for militaries because they reduce costs and provide political cover for leaders using military force (Kreps 2016). One of the most common robots currently used by the U.S. military is the unmanned aircraft system (UAS). UAS used for combat missions are cheaper than comparable manned aircraft (Mizokami 2016, Zegart 2015). More importantly, UAS take American soldiers off the battlefield, completely removing the risk of injury or mortality to the soldiers (Kaag and Kreps 2014, Singer 2009). Individuals piloting UAS are thousands of miles from the combat zone and have zero chance of getting killed by an enemy. Robots combine the lack of risk to soldiers lives with the ability to have a close presence to the battlefield, ensuring that military strikes are precise and achieve their goal. Using robots can also limit indiscriminate killing by ensuring the correct target is hit, and technology is advancing to make hitting the correct target a near certainty (Rothenberg 2016). Robots of the future will have facial scanning software that ensures they have located the correct individual (Dillow 2011). Assuming a kinetic strike has not yet occurred, robots can also abort an attack like soldiers choosing not to engage if they determine the target is incorrect (Zenko 2013). These close-up precision capabilities allow robots to be more just than conventional military operations (Emmerson 2013, Strawser 2010). Robots have the capabilities to become proficient soldiers that lack risks inherent with human soldiers. This creates political cover for leaders looking to intervene militarily in an international crisis (Plaw, Fricker, and Colon 2016). Engaging in conflict is costly to a state in terms of lives and resources lost (Bueno de Mesquita et al 2003, Eichenberg 2005, Perla Jr. 2011). Over long periods of time, these costs may become unbearable for a population, or the public may not see the utility of such actions. However, if robots could fight these conflicts without creating a risk to the citizens of the state engaging in a military operation, then the

citizens may become indifferent to military operations because they do not incur the costs (Kaag and Kreps 2014, Kreps 2016). However, robot technology has not yet evolved to the point where conflict is completely riskless. There is near consensus that technology like UAS are not currently riskless because they are cost-prohibitive for many actors and cannot operate in denied areas (places where there is a strong air defense) (Sayler 2014, Stohl, Brooks, and Abizaid 2014). Still, the key point is how the technology appears to the public. If people think that wars fought by robots have no consequences, then they are likely to support the use of force exerted by robots. If one could engage in an action with few risks which, if successful, could create a positive benefit, the public would likely support the action (Kaag and Kreps 2014). The Double-Edged Sword Robots have the power to create positive benefits for the international system. As a world superpower, the United States is frequently tasked with addressing security issues such as terrorism and humanitarian crises. However, humanitarian crises are often perceived as a waste of time by a public that does not want to use resources towards a conflict to which they will see little benefit (Eichenberg 2005, Jentleson 1992, Perla Jr. 2011). However, if U.S. citizens are indifferent about sending robots to respond to a humanitarian crisis, the United States could intervene and prevent future problems. Empirically, evidence suggests that this could occur. Walsh and Schulzke (2015) showed for example, that individuals preferred the use of UAS to sending in ground force in instances where the public did not show high levels of support for military intervention.

Yet, robots are a double-edged sword. Robots could also be used to fight internal conflicts or be used to escalate tensions in sensitive areas (Kreps 2016). For example, robots could facilitate the ease with which authoritarian governments maintain domestic control (Horowitz, Kreps, and Fuhrmann 2016). In Syria, Bashar al-assad has used the Syrian army to kill civilians. Many people in the Syrian army are not willing to kill civilians and end up deserting. 1 This would not happen with robots. Robots would continue killing without regard because they lack emotional capacity. These scenarios are just examples of the myriad ways in which robots could be used. The important point is that the technology itself allows for endless possibilities. The world could use robots to end humanitarian crises and make the world a better place, or some individuals could abuse the characteristics of robots that make them so appealing to further their individual goals. Current Research on Robots and New Directions Moving Forward With the likelihood of robotic warfare increasing, researchers have begun to consider the role of robots in conflict, with research focusing on the use of UAS as a useful starting point. Walsh (2015) looked at the role of American and civilian casualties on support for the use of force using robots, finding that preventing American casualties increased support for the use of force while increasing civilian casualties decreased this support. Horowitz (2016) found some similar results when discussing the autonomous robots. Subjects in this study supported the use of UAS at higher levels if their use could protect U.S. soldiers. 1 See http://www.independent.co.uk/news/world/middle-east/aleppo-latest-battle-un-syrian-army-forcesgovernment-regime-kill-civilians-massacre-assad-russia-a7471416.html and http://www.cnn.com/2011/world/meast/07/05/syria.sniper/index.html

Other studies have considered factors that inhibit support for robots. Kreps (2014) found that arguments about how UAS could violate international humanitarian law had a strong impact on support for the use of drones. Kreps and Wallace (2016) showed that arguments about international law could shape public support for UAS more so than discussing the effectiveness of UAS strikes. While these studies have increased our understanding about preference for robots over human soldiers, none of them focus on the central characteristic of robots: they reduce the risk of using military force. None of these studies have attempted to capture the explicit perceived risk of using UAS or robots. There is a consensus that drastically reducing risk is the causal mechanism that makes robots different from previous technological advances in warfare, yet no one has explicitly measured risk perceptions and how it relates to support for the use of force. Moving forward, it will be important to understand the relationship between risk and support when using robots. As stated above, robotic technology does not currently allow actors to use robots to operate in denied theaters of war, but advances in technology make this a strong possibility in the future (Sayler 2014, Stohl, Brooks, and Abizaid 2014). When the threat of losing too many soldiers does not deter actors from using force, how can the United States ensure that other countries do not turn to military force first rather than diplomacy? Do the laws of war need to change? Do costs shift to other considerations when soldiers are not at risk? Is there a limit to how much the public would support the use of force using robots? These questions are relevant, and need to be discussed now, but there are few cases of states using robots against each other. To mitigate this lack of data, researchers should continue the strategy used by the researchers mentioned above: create hypothetical scenarios and measure how individuals react to each scenario. These scenarios can inform researchers and policymakers

how the public will perceive future military operations without having wait until such a scenario occurs (Croson 2002). While polling data and case studies are useful, there are not enough cases concerning actors using robots in a manner commensurate with these types of studies. The United States is the largest user of robots in military operations, but without a clear comparison to another actor, a case study will not apply generally and will only provide understanding of specific instances. (Hopkins 2010). A large-n study using only data from the United States does not provide sufficient data from which to draw general conclusion. In either case, it is difficult to test whether there is a causal mechanism behind the variables of interest (i.e, risk and support). Therefore, the experimental method is most appropriate to answer the policy concerns currently under study. Experiments serve this need by allowing researchers to establish causality through careful establishment of steps to show that variable X causes Y (McDermott 2002). Experiments allow researchers to explore relationships that will benefit our understanding of using robots faster than using other methods of analysis. The Implications of Robot Militaries The importance of studying robots right now lies in the rate at which the technology is proliferating throughout the world. Again turning to UAS, the New America Foundation (Bergen et al 2017) estimates that at least 26 states have UAS that are capable of engaging in combat operations. Half of those states did not have combat-capable UAS five years ago. In addition, 5 non-state actors, most-notably the Islamic State of Iraq and Ash Sham (ISIS), developed their own rudimentary technology capable of dropping bombs and being used as an improvised explosive device. Even if actors do not have the technological capabilities of the most powerful militaries, they still have a new manner of furthering their goals without having to rely on individuals to join their cause; robots are an equalizer.

Conclusion Robots are poised to revolutionize society by reducing risks associated with dangerous tasks, but this reduction in risk can also remove the constraints on using force. If states can carry out military operations with robots, they may be poised to use force rather than rely on diplomacy. This is not the current reality, as current technology limits the efficacy of robots, but it is important to understand how using robots affects the perception of using force so that we can discover methods to mitigate the possibility of increased, robotic conflict. The issues raised in this paper are just a sample of the problems that may arise from using robots for combat but provide a good starting point to begin to understand how robots will affect international relations.

References Bergen, Peter, David Sternman, Alyssa Sims, Albert Ford, and Christopher Mellon. 2017. World of Drones: Examining the Proliferation, Development, and Use of Armed Drones. Washington DC: New America Foundation. https://www.newamerica.org/indepth/world-of-drones/ Bueno de Mesquita, Bruce, Alastair Smith, Randolph Siverson and James Morrow. 2003. The Logic of Political Survival. Cambridge: MIT Press. Croson, Rachel. 2002. "Why and How to Experiment: Methodologies from Experimental Economics." University of Illinois Law Review: 921-946. Dillow, Clay. 2011. Army Developing Drones That Can Recognize Your Face From a Distance And even recognize your intentions. September28. http://www.popsci.com/technology/article/2011-09/army-wants-drones-can-recognizeyour-face-and-read-your-mind. Accessed February 10, 2016. Eichenberg, Richard. 2005."Victory Has Many Friends: US Public Opinion and the Use of Force 1985-2001." International Security 30(1): 140-177. Emmerson, Ben. 2013. Promotion and protection of human rights and fundamental freedoms while countering terrorism. United Nations. Available at http://www.ohchr.org/en/issues/terrorism/pages/srterrorismindex.aspx. Hopkin, Jonathan. 2010. Comparative Methods. In David Marsh and Gerry Stoker's (eds) Theory and Methods in Political Science. New York: Palgrave Macmillan. Horowitz, Michael, Sarah Kreps, and Michael Fuhrmann. 2016. The Consequence of Drone Proliferation: Separating Fact from Fiction. Social Science Research Network. http://papers.ssrn.com/sol3/papers.cfm?abstract_id=2722311. Horowitz, Michael. 2016. Public Opinion and the Politics of the Killer Robot Debate. Research and Politics 3(1): 1-8 Jentleson, Bruce. 1992. The Pretty Prudent Public: Post-Vietnam American Opinion on the Use of Military Force. International Studies Quarterly 36(1): 49-73. Kaag, John and Sarah Kreps. 2014. Drone Warfare. Malden, MA: Polity Press. Kreps, Sarah and Geoffrey Wallace. 2016. International Law, Elites, and Public Support for Drones. International Organization 53(6): 830-844. Kreps, Sarah. 2014. Flying under the radar: a study of public attitudes towards unmanned aerial vehicles. Research and Politics 1(1): 1-7. Kreps, Sarah. 2016. Drones: What Everyone Needs to Know. New York: Oxford University Press.

McDermott, Rose. 2002. Experimental Methodology in Political Science." Political Analysis 10(4):325-61. Mizokami, Kyle. 2016. The F-35 Is about to Get a Lot Cheaper. Sort Of. July 11. http://www.popularmechanics.com/military/weapons/a21776/f-35-cheaper/. Accessed December 20, 2016. Perla Jr, Hector. 2011. Explaining Public Support for the Use of Military Force: The Impact of Reference Point Framing and Prospective Decision Making. International Organization 65(4): 139-167. Plaw, Avery, Matthew Fricker, and Carlos Colon. 2016. The Drone Debate: A Primer on the U.S. Use of Unmanned Aircraft Outside Conventional Battlefields. New York: Rowman and Littlefield. Rothenberg, David. 2016. Drones and the Emergence of Data-Driven Warfare. In Peter Bergen and Daniel Rothenberg s (eds) Drone Wars: Transforming Conflict, Law, and Policy. New York: Cambridge University Press. Sayler, Kelley. 2014. A World of Proliferated Drones: A Technology Primer. Washington DC: Center for New American Security. Schornig, Niklas. 2010. Robot Warriors: Why the Western Investment into Military UAS Might Backfire. Frankfurt: Peace Research Institute. Singer, Peter. 2009. Wired for War: The Robotics Revolution and Conflict in the 21st Century. New York: Penguin Press. Kindle Edition. Stohl, Rachel, Rosa Brooks, and John P. Abizaid. 2015. Recommendations and Report of The Tasks Force On US Drone Policy. Washington, D.C.: Stimson Center. Strawser, Bradley. 2010. Moral Predators: The Duty to Employ Uninhabited Aerial Vehicles, Journal of Military Ethics 9(4): 342 68, 346. Walsh, James, and Mark Schulzke. 2015. The Ethics of Drone Strikes: Does Reducing the Cost of Conflict Encourage War? Strategic Studies Institute. US Army War College. Walsh, James. 2015. "Precision Weapons, Civilian Casualties, and Support for the Use of Force." Political Psychology 36(5): 507-523. Whittle, Richard. 2014. Predator: The Secret Origins of the Drone Revolution. New York: Henry Holt and Company. Kindle Edition. Zegart, Amy. 2015. The Coming Revolution of Drone Warfare. March 15. Wall Street Journal. Zenko, Micah. 2013. "Reforming U.S. Drone Strike Policies." Chicago Council on Foreign Relations.