Autonomous Weapon Systems (AWS) are no longer limited to science fiction. Conflicts in the Ukraine and Gaza demonstrate an increased trend toward the use of autonomy in the use of force in armed conflict. This book analyses the art 36 legal review obligation and assesses how states can determine the legality of AWS. It proposes a new ‘functional’ approach to legal review that considers both weapons law and targeting rules engaged by the autonomous functionality.
This publication considers lethal autonomous weapon systems, approaching the issue from five different perspectives. It has been published ahead of the first meeting of the Group of Governmental Experts of the High Contracting Parties to the Convention on Certain Conventional Weapons mandated to examine issues related to emerging technologies in the area of lethal autonomous weapon systems in the context of the objectives and purposes of the Convention. The United Nations Office for Disarmament Affairs Occasional Papers are a series of ad hoc publications featuring, in edited form, papers or statements made at meetings, symposiums, seminars, workshops or lectures that deal with topical issues in the field of arms limitation, disarmament and international security.
The DSB Task Force on the Role of Autonomy in DoD Systems was asked to study relevant technologies, ongoing research, and the current autonomy-relevant plans of the Military Services, to assist the DoD in identifying new opportunities to more aggressively use autonomy in military missions, to anticipate vulnerabilities, and to make recommendations for overcoming operational difficulties and systemic barriers to realizing the full potential of autonomous systems. The Task Force has concluded that, while currently fielded unmanned systems are making positive contributions across DoD operations, autonomy technology is being underutilized as a result of material obstacles within the Department that are inhibiting the broad acceptance of autonomy and its ability to more fully realize the benefits of unmanned systems. Overall, the Task Force found that unmanned systems are making a significant, positive impact on DoD objectives worldwide. However, the true value of these systems is not to provide a direct human replacement, but rather to extend and complement human capability by providing potentially unlimited persistent capabilities, reducing human exposure to life threatening tasks, and with proper design, reducing the high cognitive load currently placed on operators/supervisors.Unmanned systems are proving to have a significant impact on warfare worldwide. The true value of these systems is not to provide a direct human replacement, but rather to extend and complement human capability in a number of ways. These systems extend human reach by providing potentially unlimited persistent capabilities without degradation due to fatigue or lack of attention. Unmanned systems offer the warfighter more options and flexibility to access hazardous environments, work at small scales, or react at speeds and scales beyond human capability. With proper design of bounded autonomous capabilities, unmanned systems can also reduce the high cognitive load currently placed on operators/supervisors. Moreover, increased autonomy can enable humans to delegate those tasks that are more effectively done by computer, including synchronizing activities between multiple unmanned systems, software agents and warfighters--thus freeing humans to focus on more complex decision making.1.0 Executive Summary * 1.1. Misperceptions about Autonomy are Limiting its Adoption * 1.2. Create an Autonomous Systems Reference Framework to Replace "Levels of Autonomy" * 1.3. Technical Challenges Remain, Some Proven Autonomy Capability Underutilized * 1.4. Autonomous Systems Pose Unique Acquisition Challenges * 1.5. Avoid Capability Surprise by Anticipating Adversary Use of Autonomous Systems * 2.0 Operational Benefits of Autonomy * 2.1. Unmanned Aerial Vehicles * 2.2. Unmanned Ground Systems * 2.3. Unmanned Maritime Vehicles * 2.4. Unmanned Space Systems * 2.5. Conclusion * 3.0 Technical Issues of Autonomy * 3.1. Motivation: What Makes Autonomy Hard * 3.2. Defining Levels of Autonomy is Not Useful * 3.3. Autonomous System Reference Framework * 3.4. Needed Technology Development * 3.5. Technical Recommendations * 4.0 Acquisition Issues of Autonomy * 4.1. Requirements and Development * 4.2. Test and Evaluation * 4.3. Transition to Operational Deployment * 5.0 Capability Surprise in Autonomy Technology * 5.1. Overview of Global Unmanned Systems * 5.2. Unmanned Symmetric Adversary Scenarios * 5.3. Value for Asymmetric Adversaries * 5.4. External Vulnerabilities * 5.5. Self-Imposed Vulnerabilities * 5.6. Recommendations . * Appendix A--Details of Operational Benefits by Domain * A.1. Aerial Systems Strategy * A.2. Maritime Systems * A.3. Ground Systems * A.4. Space Systems * Appendix B--Bibliography * Appendix C--Task Force Terms of Reference * Appendix D--Task Force Membership * Appendix E--Task Force Briefings * Appendix F--Glossary
Military robots and other, potentially autonomous robotic systems such as unmanned combat air vehicles (UCAVs) and unmanned ground vehicles (UGVs) could soon be introduced to the battlefield. Look further into the future and we may see autonomous micro- and nanorobots armed and deployed in swarms of thousands or even millions. This growing automation of warfare may come to represent a major discontinuity in the history of warfare: humans will first be removed from the battlefield and may one day even be largely excluded from the decision cycle in future high-tech and high-speed robotic warfare. Although the current technological issues will no doubt be overcome, the greatest obstacles to automated weapons on the battlefield are likely to be legal and ethical concerns. Armin Krishnan explores the technological, legal and ethical issues connected to combat robotics, examining both the opportunities and limitations of autonomous weapons. He also proposes solutions to the future regulation of military robotics through international law.
The new means and methods of warfare challenge the law of armed conflict. For this reason, a deep legal and operational study was required to detect the existing gaps and deregulation.
"This 50-page report outlines concerns about these fully autonomous weapons, which would inherently lack human qualities that provide legal and non-legal checks on the killing of civilians. In addition, the obstacles to holding anyone accountable for harm caused by the weapons would weaken the law's power to deter future violations"--Publisher's website.
"Because of the increasing use of Unmanned Aerial Vehicles (UAVs, also commonly known as drones) in various military and para-military (i.e., CIA) settings, there has been increasing debate in the international community as to whether it is morally and ethically permissible to allow robots (flying or otherwise) the ability to decide when and where to take human life. In addition, there has been intense debate as to the legal aspects, particularly from a humanitarian law framework. In response to this growing international debate, the United States government released the Department of Defense (DoD) 3000.09 Directive (2011), which sets a policy for if and when autonomous weapons would be used in US military and para-military engagements. This US policy asserts that only "human-supervised autonomous weapon systems may be used to select and engage targets, with the exception of selecting humans as targets, for local defense ...". This statement implies that outside of defensive applications, autonomous weapons will not be allowed to independently select and then fire upon targets without explicit approval from a human supervising the autonomous weapon system. Such a control architecture is known as human supervisory control, where a human remotely supervises an automated system (Sheridan 1992). The defense caveat in this policy is needed because the United States currently uses highly automated systems for defensive purposes, e.g., Counter Rocket, Artillery, and Mortar (C-RAM) systems and Patriot anti-missile missiles. Due to the time-critical nature of such environments (e.g., soldiers sleeping in barracks within easy reach of insurgent shoulder-launched missiles), these automated defensive systems cannot rely upon a human supervisor for permission because of the short engagement times and the inherent human neuromuscular lag which means that even if a person is paying attention, there is approximately a half-second delay in hitting a firing button, which can mean the difference for life and death for the soldiers in the barracks. So as of now, no US UAV (or any robot) will be able to launch any kind of weapon in an offensive environment without human direction and approval. However, the 3000.09 Directive does contain a clause that allows for this possibility in the future. This caveat states that the development of a weapon system that independently decides to launch a weapon is possible but first must be approved by the Under Secretary of Defense for Policy (USD(P)); the Under Secretary of Defense for Acquisition, Technology, and Logistics (USD(AT&L)); and the Chairman of the Joint Chiefs of Staff. Not all stakeholders are happy with this policy that leaves the door open for what used to be considered science fiction. Many opponents of such uses of technologies call for either an outright ban on autonomous weaponized systems, or in some cases, autonomous systems in general (Human Rights Watch 2013, Future of Life Institute 2015, Chairperson of the Informal Meeting of Experts 2016). Such groups take the position that weapons systems should always be under "meaningful human control," but do not give a precise definition of what this means. One issue in this debate that often is overlooked is that autonomy is not a discrete state, rather it is a continuum, and various weapons with different levels of autonomy have been in the US inventory for some time. Because of these ambiguities, it is often hard to draw the line between automated and autonomous systems. Present-day UAVs use the very same guidance, navigation and control technology flown on commercial aircraft. Tomahawk missiles, which have been in the US inventory for more than 30 years, are highly automated weapons with accuracies of less than a meter. These offensive missiles can navigate by themselves with no GPS, thus exhibiting some autonomy by today's definitions. Global Hawk UAVs can find their way home and land on their own without any human intervention in the case of a communication failure. The growth of the civilian UAV market is also a critical consideration in the debate as to whether these technologies should be banned outright. There is a $144.38B industry emerging for the commercial use of drones in agricultural settings, cargo delivery, first response, commercial photography, and the entertainment industry (Adroit Market Research 2019) More than $100 billion has been spent on driverless car development (Eisenstein 2018) in the past 10 years and the autonomy used in driverless cars mirrors that inside autonomous weapons. So, it is an important distinction that UAVs are simply the platform for weapon delivery (autonomous or conventional), and that autonomous systems have many peaceful and commercial uses independent of military applications"--
The deployment of Autonomous Weapons gives rise to ongoing debate in society and at the United Nations, in the context of the Convention on Certain Conventional Weapons. Yet little empirical research has been done on this topic. This volume fills that gap by offering an empirical study based on military personnel and civilians working at the Dutch Ministry of Defence. It yields insight into how Autonomous Weapons are perceived by the military and general public; and which moral values are considered important in relation to their deployment. The research approach used is the Value-Sensitive Design (VSD) method that allows for the consideration of human values throughout the design process of technology. The outcome indicates that military personnel and civilians attribute more agency (the capacity to think and plan) to an Autonomous Weapon than to a Human Operated Drone. In addition, it is clear that common ground exists between military and societal groups in their perception of the values of human dignity and anxiety. These two values arise often in the discourse, and addressing them is essential when considering the ethics of the deployment of Autonomous Weapons. The text of this volume is also offered in parallel French and German translation.
The main theme of this volume of the Yearbook of International Humanitarian Law is weapons law. In several chapters, how International Humanitarian Law (IHL) copes with old and new weapons as well as political developments in regard to military technology is discussed, while in two chapters the significance of non- or less-lethal weapons in peace-keeping and law enforcement operations as well as the legality of lethal autonomous weapon systems under IHL are analysed. Moreover, the volume describes the current status of nuclear deterrence under international law. Another layer is added by examining how IHL influences the programming of automatic target recognition systems using artificial intelligence. The second part of the book contains a historic perspective on the roots of IHL in Europe, which can be traced back to the ninth century, as well as a Year in Review describing the most important events and legal developments in the area of IHL that took place in 2018. The Yearbook of International Humanitarian Law is the world’s only annual publication devoted to the study of the laws governing armed conflict. It provides a truly international forum for high-quality, peer-reviewed academic articles focusing on this crucial branch of international law. Distinguished by contemporary relevance, the Yearbook of International Humanitarian Law bridges the gap between theory and practice and serves as a useful reference tool for scholars, practitioners, military personnel, civil servants, diplomats, human rights workers and students.
The variety, pace, and power of technological innovations that have emerged in the 21st Century have been breathtaking. These technological developments, which include advances in networked information and communications, biotechnology, neurotechnology, nanotechnology, robotics, and environmental engineering technology, have raised a number of vital and complex questions. Although these technologies have the potential to generate positive transformation and help address 'grand societal challenges', the novelty associated with technological innovation has also been accompanied by anxieties about their risks and destabilizing effects. Is there a potential harm to human health or the environment? What are the ethical implications? Do this innovations erode of antagonize values such as human dignity, privacy, democracy, or other norms underpinning existing bodies of law and regulation? These technological developments have therefore spawned a nascent but growing body of 'law and technology' scholarship, broadly concerned with exploring the legal, social and ethical dimensions of technological innovation. This handbook collates the many and varied strands of this scholarship, focusing broadly across a range of new and emerging technology and a vast array of social and policy sectors, through which leading scholars in the field interrogate the interfaces between law, emerging technology, and regulation. Structured in five parts, the handbook (I) establishes the collection of essays within existing scholarship concerned with law and technology as well as regulatory governance; (II) explores the relationship between technology development by focusing on core concepts and values which technological developments implicate; (III) studies the challenges for law in responding to the emergence of new technologies, examining how legal norms, doctrine and institutions have been shaped, challenged and destabilized by technology, and even how technologies have been shaped by legal regimes; (IV) provides a critical exploration of the implications of technological innovation, examining the ways in which technological innovation has generated challenges for regulators in the governance of technological development, and the implications of employing new technologies as an instrument of regulatory governance; (V) explores various interfaces between law, regulatory governance, and new technologies across a range of key social domains.