What happens to deterrence and escalation when decisions can be made at machine speeds and when states can put fewer human lives at risk? What are potential areas for miscalculation and unintended consequences, and unwanted escalation in particular? This exploratory report provides an initial examination of how artificial intelligence and autonomous systems could affect deterrence and escalation in conventional crises and conflicts.
Deterrence in the Age of Thinking Machines
Download eBook for Free
|PDF file||0.8 MB||
Use Adobe Acrobat Reader version 10 or higher for the best experience.
Purchase Print Copy
|Add to Cart||Paperback122 pages||$22.00||$17.60 20% Web Discount|
- What are the implications of adding thinking machines and autonomous systems to the practices that countries have developed to signal one another about the use of force and its potential consequences?
- What happens to deterrence and escalation when decisions can be made at machine speeds and are carried out by forces that do not risk human lives of the using state or actor?
- How might the rise of these capabilities weaken or strengthen deterrence?
- What are potential areas of miscalculation and unintended consequences?
The greater use of artificial intelligence (AI) and autonomous systems by the militaries of the world has the potential to affect deterrence strategies and escalation dynamics in crises and conflicts. Up until now, deterrence has involved humans trying to dissuade other humans from taking particular courses of action. What happens when the thinking and decision processes involved are no longer purely human? How might dynamics change when decisions and actions can be taken at machine speeds? How might AI and autonomy affect the ways that countries have developed to signal one another about the potential use of force? What are potential areas for miscalculation and unintended consequences, and unwanted escalation in particular?
This exploratory report provides an initial examination of how AI and autonomous systems could affect deterrence and escalation in conventional crises and conflicts. Findings suggest that the machine decisionmaking can result in inadvertent escalation or altered deterrence dynamics, due to the speed of machine decisionmaking, the ways in which it differs from human understanding, the willingness of many countries to use autonomous systems, our relative inexperience with them, and continued developments of these capabilities. Current planning and development efforts have not kept pace with how to handle the potentially destabilizing or escalatory issues associated with these new technologies, and it is essential that planners and decisionmakers begin to think about these issues before fielded systems are engaged in conflict.
Insights from a wargame involving AI and autonomous systems
- Manned systems may be better for deterrence than unmanned ones.
- Replacing manned systems with unmanned ones may not be seen as a reduced security commitment.
- Players put their systems on different autonomous settings to signal resolve and commitment during the conflict.
- The speed of autonomous systems did lead to inadvertent escalation in the wargame.
Implications for deterrence
- Autonomous and unmanned systems could affect extended deterrence and our ability to assure our allies of U.S. commitment.
- Widespread AI and autonomous systems could lead to inadvertent escalation and crisis instability.
- Different mixes of human and artificial agents could affect the escalatory dynamics between two sides.
- Machines will likely be worse at understanding the human signaling involved deterrence, especially deescalation.
- Whereas traditional deterrence has largely been about humans attempting to understand other humans, deterrence in this new age involves understanding along a number of additional pathways.
- Past cases of inadvertent engagement of friendly or civilian targets by autonomous systems may offer insights about the technical accidents or failures involving more-advanced systems.
- Conduct further work on deterrence theory and other frameworks to explicitly consider the potential effects of AI and autonomous systems.
- Evaluate the escalatory potential of new systems.
- Evaluate the escalatory potential of new operating concepts.
- Wargame additional scenarios at the operational and strategic levels.
Table of Contents
Key Deterrence Concepts
Artificial Intelligence and Autonomous Systems
Potential Futures in a World of Proliferated Artificial Intelligence and Autonomous Systems
A Wargame of Artificial Intelligence and Autonomous Systems
Wargame Insights and Debates
Implications for Deterrence
Implications for Decisionmaking
Conclusion and Areas for Further Research
General Morphological Analysis
Research conducted by
Funding for this research was provided by gifts from RAND supporters and income from operations. The research was conducted within the International Security and Defense Policy Center of the RAND National Defense Research Institute.
This report is part of the RAND Corporation Research report series. RAND reports present research findings and objective analysis that address the challenges facing the public and private sectors. All RAND reports undergo rigorous peer review to ensure high standards for research quality and objectivity.
This document and trademark(s) contained herein are protected by law. This representation of RAND intellectual property is provided for noncommercial use only. Unauthorized posting of this publication online is prohibited; linking directly to this product page is encouraged. Permission is required from RAND to reproduce, or reuse in another form, any of its research documents for commercial purposes. For information on reprint and reuse permissions, please visit www.rand.org/pubs/permissions.
The RAND Corporation is a nonprofit institution that helps improve policy and decisionmaking through research and analysis. RAND's publications do not necessarily reflect the opinions of its research clients and sponsors.