How will advances in artificial intelligence impact strategic stability? A growing number of studies and reports assessing the ways that advances in AI could influence global politics focus on the potential risks to strategic stability from integration of AI into the nuclear domain, particularly in large language models and frontier AI. These risks come from multiple potential sources, including miscalculation by machines, sidestepping of human firebreaks to escalation, AI-induced accidents, the speed of AI-enabled warfare, and other mechanisms. The relationship between AI integration and strategic stability may change over time as knowledge and experience with AI systems increases, thus decreasing the likelihood of automation bias, but fundamentally the relationship will depend on second-strike capabilities. While there is inherent uncertainty since we are still early in the age of AI, at this point it appears as though the higher the confidence nuclear-armed states have in their second-strike capabilities, the lower the probability that they integrate AI in dangerous ways that make escalation at machine speed more likely, and vice versa.
The article initiates a critical discussion on the impact of artificial intelligence (AI) on strategic stability, particularly concerning its integration into the nuclear domain. It highlights a dichotomy in current assessments: some studies predict increased risks from AI due to factors like machine miscalculation, human override bypass, AI-induced accidents, and the accelerated pace of AI-enabled warfare. Conversely, it's argued that AI integration might foster greater caution in decision-makers, potentially reducing nuclear escalation risks. The piece posits that the relationship between AI and strategic stability is non-linear and evolves with knowledge and experience, fundamentally hinging on the confidence nuclear-armed states have in their second-strike capabilities. It also stresses the psychological and organizational behavioral aspects of human interaction with AI and suggests that confidence-building measures (CBMs) could be crucial for enhancing international security, citing examples like a US-China agreement on human control over nuclear weapons and the potential for an Autonomous Incidents Agreement.
This section delves into the foundational concepts, tracing the evolution of artificial intelligence from its modern origins post-World War II to its current advancements driven by machine learning, neural networks, and deep learning. The author clarifies that "artificial intelligence" refers to tasks machines perform that traditionally required human intelligence, focusing on specific "AI use cases" rather than broader Artificial General Intelligence (AGI), acknowledging the "AI effect" where AI is always defined as what computers cannot yet do. Concurrently, strategic stability is defined primarily as "first-strike stability," aiming to prevent nuclear war by ensuring no adversary feels compelled to launch a preventive strike. The core idea is that in a stable scenario, war would only occur if deliberate, not accidental.
The article acknowledges the difficulty in empirically assessing AI's impact on strategic stability due to its nascent stage of integration into the nuclear domain. It then outlines the two main schools of thought: those who believe AI will undermine stability and those who believe it will not. This section serves as an introduction to the detailed arguments presented in the subsequent subsections.
This part elaborates on several critical ways AI could heighten nuclear risk. Firstly, it discusses the dangers of AI in early warning systems, drawing parallels with historical incidents like the Soviet Oko system's false alarm, which nearly triggered nuclear war due to machine error. It highlights the risks of AI algorithms being trained on insufficient simulated data, vulnerability to hacking, and the potential for automation bias to override human judgment, thereby eliminating human firebreaks to escalation. Secondly, it examines how enhanced AI surveillance capabilities, combined with faster and more accurate conventional munitions, could threaten second-strike capabilities by enabling real-time tracking of mobile transporter erector launchers (TELs) and ballistic missile submarines (SSBNs), leading to pressures for preemptive strikes. Thirdly, the deployment of uncrewed platforms with nuclear weapons, like Russia's Poseidon, raises concerns about eliminating positive human control and increasing accident risks. Lastly, the article addresses how the increasing speed of conventional warfare, driven by AI, could pressure national command systems, potentially incentivizing less stable nuclear launch postures like 'launch on warning' or pre-delegation, exacerbating rapid decision-making in crises.
This section offers a counter-narrative, suggesting that the risks posed by AI to strategic stability might be overstated. It argues that criticisms often assume suboptimal behavior from states, ignoring the strong military incentives for stringent testing, evaluation, and maintaining human control, especially for nuclear systems. The U.S. Department of Defense's 'human in the loop' policy for nuclear weapons decisions is cited as evidence. It proposes that if developed effectively, AI could improve early warning by speeding up pattern recognition, thus 'buying back time' for human decision-makers and potentially reducing hasty, error-prone responses. Doubts are cast on AI's ability to genuinely undermine second-strike capabilities, citing the practical difficulties of simultaneous real-time tracking and targeting of mobile TELs and SSBNs, given factors like asset movement, communication delays, and undersea physics. Furthermore, it suggests that uncrewed nuclear platforms, by offering greater endurance, could enhance second-strike reliability for states with limited arsenals, provided human control is preserved. Finally, it notes that military planning already accounts for conventional defeat scenarios in nuclear contingencies, and that AI might even reduce human-induced errors in complex warfare.
This section addresses the ambiguity in predicting AI's impact on strategic stability due to conflicting arguments and inherent uncertainties in AI development. It emphasizes that human interaction with AI, influenced by psychology and organizational behavior, will be crucial. A three-stage conceptual model is introduced: 1) the "hype and overconfidence period," where perceived technological effectiveness exceeds reality, leading to automation bias and risky decision-making (e.g., over-relying on AI for early warning, making miscalculation more likely); 2) the "trust gap period," where actual technology improves but trust lags, potentially causing missed opportunities for useful AI integration that could mitigate risks (e.g., faster AI information processing to detect false warnings); and 3) the "calibrated period," where knowledge and experience align expectations with technological reality, leading to optimal adoption. The article notes that accidents in earlier stages can lead to "adoption backsliding," stressing the importance of training and education for AI users to understand capabilities and limitations, thereby fostering better calibration and reducing automation bias.
This section expands on other contextual variables that will shape AI's influence on strategic stability, including technological progress, confidence in second-strike capabilities, regime type, and assumptions about the probability of accidents.
The pace and nature of AI research advancements are identified as a primary driver. The article notes the significant disagreement among AI experts regarding development timelines and capabilities. It argues that faster AI development might mitigate safety concerns by making algorithms more reliable, while excessively slow development would limit deployment and risk. The greatest risk lies in premature deployment fueled by overenthusiasm, where algorithms still harbor significant safety and reliability issues. The speed of AI advances also influences the feasibility of scenarios like AI-based surveillance undermining second-strike capabilities, emphasizing that the slower these advances, the lower their impact on stability.
A state's pre-existing confidence in its second-strike capabilities (e.g., ability to retaliate after a first strike) is presented as a crucial determinant. Countries highly confident in their second-strike deterrent are less likely to adopt risky AI applications that offer speed or endurance at the cost of reliability in the nuclear domain. Conversely, states perceiving themselves as vulnerable (e.g., Russia, North Korea, Pakistan) might be more inclined to embrace automation to bolster their perceived deterrence, potentially increasing risk. This highlights how AI adoption choices are deeply intertwined with a nation's strategic vulnerability assessment.
The article posits that autocratic regimes might have a unique incentive to integrate AI into their military and nuclear systems. Driven by concerns about 'coup-proofing' and distrust of their own military personnel, autocracies could use AI-enabled robotic systems to centralize control of force, reducing reliance on potentially disloyal human soldiers. This desire for centralized decision-making, even over nuclear use, might lead autocratic leaders to be more risk-acceptant regarding dangerous AI applications within their nuclear infrastructure compared to democracies, potentially increasing instability due to their internal political dynamics.
This part considers the applicability of 'normal accident theory' to AI integration, questioning whether increased system complexity and coupling inherent in digital infrastructure, particularly with AI, will inevitably raise accident risks. It discusses various AI failure modes, such as inadequate training data, data poisoning, hacking, and the inability of algorithms to handle unforeseen scenarios. The 'timing' of AI adoption (overconfidence vs. trust gap) is seen as influencing the likelihood of accidents. The section suggests that if past nuclear war avoidance was largely due to luck, then AI integration is inherently risky. It concludes by stressing that extensive planning, education, and training of human operators on AI's strengths and weaknesses are essential to manage these accident risks effectively.
This final section focuses on practical strategies to mitigate the strategic stability risks posed by military AI. It advocates for the implementation of confidence-building measures (CBMs), drawing lessons from the Cold War era. CBMs, including information sharing, inspections, 'rules of the road,' and limits on military operations, are proposed as tools to foster cooperation between competitors like the United States and China, who share an interest in preventing inadvertent escalation. Specific recommendations include agreeing to maintain 'positive human control' over nuclear launch decisions, thus opposing fully autonomous 'dead hand' systems, and negotiating an 'Autonomous Incidents Agreement' (modeled after the Incidents at Sea Agreement) to facilitate information sharing about AI-enabled autonomous systems, especially in peacetime, to reduce miscalculation risks from accidents. The article acknowledges challenges, such as potential information disclosure and the varying willingness of less secure states to adopt such measures.