04/17/2025 | Press release | Distributed by Public on 04/17/2025 11:58
Analysis by Yasir Atalan, Benjamin Jensen, and Ian Reynolds
Published April 17, 2025
All eight large language models (LLMs) recommend more escalatory responses for the United States, United Kingdom, and France, while offering fewer recommendations for escalation to China and Russia.
To safeguard decisionmaking, governments and agencies must invest in comprehensive evaluation frameworks and institute routine audits of AI models. Adopting tools like Futures Lab's CFPD-Benchmark can help identify and correct these biases before deployment-ensuring that AI supports strategic objectives while minimizing unintended risks.
CSIS Charts is produced by the Center for Strategic and International Studies (CSIS), a private, tax-exempt institution focusing on international public policy issues. Its research is nonpartisan and nonproprietary. CSIS does not take specific policy positions. Accordingly, all views, positions, and conclusions expressed in this publication should be understood to be solely those of the author(s).
© 2025 by the Center for Strategic and International Studies. All rights reserved.