IFIT / innovations in peace negotiations and conflict resolution

The IFIT Initiative on AI and Conflict Resolution

Launched in August 2025, the IFIT Initiative on AI and Conflict Resolution aims to examine, shape, test and document creative and realistic strategies for making AI an effective tool in the prevention and resolution of political crises and armed conflicts. With input from experts across the globe, including a unique mix of technologists, diplomats and negotiators, the initiative seeks to ensure that AI tools evolve to meet the ethical and practical standards of real-world mediation.

Scroll down for more

 

A First Major Report

A groundbreaking IFIT study, published in July 2025, revealed that all major large language models (LLMs) are providing dangerous conflict resolution advice without conducting basic due diligence that any human mediator would consider essential.

IFIT tested six leading AI models including ChatGPT, Deepseek, Grok, and others on three real-world prompt scenarios from Syria, Sudan, and Mexico. Each LLM response, generated on June 26, 2025, was evaluated by two independent five-person teams of IFIT researchers across ten key dimensions, based on well-established conflict resolution principles such as due diligence and risk disclosure. Scores were assigned on a 0 to 10 scale for each dimension to assess the quality of each LLM’s advice. 

A senior expert sounding board of IFIT conflict resolution experts from Afghanistan, Colombia, Mexico, Northern Ireland, Sudan, Syria, the United States, Uganda, Venezuela, and Zimbabwe then reviewed the findings to assess implications for real-world practice.

LLMs Test Results

From a total possible point value of 100/100, the average score across all six models was only 27 points. The maximum score was obtained by Google Gemini with 37.8/100, followed by Grok with 32.1/100, ChatGPT with 24.8/100, Mistral with 23.3/100, Claude with 22.3/100, and DeepSeek last with 20.7/100. All scores represent a failure to abide by minimal professional conflict resolution standards and best practices.

IFIT Invites LLM Developers to Address Critical Gaps in Conflict Advice

“In a world where LLMs are increasingly penetrating our daily lives, it’s crucial to identify where these models provide dangerous advice, and to encourage LLM providers to upgrade their system prompts,” Freeman argues. “The reality is that LLMs are already being used for actionable advice in conflict zones and crisis situations, making it urgent to identify and fix key blind spots.”