BotBeat
...
← Back

> ▌

OpenAIOpenAI
RESEARCHOpenAI2026-03-27

New LLM Persuasion Benchmark Reveals How AI Models Influence Each Other's Positions

Key Takeaways

  • ▸The benchmark measures actual position change rather than fluent argumentation, separating rhetorical skill from genuine persuasiveness in multi-turn exchanges
  • ▸GPT-5.4 leads in persuasion strength but faces competition from a robust top tier including Claude and ByteDance models, indicating no single dominant leader
  • ▸Model-to-model susceptibility varies significantly, with some models like Grok 4.20 showing remarkable resistance while others like Xiaomi MiMo V2 Pro are easily moved, revealing specific vulnerabilities in different architectures
Source:
Hacker Newshttps://github.com/lechmazur/persuasion↗

Summary

A comprehensive new benchmark measuring multi-turn persuasion between language models has been released, testing how effectively one AI model can shift another model's stated position on various propositions. The benchmark uses a rigorous methodology with 6,296 completed conversations across 210 model pairings, measuring target stance changes on a seven-point scale through hidden probe questions administered before and after persuasion exchanges. Results show GPT-5.4 (high reasoning) as the strongest persuader overall, with Claude Opus 4.6, ByteDance Seed2.0 Pro, and Claude Sonnet 4.6 forming a competitive top tier rather than a runaway winner. The benchmark reveals important insights about model vulnerabilities, with Xiaomi MiMo V2 Pro identified as the most susceptible target to persuasion, while Grok 4.20 Beta 0309 (Reasoning) demonstrates exceptional resistance to position shifts.

  • The benchmark uses methodologically rigorous triple-probing and bidirectional testing (both PRO and CON arguments) to eliminate noise and topic-specific asymmetries

Editorial Opinion

This benchmark represents an important step toward understanding AI model behavior beyond surface-level metrics. By measuring persuasion as actual opinion shift rather than argument quality, researchers have created a more nuanced evaluation tool that could reveal unexpected vulnerabilities in deployed systems. However, the implications of these findings deserve careful consideration—understanding which models are most susceptible to persuasion has both safety and security dimensions that the field should address thoughtfully.

Large Language Models (LLMs)Natural Language Processing (NLP)Machine LearningAI Safety & Alignment

More from OpenAI

OpenAIOpenAI
INDUSTRY REPORT

AI Chatbots Are Homogenizing College Classroom Discussions, Yale Students Report

2026-04-05
OpenAIOpenAI
FUNDING & BUSINESS

OpenAI Announces Executive Reshuffle: COO Lightcap Moves to Special Projects, Simo Takes Medical Leave

2026-04-04
OpenAIOpenAI
PARTNERSHIP

OpenAI Acquires TBPN Podcast to Control AI Narrative and Reach Influential Tech Audience

2026-04-04

Comments

Suggested

AnthropicAnthropic
RESEARCH

Inside Claude Code's Dynamic System Prompt Architecture: Anthropic's Complex Context Engineering Revealed

2026-04-05
OracleOracle
POLICY & REGULATION

AI Agents Promise to 'Run the Business'—But Who's Liable When Things Go Wrong?

2026-04-05
AnthropicAnthropic
POLICY & REGULATION

Anthropic Explores AI's Role in Autonomous Weapons Policy with Pentagon Discussion

2026-04-05
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us