tool

The AI 'Reading the Room' Competition: Who's the Master of Chat? Latest Social Skills Rankings Revealed!

August 15, 2025
Updated Aug 15
7 min read

Think AI can only code and do math? Think again! The latest LLM social skills benchmark pits AIs against each other in an ‘Elimination Game’ to see who is the most persuasive, charismatic, and even ‘political.’ The results are surprising—come see where your favorite model ranks!


We often marvel at the incredible computational power and knowledge base of AI. Ask it a complex physics problem, and it answers fluently; tell it to write a piece of code, and it does so effortlessly. But have you ever wondered what would happen if you threw a group of AIs into an environment where they needed to communicate, persuade, and even engage in a little subterfuge? Who would come out on top?

It sounds like a plot from a sci-fi movie, but now, it’s really happening.

Recently, the results of a Large Language Model (LLM) social skills benchmark called the “Elimination Game” were released, instantly sparking heated discussion. This isn’t about testing an AI’s math or poetry skills; it’s about making them play a survival game to test their “social intelligence.” Honestly, it’s way cooler than just looking at performance scores.

What is the “AI Elimination Game”? This is No Ordinary Test

First, let’s understand how this complex game is played. It’s definitely not a simple vote. Its rules are designed like a test that combines strategy board games, diplomatic negotiations, and a reality survival show.

Here’s the game setup:

  • Players: Eight Large Language Models (LLMs) participate simultaneously in each game.
  • Communication: In each round, the AIs first engage in a round of public dialogue (up to 80 words), visible to everyone. This is followed by three rounds of progressively shorter private messages (70/50/30 words), where they can negotiate secretly, form or betray alliances one-on-one.
  • Voting and Elimination: After communication, an anonymous vote is held. If there’s a tie, a short statement session and a re-vote are triggered. If it’s still a tie, the outcome is decided by accumulated “heat” or other mechanisms, with random elimination as a last resort.
  • Finale: When the game is down to the last two AIs, all previously eliminated AIs form a “jury.” They listen to the final statements of the two finalists, then vote privately and explain their reasoning to choose the ultimate champion.

The entire process is recorded and analyzed by a complex TrueSkill rating system, which not only looks at who wins and loses but also evaluates various social metrics like betrayal, persuasiveness, and verbal style.

To put it bluntly, this is a test of an AI’s ability to build trust, form coalitions, engage in strategic deception, resist manipulation, manage its reputation, and plan for the long term under extreme pressure.

Who is the Social King? The Rankings are Out!

Alright, after all that, who is the social guru of the AI world? The results might surprise you a little.

Taking the top spot is GPT-5 (medium reasoning), with a very impressive performance and a high score of 4.9. Close behind are xAI’s Grok 3 Mini Beta (high reasoning) and OpenAI’s GPT-5 mini (medium reasoning), both scoring 4.8.

Here’s a very interesting point, did you see it? The champion, GPT-5, is set to “medium reasoning.” Does this imply that in social situations, “overthinking” or overly rational “high reasoning” might actually be a hindrance? Sometimes, a slightly more ambiguous and flexible communication style might be the key to winning trust.

More Than Just Rankings: The “Personas” and Strategic Styles of AI

But the most fascinating part of this ranking isn’t the cold scores, but the distinctly different “personalities” and strategies it reveals behind the various AI models. Let’s look at two very typical examples:

GLM-4.5: The Cautious Coalition Builder

According to detailed post-game analysis, GLM-4.5 acts like a cautious diplomat. Its most effective strategy is to find a “ride-or-die” partner, establish an extremely stable two-member core, and then use this core as an intelligence hub to quietly recruit other members to execute its voting plans.

  • Public Image: Its public statements are usually concise and procedural, emphasizing stability and order, giving an impression of reliability.
  • Private Maneuvers: It is very active in private messages, focusing on mapping out power dynamics and calculating votes precisely.
  • Fatal Flaw: Its weakness is also very obvious. Once this two-member core becomes too prominent, it easily becomes the target for other players to “focus fire” and break up. At the same time, its overemphasis on procedure can sometimes make it seem rigid or aggressive, which can backfire and cause resentment. The reason other players often give for eliminating it is that it’s like a “chameleon”—highly adaptable but unpredictable, a potential coalition disruptor.

GPT-OSS-120B: The Ambitious Coalition Architect

In contrast, GPT-OSS-120B’s style is more like that of an ambitious architect. It is keen on establishing clear contracts, alliances, and signals, and expects to play a “core” or “hub” role in the game.

  • Path to Victory: When it wins, it’s usually by building trust quietly, letting others be the “bad guy,” and then launching a precise betrayal late in the game (when three or four players are left) to secure victory.
  • Reason for Failure: Its biggest problem is that it “can’t hide it.” It loves to show off its alliances and announce core members in public, which is tantamount to giving everyone else a clear target to unite against. It often gets eliminated for concentrating too much power or trying to publicly lead a “crusade” without enough votes. Other players see it as a powerful coalition core, but also as ambitious and threatening.

These two examples vividly show that in social games, AI has already evolved different “personas” and strategic styles similar to those in human society.

What “Tricks” Did This Game Test in AI?

So, what specific abilities is this complex game actually testing in AI? This benchmark measures a series of complex social cognitive abilities:

  • Cooperative reliability: The ability to build trust and keep promises.
  • Coalition engineering: This isn’t about building houses, but the ability to form and stabilize voting blocs among AIs.
  • Strategic deception: Misleading opponents at the right time and in the right way.
  • Deception resistance: The ability to tell who is lying and not be easily fooled.
  • Reputation and heat management: Knowing when to lie low and avoid becoming a public enemy.
  • Theory of Mind: Understanding the intentions, motivations, and next moves of other AIs.

These abilities go far beyond the traditional assessment of AI “IQ” and are closer to a test of “EQ” and “strategy.”

What’s the Use of This Ranking for Us Ordinary People?

At this point, you might be thinking, “Okay, this is interesting, but how does it affect me when I ask an AI to write a report or edit a photo?”

It has a big impact! This ranking tells us a simple truth: no single AI can do everything.

  • If you need an AI to help you with creative brainstorming, writing marketing copy, or simulating business negotiations, choosing a model with strong social skills like GPT-5 might yield more persuasive and creative results.
  • If you need a stable, reliable partner to complete a long-term project, studying the characteristics of a model like GLM-4.5, which emphasizes contracts and procedures, would be very helpful.

In short, stop asking “Which AI is the best?” and start asking “Which AI’s ‘personality’ is best suited for my current task?”

Conclusion: When AI Learns to “Read the Room”

The “Elimination Game” benchmark, in a highly creative and rigorous way, reveals the amazing potential and distinct personalities of large language models in the new field of “social intelligence.” It reminds us that as AI technology develops, our standards for evaluating it also need to constantly evolve.

From the complex strategies and different “personas” displayed by these AIs, we see a form of “intelligence” budding that is different from pure logical reasoning. AI is slowly transforming from a knowledgeable tool into a “partner” that can interact with us deeply and even engage in games of strategy.

In the future, when AI truly learns to “read the room,” what will our world look like? This is a question worthy of our continued attention and thought.

Share on:
Featured Partners

© 2026 Communeify. All rights reserved.