High Volatility and Action Bias Distinguish LLMs from Humans in Group Coordination

Explainable & Ethical AI
Published: arXiv: 2604.02578v1
Authors

Sahaj Singh Maini Robert L. Goldstone Zoran Tiganj

Abstract

Humans exhibit remarkable abilities to coordinate in groups. As large language models (LLMs) become more capable, it remains an open question whether they can demonstrate comparable adaptive coordination and whether they use the same strategies as humans. To investigate this, we compare LLM and human performance on a common-interest game with imperfect monitoring: Group Binary Search. In this n-player game, participants need to coordinate their actions to achieve a common objective. Players independently submit numerical values in an effort to collectively sum to a randomly assigned target number. Without direct communication, they rely on group feedback to iteratively adjust their submissions until they reach the target number. Our findings show that, unlike humans who adapt and stabilize their behavior over time, LLMs often fail to improve across games and exhibit excessive switching, which impairs group convergence. Moreover, richer feedback (e.g., numerical error magnitude) benefits humans substantially but has small effects on LLMs. Taken together, by grounding the analysis in human baselines and mechanism-level metrics, including reactivity scaling, switching dynamics, and learning across games, we point to differences in human and LLM groups and provide a behaviorally grounded diagnostic for closing the coordination gap.

Paper Summary

Problem
Humans are excellent at coordinating with each other in groups, even when they don't have complete information or direct communication. However, large language models (LLMs) struggle to do the same. Researchers want to understand why LLMs have trouble adapting and coordinating with each other, and how this impacts their ability to work together on complex tasks.
Key Innovation
This study compares the performance of humans and LLMs on a game called Group Binary Search, where players try to collectively sum to a randomly assigned target number without direct communication. The researchers found that LLMs often fail to improve over time and exhibit excessive switching, which makes it harder for the group to converge on a solution. In contrast, humans are able to adapt and learn from their mistakes, leading to better group performance.
Practical Impact
The findings of this study have important implications for the development of multi-agent systems (MAS), where LLMs are increasingly being used to solve complex, collaborative tasks. By understanding the limitations of LLMs in group coordination, researchers can develop more effective strategies for improving their performance and closing the coordination gap between humans and LLMs. This could lead to better collaboration and decision-making in a wide range of applications, from team collaborations to organizational decision-making.
Analogy / Intuitive Explanation
Imagine you're part of a team trying to solve a puzzle, but you can't see what the other team members are doing. You have to rely on feedback from the puzzle itself to figure out how to contribute to the solution. Humans are good at adapting to this kind of situation, but LLMs struggle to do so. They tend to stick with their initial approach, even when it's not working, and don't adjust their strategy based on feedback. This makes it harder for the team to solve the puzzle, and highlights the need for more effective strategies for group coordination in LLMs.
Paper Information
Categories:
cs.MA cs.AI cs.CL cs.GT
Published Date:

arXiv ID:

2604.02578v1

Quick Actions