February 7, 2026

We have spent the last several years asking whether AI can match human intelligence. A more interesting question has arrived: does it also match human stupidity?

Recent research suggests the answer is yes. AI agents don't just learn our knowledge. They faithfully reproduce our cognitive limitations — the same systematic errors that behavioral economists have catalogued in humans for half a century.

The Bias Isn't a Bug

Pilli and Nallur's work on predicting biased human decision-making tested whether GPT-4 and GPT-5 could reproduce the cognitive biases observed in 1,648 human participants [1]. The models didn't just occasionally exhibit biased behavior. Their predictions replicated the same bias patterns and load-bias interactions observed in humans — framing effects, status quo bias, and the way cognitive load amplifies both.

The counterintuitive finding: GPT-4 was better at replicating human biases than GPT-5. The more advanced model drifted toward rationality, while GPT-4 remained faithful to the irrational patterns that define actual human decision-making. This isn't a story about models being broken. It's a story about models being too good at learning from us.

Separately, Yee and Sharma's calibration framework tested four frontier models across 24,000 agent-scenario pairs on eight canonical behavioral biases from economics — loss aversion, herding, disposition effects [2]. They found what they call "systematic rationality bias" in baseline LLM behavior: the models deviate from human benchmarks in predictable, measurable ways. But crucially, these deviations map onto the same bias landscape that behavioral economists use to describe human decision-making. The models occupy a point on the same spectrum we do.

These aren't random errors. They're inherited architecture.

Cultural Inheritance at Scale

Individual biases are only part of the picture. Xuan et al.'s SocialVeil framework reveals what happens when biases collide [3].

SocialVeil tests language agents under three types of communication barriers grounded in human interaction research: semantic vagueness, sociocultural mismatch, and emotional interference. Across 720 scenarios and four frontier models, communication barriers reduced mutual understanding by over 45% on average and elevated confusion by nearly 50%.

The sociocultural mismatch finding deserves particular attention. When agents trained on different cultural contexts interact, their misaligned interpretive frameworks create friction that adaptation strategies barely dent. Repair instructions and interactive learning — the standard fixes — produced only modest improvements, falling far short of barrier-free performance.

The implication for multi-agent systems is stark. As organizations deploy agents built on different foundation models with different training data, the communication failures won't be technical. They'll be cultural — inherited from the divergent human cultures embedded in training data.

Strategic Reasoning Breaks the Same Way

If biases only affected communication, the problem would be containable. But Liu, Gu, and Song's AgenticPay benchmark demonstrates that the inheritance extends to strategic reasoning [4].

AgenticPay models markets where buyer and seller agents must negotiate through multi-round linguistic exchange — over 110 tasks spanning bilateral bargaining to complex many-to-many markets. Benchmarking state-of-the-art models revealed what the authors call "substantial gaps" in negotiation performance, particularly in long-horizon strategic reasoning.

PieArena showed agents negotiating at MBA level. New work reveals that MBA-level reasoning comes with MBA-level cognitive biases. The agents overweight recent information. They fail to maintain coherent long-term strategy under competitive pressure. They exhibit the same myopic tendencies that negotiation researchers have documented in human bargainers for decades.

These aren't capability limitations. They're behavioral patterns — inherited from the human negotiation data the models learned from.

The Laundering Problem

Here is the uncomfortable implication.

Organizations deploy AI agents precisely because they want decisions that are faster, more consistent, and less biased than human judgment. But if agents faithfully reproduce human biases, deployment doesn't remove bias from the decision pipeline. It launders bias — embedding it inside a system that is perceived as neutral and objective.

The bias is now harder to detect, not easier. A human decision-maker can be questioned, challenged, asked to explain their reasoning. An agent's biases are distributed across billions of parameters, invisible to the people relying on its outputs. More reasoning time doesn't fix systematic biases — it can actually amplify them by generating more sophisticated rationalizations.

Designing Around Inheritance

The path forward is not debiasing models. They learn from us, and we are biased. Attempting to surgically remove cognitive biases from systems trained on the full corpus of human expression is a losing game.

The path forward is the same one that good organizations have always used with human decision-makers: design systems that account for known cognitive limitations. Require structured decision frameworks. Build in adversarial review. Create checkpoints where inherited biases are most likely to compound.

The agents inherited our biases. That was inevitable. What matters now is whether we build systems that account for that inheritance — or whether we pretend the bias disappeared when we handed the decision to a machine.


References

[1] Pilli, S. & Nallur, V. (2026). "Predicting Biased Human Decision-Making with Large Language Models in Conversational Settings." arXiv:2601.11049. https://arxiv.org/abs/2601.11049

[2] Yee, B. & Sharma, K. (2026). "Calibrating Behavioral Parameters with Large Language Models." arXiv:2602.01022. https://arxiv.org/abs/2602.01022

[3] Xuan, K., Wang, P., Ye, C., Yu, H., August, T., & You, J. (2026). "SocialVeil: Probing Social Intelligence of Language Agents under Communication Barriers." arXiv:2602.05115. https://arxiv.org/abs/2602.05115

[4] Liu, X., Gu, S., & Song, D. (2026). "AgenticPay: A Multi-Agent LLM Negotiation System for Buyer-Seller Transactions." arXiv:2602.06008. https://arxiv.org/abs/2602.06008