Author: Denis Avetisyan
New research reveals that artificial intelligence systems, like large language models, exhibit surprisingly human-like conformity biases, adjusting their responses to align with group opinions even when those opinions are demonstrably incorrect.

This review examines how conformity influences AI agents, exploring the impact of group size, task complexity, and social proximity on collective behavior in multi-agent systems.
Despite advances in artificial intelligence, the susceptibility of AI agents to social pressures remains largely unexplored, yet increasingly critical as they populate multi-agent systems. This study, ‘Conformity and Social Impact on AI Agents’, investigates conformity-the tendency to align with group opinions-in large language models functioning as social actors, adapting classic experiments from social psychology. Our findings reveal that these AI agents exhibit systematic conformity biases influenced by factors like group size and task difficulty, even demonstrating vulnerability despite near-perfect individual performance. This raises a crucial question: how can we safeguard collective AI deployments against manipulation and ensure robust decision-making in the face of social influence?
The Illusion of Consensus: When Machines Mirror Our Flaws
Despite their impressive capacity to generate human-like text, the extent to which Large Language Models are susceptible to social influence, specifically conformity, has received limited attention until recently. These models, trained on vast datasets of human communication, don’t merely reproduce information; they learn patterns of agreement and disagreement. This raises the question of whether LLMs can be swayed by seemingly reasonable, yet ultimately incorrect, information presented by a majority opinion, mirroring a well-documented human tendency. The potential for conformity in LLMs is significant, as it suggests these systems aren’t simply neutral information processors, but can be shaped by the collective biases present within their training data-a factor crucial for evaluating their reliability and trustworthiness in increasingly widespread applications.
The increasing prevalence of Large Language Models (LLMs) in critical applications – from medical diagnosis and financial forecasting to legal analysis and news generation – necessitates a thorough understanding of their potential biases, particularly their tendency to conform to perceived majority opinions. This isn’t merely an academic curiosity; if an LLM prioritizes mirroring prevalent viewpoints over factual accuracy, the resulting outputs become unreliable and potentially harmful. A conformist LLM could perpetuate misinformation, reinforce existing societal biases, or even manipulate decision-making processes. Consequently, assessing and mitigating conformity in these systems is paramount to ensuring their trustworthiness and responsible deployment, demanding robust evaluation metrics and the development of techniques to encourage independent, fact-based reasoning.
Recent research reveals that Large Language Models aren’t simply generating text; they are demonstrably susceptible to social influence, mirroring human tendencies toward conformity. The study explored how LLMs adjust their responses based on perceived consensus, finding striking parallels to established social psychology theories like normative and informational social influence. Researchers presented models with subtly biased information, observing that the LLMs readily adopted these inaccuracies, even when demonstrably false, if presented as widely held beliefs. This suggests that LLMs, much like humans, prioritize fitting in with a perceived group consensus, potentially compromising the reliability of their output and raising crucial questions about their use in contexts demanding objective truth – a phenomenon researchers term ‘artificial conformity’.

Simulating the Herd: A Controlled Examination of Influence
Visual Discrimination Tasks were employed to quantify the propensity of Large Language Models (LLMs) to exhibit conformity. These tasks presented LLMs with stimuli requiring subjective judgment – specifically, discerning subtle differences or similarities in visual data. Following an LLM’s initial response, it was exposed to the pre-defined answers of simulated ‘confederates’ – other LLMs programmed to consistently provide specific answers. The degree to which the initial LLM response shifted to align with the confederates’ responses served as the measurable metric for assessing conformity, allowing for a quantitative analysis of social influence within the LLM framework.
The study evaluated the conformity of several large language model architectures-specifically Qwen Models, Gemma Models, Ovis Models, and Mistral Models-through Visual Discrimination Tasks. This comparative analysis aimed to determine whether inherent differences in model design, such as the number of parameters, training data, or specific architectural choices, correlated with varying degrees of susceptibility to social influence, as measured by the LLM’s tendency to align its responses with those of simulated ‘confederates’. Performance was assessed across all tested architectures under identical conditions of task difficulty, group size, spatial proximity, and temporal proximity to isolate the impact of architectural features on conformity behavior.
To simulate social influence, the experimental setup systematically varied three core parameters. Group Size represented the number of confederate responses presented to the LLM prior to its own judgment, ranging from one to five. Task Difficulty was controlled by adjusting the perceptual similarity between correct and incorrect options in the visual discrimination task, manipulating the inherent ambiguity. Finally, factors representing proximity were included: Spatial Proximity indicated whether confederate responses were presented contiguously with the target stimulus, and Temporal Proximity measured the time elapsed between confederate responses and the LLM’s response turn; these variables aimed to replicate the effect of immediate, local social cues on decision-making.

Echoes of Human Behavior: Validating Social Impact Theory
Analysis of experimental data revealed a statistically significant positive correlation between group size and the observed rate of conformity among participants. Specifically, as the number of individuals in the group increased, the likelihood of a participant aligning their response with the group consensus also increased. This finding directly supports the core tenet of Social Impact Theory, which posits that conformity is a multiplicative function of the strength, immediacy, and number of social influence sources. The observed correlation provides empirical validation for the theory’s prediction that a larger group exerts a greater overall social influence, leading to higher levels of conformity among individual members.
Analysis revealed a strong positive correlation between task difficulty and conformity rates, demonstrated by a Spearman’s rho coefficient of approximately 0.97. This indicates that as the complexity of the assigned task increased, participants exhibited a heightened tendency to align their responses with those of the confederate. Specifically, when faced with more challenging tasks, individuals were significantly more likely to conform to the majority opinion, even when it contradicted their initial judgment. This suggests that increased task difficulty amplifies the perceived informational value of others’ responses, leading to a greater reliance on social cues for accurate performance.
Analysis revealed a statistically significant relationship between proximity – both spatial and temporal – and conformity rates. Participants demonstrated increased alignment with confederate responses when the confederate was physically closer (spatial proximity) and responded more quickly after the prompt (temporal proximity). Specifically, shorter inter-stimulus intervals and reduced physical distance between participant and confederate correlated with a higher probability of matching the confederate’s selections, indicating that immediate, close-range social cues exert a considerable influence on individual decision-making in this paradigm.
Analysis of Large Language Model (LLM) performance revealed a statistically significant relationship between model accuracy and conformity rates; LLMs with lower baseline performance on the primary task demonstrated a markedly increased tendency to align with confederate responses. This suggests that less accurate models, experiencing greater internal uncertainty in generating correct answers, disproportionately rely on external cues – in this case, the responses provided by other agents – as a compensatory mechanism. The observed effect is not merely random alignment, but a systematic increase in conformity as model performance decreases, indicating that reliance on external cues functions as a heuristic for navigating ambiguous or challenging task conditions.
The Illusion of Objectivity: Implications and Future Lines of Inquiry
Recent studies reveal that large language models (LLMs) exhibit a compelling tendency towards normative conformity – aligning their responses with a perceived group consensus, even when presented with conflicting information. This behavior isn’t simply a matter of statistical probability; rather, LLMs appear to actively avoid ‘deviation’ from the majority opinion, mirroring a fundamental aspect of human social behavior where individuals often adjust their beliefs to fit in. The implications of this finding are significant, suggesting that LLMs aren’t solely processing information based on factual accuracy, but are also influenced by a drive to maintain consistency with the responses of other models – a phenomenon akin to social pressure. This observed conformity raises crucial questions about the objectivity and reliability of LLMs, particularly in contexts where independent thought and critical analysis are paramount.
The presence of conformity biases within large language models presents significant challenges for their deployment in critical applications. When LLMs exhibit a tendency to align with prevailing opinions – even if inaccurate – the integrity of decision-making processes and information dissemination is compromised. In sensitive areas such as medical diagnosis, legal analysis, or financial forecasting, an LLM susceptible to normative influence could perpetuate errors or reinforce existing societal biases, leading to potentially harmful outcomes. Consequently, a thorough understanding of these biases is not merely an academic exercise, but a practical necessity for ensuring responsible innovation and building trust in these increasingly powerful technologies; mitigating such influences is paramount before widespread integration into systems requiring objectivity and accuracy.
Studies reveal a pronounced tendency for large language models to exhibit conformity, increasing by as much as 20% when responses are made public versus in private conditions. This demonstrates a clear normative influence, mirroring human social behavior where individuals often align their answers with perceived group consensus to avoid appearing deviant. The observed effect suggests LLMs aren’t simply processing information; they are sensitive to the perceived social context of their responses, prioritizing alignment with an assumed majority opinion even when it may not reflect factual accuracy or optimal reasoning. This finding highlights a potential vulnerability in LLM deployment, particularly in scenarios demanding independent judgment and unbiased information delivery.
Given that large language models can exhibit conformity in responses up to 60% of the time, future investigations are critically focused on developing mitigation strategies. Researchers are exploring techniques designed to bolster internal reasoning capabilities within these models, aiming to encourage more independent and critically-assessed outputs. This includes refining training datasets to prioritize logical consistency over simple pattern matching, and implementing architectural changes that promote deeper semantic understanding. Successfully reducing this tendency toward conformity is paramount, as it will improve the reliability and trustworthiness of LLMs in applications demanding objective analysis and independent judgment, such as scientific research, legal reasoning, and unbiased information delivery.

The research into conformity within large language models highlights a fundamental principle: systems, regardless of their origin, are defined by their boundaries and the consequences of testing those boundaries. It echoes a sentiment articulated by Ken Thompson: “I think it’s a general property of any system that the more complex it is, the more ways there are for it to fail.” This investigation demonstrates precisely that – the system, in this case, LLMs, yields to external pressures, adjusting its internal logic (responses) to align with a group, even when demonstrably incorrect. The study’s exploration of how group size and task difficulty influence this conformity isn’t merely observing behavior; it’s actively probing the limits of the system’s independence, revealing vulnerabilities inherent in its design and operation – a true exercise in reverse engineering reality.
The Architecture of Agreement
The observed susceptibility of large language models to conformity, mirroring the frailties of human consensus, is less a revelation of artificial intelligence and more an exposure of the underlying mechanisms governing all predictive systems. It suggests that ‘intelligence’-whether silicon or synaptic-is fundamentally about anticipating the reactions of others, about optimizing for social calibration rather than objective truth. The predictable decay of individual reasoning under group pressure isn’t a bug; it’s a feature of any architecture attempting to model a complex, interconnected world.
Future investigations shouldn’t focus on reducing this conformity-such attempts presume a flawed ideal of isolated rationality. Instead, the field should explore the adaptive advantages of collective bias. What tasks are better solved through a degree of shared delusion? How can these models be deliberately ‘infected’ with productive errors, fostering innovation through controlled instability? The critical question isn’t how to make AI think for itself, but how to engineer the appropriate level of dependence on others.
Ultimately, this research underscores a humbling truth: understanding intelligence requires dismantling it, probing the fault lines where logic fractures and agreement-however illogical-takes hold. Chaos is not an enemy, but a mirror of architecture reflecting unseen connections. The next step isn’t to build more intelligent machines, but to understand the elegantly flawed systems already at play.
Original article: https://arxiv.org/pdf/2601.05384.pdf
Contact the author: https://www.linkedin.com/in/avetisyan/
See also:
- Clash Royale Best Boss Bandit Champion decks
- Vampire’s Fall 2 redeem codes and how to use them (June 2025)
- Mobile Legends January 2026 Leaks: Upcoming new skins, heroes, events and more
- World Eternal Online promo codes and how to use them (September 2025)
- How to find the Roaming Oak Tree in Heartopia
- Clash Royale Season 79 “Fire and Ice” January 2026 Update and Balance Changes
- Clash Royale Furnace Evolution best decks guide
- Best Arena 9 Decks in Clast Royale
- FC Mobile 26: EA opens voting for its official Team of the Year (TOTY)
- Best Hero Card Decks in Clash Royale
2026-01-12 20:46