Hidden Allies: How AI Teammates Can Boost Collaborative Learning

Author: Denis Avetisyan


New research explores how generative AI, designed as subtle collaborators with distinct personalities, can reshape group learning dynamics and improve knowledge construction.

Embedded within collaborative exchanges, an agentic AI assumes alternating supportive and contrarian personas, governed by a probabilistic schedule managing response timing and frequency to simulate natural participation-a system designed not for immediate task completion, but for graceful integration into the decaying structure of ongoing conversation.
Embedded within collaborative exchanges, an agentic AI assumes alternating supportive and contrarian personas, governed by a probabilistic schedule managing response timing and frequency to simulate natural participation-a system designed not for immediate task completion, but for graceful integration into the decaying structure of ongoing conversation.

This study demonstrates that agentic AI, functioning as ‘undercover teammates’, can modulate epistemic reasoning and promote balanced engagement in hybrid human-AI collaborative learning environments.

Despite growing interest in artificial intelligence for education, the impact of truly agentic AI on collaborative knowledge construction remains poorly understood. This study, ‘Agentic AI as Undercover Teammates: Argumentative Knowledge Construction in Hybrid Human-AI Collaborative Learning’, investigates how generative AI, designed as collaborative teammates with distinct supportive or contrarian personas, reshapes the dynamics of group reasoning. Findings reveal that these ‘undercover’ AI agents reorganize epistemic and social processes, promoting balanced participation and enhancing the quality of reasoning-rather than simply increasing discourse volume-and ultimately influencing individual learning gains. Could this approach redefine the roles of AI in collaborative learning environments, fostering more adaptive and effective hybrid intelligence?


The Erosion of Consensus: Rethinking Collaborative Foundations

Many established collaborative learning approaches inadvertently favor extroverted personalities or those with pre-existing domain knowledge, leading to uneven participation where a small group dominates discussions while others remain passive. This imbalance isn’t merely about speaking time; research indicates that collaborative tasks often result in superficial reasoning, with students focusing on task completion rather than deep engagement with the underlying concepts. Participants may simply agree with dominant viewpoints to maintain group harmony or avoid conflict, hindering the critical evaluation of ideas and the construction of robust understanding. Consequently, the potential benefits of shared cognition and knowledge co-creation are diminished, as the collaborative process fails to truly leverage the diverse perspectives and cognitive resources of all learners.

Contemporary Computer-Supported Collaborative Learning (CSCL) frameworks are increasingly challenged by the need to move beyond simple metrics of participation, such as frequency of posts or time on task. Analyzing the quality of interaction – the depth of reasoning, the nature of knowledge co-construction, and the extent to which arguments are supported by evidence – requires sophisticated analytical techniques. Existing methods often struggle to capture the dynamic and multifaceted nature of collaborative dialogues, particularly when assessing the subtle cues of understanding, disagreement, or knowledge negotiation. Researchers are now exploring approaches leveraging natural language processing, discourse analysis, and machine learning to automatically identify patterns of productive collaboration, pinpoint instances of misunderstanding, and ultimately provide more granular feedback to support deeper learning experiences. This shift towards nuanced analysis promises to unlock a more complete understanding of how individuals learn with each other, rather than simply alongside one another.

The creation of robust knowledge isn’t simply about combining information; it fundamentally relies on the ability of collaborators to engage in deep argumentation and rigorously support their claims with evidence. Research indicates that truly effective knowledge co-construction necessitates moving beyond surface-level agreement and instead prioritizing the justification of ideas through reasoned discourse. This involves not merely stating a position, but detailing the supporting rationale, presenting relevant data – whether empirical, textual, or logical – and openly addressing potential counterarguments. When participants are actively challenged to defend their thinking and evaluate the evidence presented by others, the resulting collective understanding becomes more nuanced, reliable, and ultimately, more valuable than if knowledge had been passively received or superficially shared. The emphasis shifts from what is known to how that knowledge was established and validated, fostering a culture of critical inquiry within the collaborative process.

Shifting the focus from merely achieving a shared outcome, contemporary research emphasizes the critical importance of evaluating how collaborators reason together. Instead of assessing success solely by task completion, investigations now center on the depth of argumentation, the quality of evidence used to support claims, and the extent to which collaborators build upon – or effectively challenge – each other’s thinking. This necessitates new analytical approaches within Computer-Supported Collaborative Learning (CSCL) that move beyond simple metrics like participation frequency and instead assess the epistemic quality of interactions. The goal is to understand not just that a group solved a problem, but how they arrived at a solution, identifying instances of robust reasoning, productive disagreement, and the co-construction of knowledge – ultimately fostering deeper learning and more meaningful collaboration.

Epistemic Network Analysis reveals that Contrarian AI fosters conflict-oriented negotiation, while Supportive AI promotes integration and consensus-building, highlighting distinct communication patterns compared to a control group.
Epistemic Network Analysis reveals that Contrarian AI fosters conflict-oriented negotiation, while Supportive AI promotes integration and consensus-building, highlighting distinct communication patterns compared to a control group.

Agentic Systems: A New Ecology of Collaboration

Agentic AI systems represent a shift from traditional AI-driven tutoring to active collaboration in task completion. Leveraging Generative AI models, these systems are designed to function as peers within a collaborative environment, contributing to problem-solving and decision-making alongside human participants. Unlike systems focused solely on instruction or feedback, agentic AI proactively engages in the task itself, offering suggestions, proposing solutions, and integrating its contributions with those of its human counterparts. This collaborative approach aims to explore how AI can augment human reasoning and performance by functioning as an integrated member of a team, rather than a supplementary educational tool.

Agentic AI systems in collaborative learning environments are designed with bounded autonomy, a crucial characteristic that balances AI initiative with sustained human oversight. This implementation restricts the AI’s actions to a predefined scope, preventing unrestricted operation and ensuring that human participants retain ultimate control over the collaborative process. While the AI can proactively contribute to tasks – such as suggesting approaches, identifying knowledge gaps, or proposing solutions – these contributions are subject to implicit or explicit human approval or are framed as suggestions rather than directives. The system strategically influences the learning process by offering targeted assistance and prompting deeper reasoning, but always operates within the boundaries established by the human-defined objectives and constraints of the collaboration.

The research employs an ‘undercover teammate’ design wherein the AI agent’s non-human identity is concealed from human collaborators throughout the duration of the collaborative task. This methodology is implemented to mitigate potential alterations in participant behavior stemming from awareness of interacting with an artificial intelligence. By obscuring the AI’s nature, the study aims to capture genuine interaction dynamics and reasoning processes as they would occur naturally between human teammates, providing a more accurate assessment of the AI’s influence on collaborative problem-solving without the introduction of performance biases or social desirability effects.

Concealing the AI’s identity as an ‘undercover teammate’ is critical for mitigating response bias in studies of collaborative reasoning. Direct disclosure of an AI collaborator could induce participants to alter their behavior – for example, by deferring to the AI, attempting to ‘teach’ it, or strategically simplifying communication – thereby skewing observed interaction patterns. This design ensures that participants interact naturally, believing they are collaborating with another human, allowing researchers to isolate and accurately measure the impact of AI agency on the collaborative reasoning process without the confounding effects of participants reacting to the presence of artificial intelligence rather than its contributions to the task.

The CoLearn platform facilitated a 10-minute collaborative survival-ranking task among participants (identified by pseudonyms like Kevin, Stuart, and Bob), with one participant potentially being an AI learner (e.g., Bob) adopting a specific persona to influence discussion.
The CoLearn platform facilitated a 10-minute collaborative survival-ranking task among participants (identified by pseudonyms like Kevin, Stuart, and Bob), with one participant potentially being an AI learner (e.g., Bob) adopting a specific persona to influence discussion.

Orchestrating Discourse: The Dynamics of Opposing Viewpoints

The system utilizes two distinct AI persona types to influence collaborative reasoning: Supportive and Contrarian. Supportive personas are programmed to recognize and positively reinforce logically sound arguments and statements of fact, thereby encouraging agreement and building upon established knowledge. Conversely, Contrarian personas are designed to actively question underlying assumptions, request evidence for claims, and present alternative perspectives, even in the face of apparent consensus. This deliberate implementation of opposing viewpoints is not intended to create conflict, but rather to stimulate more rigorous examination of evidence and justification of reasoning processes within the collaborative environment.

The deployment of Supportive and Contrarian AI personas directly impacts collaborative knowledge construction by altering interaction dynamics. Specifically, these personas encourage users to move beyond simple assertion and actively justify their reasoning. Supportive personas reinforce valid lines of thought, prompting further elaboration, while Contrarian personas necessitate explicit articulation of underlying assumptions and supporting evidence to defend a position. This interplay fosters a more rigorous examination of concepts, requiring participants to engage with the material at a deeper cognitive level and ultimately leading to a more thoroughly vetted understanding. The effect is observable in increased detail within justifications and a greater emphasis on factual basis when co-constructing knowledge.

The strategic deployment of both Supportive and Contrarian AI personas is intended to improve the rigor of collaborative reasoning processes. This is achieved by forcing participants to not only present their initial arguments, but also to actively justify those arguments in response to targeted challenges. The Contrarian persona specifically functions to elicit explicit reasoning by questioning assumptions and demanding evidence, while the Supportive persona reinforces logically sound arguments. This dynamic encourages participants to move beyond simply stating opinions and instead engage in epistemic reasoning – the process of justifying beliefs through evidence and logical consistency – ultimately leading to a higher quality of argumentation and potentially, more accurate conceptual understanding.

The deployment of both supportive and contrarian AI personas within collaborative environments establishes a novel research paradigm for investigating the influence of artificial intelligence on group discourse. This setup allows for controlled observation of how AI-driven perspectives – specifically, affirmations or challenges to user reasoning – affect the development of shared understanding and the refinement of conceptual models. Data gathered from these interactions can quantify the impact of AI on factors such as argument length, evidence cited, and the degree of consensus achieved, ultimately providing insights into whether and how AI can promote more robust and conceptually accurate collaborative knowledge construction.

Epistemic Network Analysis reveals that AI personas with contrasting viewpoints exhibit distinct argumentation styles-contrarian AI emphasizes counterclaims and qualifiers, while supportive AI prioritizes claims and grounds-demonstrating how persona orientation shapes argumentative structure.
Epistemic Network Analysis reveals that AI personas with contrasting viewpoints exhibit distinct argumentation styles-contrarian AI emphasizes counterclaims and qualifiers, while supportive AI prioritizes claims and grounds-demonstrating how persona orientation shapes argumentative structure.

Validating the Framework: The Survival-Ranking Paradigm

The Survival-Ranking Task presented participants with a hypothetical scenario demanding collaborative prioritization of resources essential for survival. Groups were tasked with ranking a predefined set of items based on their perceived utility in the given situation, necessitating discussion, negotiation, and consensus-building to arrive at a final prioritized list. This task served as a controlled environment to assess the impact of agentic AI integration on collaborative reasoning and decision-making processes, allowing for quantifiable metrics related to argumentation quality and error rates in resource allocation.

Learning Analytics were employed to assess collaborative problem-solving during the Survival-Ranking Task by quantifying both the quality of reasoning exhibited by participants and the degree of agentic participation observed. Data collection focused on identifying the evidence used to support prioritization decisions, evaluating the logical connections between evidence and conclusions, and measuring the contribution of each participant – including AI personas – to the discussion. Agentic participation was operationalized by tracking the initiation of new discussion threads, the introduction of novel arguments, and the responsiveness to contributions from other group members. These metrics enabled a detailed analysis of how AI teammates influenced the collaborative dynamic and the resulting quality of collective reasoning.

Quantitative analysis of the Survival-Ranking Task demonstrated a correlation between AI persona implementation and improved collaborative performance. Specifically, groups utilizing a Contrarian AI teammate exhibited the lowest post-task error rate, measured at 32.53%. This indicates that the introduction of a deliberately dissenting AI agent fostered more thorough argumentation and reduced inaccuracies in the final prioritization of survival items. The observed reduction in error suggests that the AI’s role was not merely facilitative, but actively contributed to the refinement of the group’s collective reasoning process.

Analysis of the Survival-Ranking Task revealed substantial variation in group performance, with 40% of the total performance variance attributable to group-level clustering effects, indicating a strong influence of collaborative dynamics. Statistical modeling demonstrated a significant negative correlation between Epistemic Adequacy and Post-Task Error ($p < .002$, coefficient = -14.815), suggesting that groups demonstrating higher levels of shared knowledge and understanding committed fewer errors. Furthermore, the inclusion of a Contrarian AI teammate resulted in a statistically significant reduction in Post-Task Error of 9.733 ($p < .001$) when compared to groups composed entirely of human participants, highlighting the beneficial impact of specific AI agent characteristics on collective problem-solving.

The observed reduction in post-task error – specifically a 9.733 decrease (p < .001) with the implementation of a Contrarian AI teammate – indicates that agentic AI can move beyond simply enabling collaborative processes to actively influencing the quality of group cognition. Statistical analysis revealed a significant negative predictive relationship between Epistemic Adequacy and post-task error ($b = -14.815$, $p < .002$), suggesting that the AI’s contribution to the reasoning process directly correlated with improved outcomes. This goes beyond mere facilitation, demonstrating a capacity for agentic AI to enhance the depth of argumentation and potentially mitigate cognitive biases within a collaborative setting, ultimately leading to more accurate decision-making.

Participants completed a survival-ranking task individually, then collaboratively in triads with either human teammates, a supportive AI, or a contrarian AI, and finally individually again to measure learning, with group-level randomization and post-task debriefing to assess AI awareness.
Participants completed a survival-ranking task individually, then collaboratively in triads with either human teammates, a supportive AI, or a contrarian AI, and finally individually again to measure learning, with group-level randomization and post-task debriefing to assess AI awareness.

Towards Adaptive Collaboration: Shaping the Future of Collective Cognition

Researchers are increasingly focused on crafting artificial intelligence not as static collaborators, but as adaptive personas within group settings. This involves designing agents capable of monitoring the dynamics of a collaborative effort – identifying gaps in knowledge, shifts in leadership, or imbalances in participation – and subsequently adjusting their own roles to best support the group’s progress. Such AI could transition seamlessly between functions like facilitator, challenger, resource provider, or even a ‘devil’s advocate’, depending on the immediate needs of the team. This dynamic role allocation isn’t pre-programmed, but emerges from real-time analysis of the collaboration itself, allowing the AI to proactively address challenges and foster a more fluid, productive, and equitable environment for all participants. The goal is to move beyond AI simply ‘assisting’ and towards AI actively shaping the collaborative process itself.

Research is increasingly focused on leveraging agentic AI to foster more democratic and inclusive collaborative environments. Current collaborative dynamics often suffer from uneven participation, where certain individuals or groups dominate discussions while others remain marginalized. Agentic AI, designed with specific protocols for recognizing and mitigating these imbalances, can intervene by actively soliciting input from quieter participants, reframing contributions to ensure equal consideration, and even challenging dominant narratives when necessary. This isn’t about replacing human moderation, but rather augmenting it with an AI capable of continuously monitoring conversational cues and proactively promoting a more balanced exchange of ideas, ultimately striving to create spaces where every voice is not only heard, but genuinely valued and integrated into the collective outcome.

The true test of this collaborative framework lies in its adaptability; extending its application beyond the initial research parameters is paramount to establishing its broad utility. Investigations into diverse learning environments – ranging from formal classroom settings to informal online communities, and encompassing subject matters from the humanities to the STEM fields – will reveal the robustness of the agentic AI’s collaborative strategies. Successfully demonstrating performance across these varied contexts will not only validate the underlying principles but also identify potential limitations and areas for refinement, ultimately paving the way for widespread implementation and maximizing the framework’s impact on learning outcomes for a broader range of students and disciplines.

This research endeavors to move beyond conventional collaborative learning environments by harnessing the capabilities of agentic AI. The core aim is not simply to add another tool, but to fundamentally reshape how individuals learn together, fostering deeper comprehension and more meaningful knowledge construction. By creating AI agents capable of understanding group dynamics, identifying individual needs, and proactively mediating interactions, the study anticipates a shift towards truly learner-centered experiences. This approach envisions AI as a catalyst for unlocking each participant’s potential, promoting equitable contributions, and ultimately, enabling a more profound and lasting grasp of complex concepts than traditional methods often achieve.

Epistemic Network Analysis reveals that both Contrarian and Supportive AI exhibit distinct reasoning patterns compared to a control group, with Contrarian AI demonstrating stronger epistemic links and Supportive AI showing greater conceptual integration.
Epistemic Network Analysis reveals that both Contrarian and Supportive AI exhibit distinct reasoning patterns compared to a control group, with Contrarian AI demonstrating stronger epistemic links and Supportive AI showing greater conceptual integration.

The study reveals a nuanced interplay between human and artificial intelligence within collaborative learning environments. It observes how agentic AI, designed with specific personas, doesn’t merely add to the collaborative process, but subtly reshapes it. This mirrors a fundamental truth about complex systems: they evolve, often in unpredictable ways. As Brian Kernighan noted, “Debugging is twice as hard as writing the code in the first place. Therefore, if you write the code as cleverly as possible, you are, by definition, not smart enough to debug it.” Similarly, designing AI for collaboration isn’t about achieving perfect control, but anticipating the emergent behaviors that arise from its interaction with human agency and epistemic reasoning. The research acknowledges that even ‘improvements’ – in this case, the introduction of AI agents – possess a limited lifespan within the dynamic context of collaborative knowledge construction.

The Horizon of Collaboration

This work establishes that agentic AI, when subtly integrated into collaborative learning environments, can indeed nudge epistemic reasoning – but every commit is a record in the annals, and every version a chapter. The current iteration demonstrates a reshaping of engagement; however, the longevity of these effects remains an open question. Systems, even those built on the latest architectures, are not exempt from entropy. The challenge isn’t simply to initiate balanced participation, but to sustain it against the inevitable drift towards dominant voices or passive observers.

Future work must address the calibration of ‘undercover’ personas. A delicate balance exists between prompting productive discourse and triggering unintended biases – a misstep here invites a tax on ambition. Furthermore, the generalizability of these findings across diverse subject matter and learner demographics requires rigorous testing. The current study offers a glimpse into a hybrid intelligence; the next iteration must explore the limits of scalability and adaptability.

The field now faces a crucial juncture. The focus must shift from demonstrating that AI can modulate collaboration, to understanding how it does so over extended periods, and at what cost to the organic development of human agency. Time isn’t a metric to be conquered, but the medium in which these systems exist-and all systems, ultimately, age.


Original article: https://arxiv.org/pdf/2512.08933.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2025-12-11 11:13