Who’s Driving? How Separate Robot ‘Personas’ Affect Trust and Blame

Author: Denis Avetisyan


New research explores how assigning distinct identities to different components of a single robot influences human perceptions of responsibility when things go wrong.

This study investigates the impact of split-embodiment in human-robot interaction on trust, failure attribution, and the compartmentalization of blame across robotic domains.

While humans readily attribute agency and responsibility to distinct individuals, replicating this nuanced social interaction with robots presents a unique challenge. This is explored in ‘”Meet My Sidekick!”: Effects of Separate Identities and Control of a Single Robot in HRI’, which investigates how presenting multiple, differentiated “personas” controlling separate components of a single robot impacts human perception of trust and accountability. The study demonstrates that users can effectively compartmentalize responsibility, attributing failures to specific robot “agents” based on which component-head or gripper-caused the error. Could leveraging this split-embodiment approach enable more intuitive and trustworthy human-robot collaboration, ultimately allowing a single robot to fulfill multiple social roles?


The Fragility of Trust in Collaborative Systems

Human collaboration with robotic teammates hinges on a fundamental element: trust. However, this trust proves remarkably fragile, susceptible to erosion even with minor robotic errors. Unlike interactions with human colleagues, where intentions and capabilities are generally understood, robotic actions can be perceived as opaque or unpredictable, amplifying the impact of mistakes. A single failure can trigger disproportionate skepticism, not simply regarding the immediate task, but concerning the robot’s overall reliability and competence. This is because humans naturally assess trustworthiness based on both demonstrated capability and perceived intent; when a robot errs, attributing the cause – whether a technical malfunction, a programming flaw, or a lack of situational awareness – becomes critical, and a negative attribution swiftly diminishes confidence in future performance. Consequently, designing robotic systems that not only avoid errors but also effectively communicate their reasoning and limitations is paramount to sustaining productive human-robot teamwork.

When robotic teammates falter, a significant challenge arises from the difficulty in establishing clear accountability – a hurdle traditional single-agent robots are ill-equipped to overcome. Unlike human collaborators who can offer explanations and accept responsibility, these robots often lack the capacity for nuanced communication regarding the why behind an error. This absence of transparent reasoning hinders the process of trust recovery, as humans are left to speculate about the cause of the failure and the robot’s ability to prevent future mistakes. The resulting ambiguity can foster suspicion and erode confidence in the robotic teammate, ultimately diminishing the effectiveness of the human-robot team. Consequently, designing robots capable of articulating their internal reasoning and accepting responsibility for actions-or lack thereof-is paramount for fostering sustained trust and optimizing collaborative performance.

The capacity to discern responsibility following a robotic error is central to fostering sustained human trust in collaborative robots. Research indicates that humans don’t simply register a failure; they actively seek to understand why it occurred, attributing blame to factors like the robot’s programming, sensor malfunction, or even external environmental disturbances. This attribution process isn’t purely logical; it’s heavily influenced by expectations regarding the robot’s capabilities and the perceived intent behind its actions. Consequently, designing robots capable of transparently communicating the cause of an error-and accepting appropriate accountability-is paramount. A robot that can articulate its limitations and explain the rationale behind a mistake is far more likely to retain a human partner’s confidence than one that remains opaque or deflects responsibility, ultimately improving team performance and fostering more effective human-robot collaboration.

The efficacy of human-robot teams is deeply intertwined with the psychological response to robotic failures; a single error doesn’t just represent a task completion issue, but a potential disruption of the collaborative dynamic. Research indicates that humans don’t simply assess whether a robot fails, but how the failure occurs and the robot’s subsequent behavior profoundly influences continued cooperation. A robot perceived as lacking awareness of its mistakes, or exhibiting no attempt at correction, can quickly erode human confidence and willingness to collaborate, even if the error is minor. Consequently, designing for robust human-robot interaction necessitates proactive consideration of these psychological impacts, prioritizing transparency, error recovery mechanisms, and demonstrable robot ‘awareness’ to foster continued trust and maintain optimal team performance, even in the face of inevitable setbacks.

Distributed Agency: The Architecture of Accountability

Split Embodiment represents a departure from traditional robotic architectures by distributing control across multiple independent entities, termed Persona. In this design, each Persona is responsible for governing a specific set of physical subsystems within the robot. This means, for example, one Persona might control locomotion, while another manages manipulation, and a third handles sensing and perception. This division of labor is implemented through dedicated hardware and software interfaces, allowing each Persona to operate with a degree of autonomy while contributing to the overall robotic task. The resulting system differs fundamentally from single-controller designs where all subsystems are directly managed by a singular processing unit.

Traditional multi-agent systems operate within computational environments; this work extends that paradigm by physically realizing a multi-agent system through robotics. Instead of software agents coordinating within a single process, distinct computational agents – termed ‘Personas’ – directly govern separate physical subsystems of a robot. This creates a distributed robotic agent where processing and actuation are not centralized, but rather distributed across multiple hardware components, each under the control of an independent agent. The resulting architecture represents a departure from monolithic robot control systems and enables a genuinely distributed approach to robotic task execution and accountability.

Assigning distinct Control Domains to each Persona within a robotic system facilitates a clear delineation of responsibility during task execution. This means each Persona is exclusively responsible for managing and controlling a specific set of the robot’s physical subsystems – for example, locomotion, manipulation, or perception. Consequently, when an action is performed, the originating Persona, and thus its associated Control Domain, can be directly identified. This contrasts with a single-agent system where tracing the origin of an action requires analyzing the entire control stack, and simplifies debugging, performance analysis, and the assignment of accountability in complex scenarios. The granular control offered by this approach also allows for independent operation and potential specialization of each Persona within the broader robotic system.

Single Agent Embodiment represents a traditional robotic architecture wherein a singular, centralized controller processes all sensory input and generates commands for the entirety of the robot’s physical actuators. This contrasts with distributed approaches by placing complete operational responsibility, including perception, planning, and control, within a single computational unit. Consequently, any failure or limitation within this central controller directly impacts the robot’s overall functionality and introduces a single point of failure. Debugging and isolating the source of errors can be complex due to the interconnected nature of all control processes within this unified system, and scaling complexity often necessitates increased computational resources for the single controller.

Blame Compartmentalization: A Pathway to Recovered Trust

Experiments utilizing a `Split Embodiment` robotic system demonstrated the facilitation of `Blame Compartmentalization` in human subjects following task failures. Specifically, participants were able to attribute responsibility to a particular robotic persona, even when a single physical robot executed the actions. This isolation of blame occurred across multiple collaborative tasks – including sorting, arrangement, and motivational support – and suggests that users perceive and assess the competence of distinct personas independently. The observed effect indicates a cognitive mechanism wherein failures are not generalized to the entire system, but rather localized to the specific persona deemed responsible, which has implications for trust recovery in multi-agent robotic systems.

Collaborative tasks were conducted utilizing the Hello Robot Stretch 3 platform, equipped with YOLO Object Detection for environmental perception. Participants engaged in three distinct task types: a Sorting Task requiring object categorization and placement; an Arrangement Task involving spatial organization of objects; and a Motivational Support Task designed to assess collaborative problem-solving. The Stretch 3 robot served as the physical agent within these tasks, while YOLO Object Detection enabled it to identify and interact with objects within the workspace, facilitating the completion of the assigned activities by the human-robot team.

Experimental results demonstrate a statistically significant improvement in trust recovery when utilizing a split-embodiment robot compared to traditional single-agent designs. Specifically, the Bayes Factor of 182.33 provides strong evidence that users differentiate between personas within the split-embodiment system when assessing competence following a failure. This indicates that clear attribution of responsibility to a specific persona facilitates a more nuanced evaluation, allowing users to maintain trust in the overall system even after an individual component (persona) makes an error. The magnitude of the Bayes Factor suggests this differentiation is not merely a statistical artifact but a robust finding supported by the data.

Statistical analysis using Fisher’s Exact Test demonstrated a significant difference in how mistakes were attributed between the split-embodiment and co-embodiment conditions (p-value = 0.04). This result indicates that participants were more likely to attribute errors to a specific persona when the robot utilized a split-embodiment approach, as opposed to viewing the robot as a single, unified entity in the co-embodiment condition. This differentiation in mistake attribution provides empirical support for the hypothesis that split embodiment facilitates blame compartmentalization, allowing users to isolate responsibility to individual components of the robotic system.

Statistical analysis using a Bayes Factor of 100.13 provides strong evidence that competence assessments are significantly influenced by both the specific task being performed and the persona embodying the robot. This model demonstrates that task difficulty and the perceived capabilities of the persona are independent factors impacting user evaluation. Critically, the interaction effect between task and persona indicates that the assessment of competence is not simply additive; rather, the perceived competence of a given persona varies depending on the demands of the task, suggesting a nuanced user evaluation process.

The Future of Collaboration: Towards Accountable Machines

Sustained collaboration between humans and robots in challenging environments hinges on a robust capacity to navigate instances of error and assign responsibility appropriately. When a robot fails, simply correcting the issue isn’t enough; humans require an understanding of why the failure occurred and whether the robot can be trusted to avoid similar errors in the future. Research indicates that effectively managing blame – meaning, avoiding disproportionate or unfair attribution of fault – is directly correlated with maintaining human trust over time. A robot perceived as transparent in its actions and accountable for its mistakes fosters a collaborative dynamic, while one that deflects responsibility erodes confidence and hinders long-term partnership, even if the error itself is minor. Consequently, designing robotic systems that prioritize clear explanations and demonstrable learning from failures is paramount for successful and enduring human-robot interaction.

Successful human-robot collaboration hinges on more than just a robot’s technical skill; it fundamentally relies on the psychological landscape of human trust. This research demonstrates that even demonstrably capable robots can erode user confidence if accountability for errors isn’t clearly established. The study reveals that humans don’t simply assess what a robot can do, but how it handles failures and accepts responsibility – or deflects it. Ignoring these crucial psychological factors risks creating a dynamic where users are hesitant to rely on robotic assistance, even when it’s objectively beneficial, ultimately hindering the potential for effective and sustained teamwork. Therefore, designing robots with transparent decision-making processes and clear mechanisms for acknowledging mistakes is paramount for building robust and productive human-robot partnerships.

Extending the applicability of this collaborative architecture represents a vital next step, demanding investigation across diverse task domains – from surgical assistance and disaster response to complex manufacturing and space exploration. Crucially, future studies must account for variations in human expertise; a novice user will likely require a different level of robotic transparency and explanation than an experienced professional. Adapting the system’s feedback mechanisms and levels of autonomy to match the operator’s skill – offering more detailed guidance to beginners and greater control to experts – promises to optimize performance and build enduring trust. This necessitates developing flexible interfaces and algorithms capable of dynamically adjusting to the user’s cognitive load and evolving skillset, ultimately paving the way for truly personalized and effective human-robot partnerships.

The potential for truly effective human-robot collaboration hinges on building machines that operate with transparency and accountability. When a robot’s actions and reasoning are readily understandable – revealing why a decision was made, not just what was decided – humans are better equipped to anticipate its behavior and intervene appropriately. This clarity is paramount for establishing trust, particularly when errors occur; a robot that can explain its mistakes, and demonstrate an understanding of the consequences, allows for constructive feedback and shared learning. Such a dynamic fosters a collaborative partnership, moving beyond simple task allocation to a synergistic relationship where humans and robots leverage each other’s strengths, ultimately boosting productivity and innovation in complex environments.

The study meticulously charts how humans navigate the complexities of attributing failure within a single, yet fragmented, robotic system. This echoes a fundamental truth regarding all complex systems – their inherent fragility over time. Robert Tarjan aptly observed, “Debugging is twice as hard as writing the code in the first place. Therefore, if you write the code as cleverly as possible, you are, by definition, not smart enough to debug it.” The researchers demonstrate that humans readily compartmentalize responsibility, effectively assigning blame to specific ‘domains’ of the robot – a pragmatic approach to managing expectations when a unified entity falters. It’s a recognition that even within a seemingly cohesive system, components degrade, and understanding where the failure originates becomes crucial – a concept mirrored in managing the version history of any evolving creation.

The Long Game

This work demonstrates a human capacity to ascribe agency – and importantly, to compartmentalize fault – even within a demonstrably unified system. Every architecture lives a life, and this one reveals a preference for narrative coherence over mechanistic understanding. The finding that users differentiate between ‘sides’ of a single robot suggests a fundamental tendency to populate the world with intentional actors, regardless of the underlying reality. This is not necessarily progress; it is simply observation.

Future explorations must acknowledge the transient nature of these perceptions. The ease with which these separate identities are established, and the durability of that perception over extended interaction, remains unknown. Improvements age faster than one can understand them. A crucial next step involves examining how these attributions shift when the ‘split’ robot experiences repeated, varied failures – does the compartmentalization hold, or does the system inevitably converge on a single point of blame, effectively erasing the constructed personas?

Ultimately, the question isn’t whether humans can be deceived by clever embodiment, but how long before the illusion frays. All systems decay. The interesting metric will not be the initial success of this technique, but the rate at which the carefully constructed identities are eroded by the inevitable imperfections of the machine itself.


Original article: https://arxiv.org/pdf/2602.07598.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-02-10 18:16