Author: Denis Avetisyan
A new framework leverages real-to-sim transfer and lightweight machine learning to significantly improve the efficiency and reliability of remote manipulation tasks.

This work presents a shared autonomy system employing residual learning, a kNN human surrogate, and effective real-to-sim transfer for robust contact-rich manipulation.
Despite advances in robotics, fine-grained, contact-rich manipulation remains challenging for human operators due to its inherent slowness and potential for error. This paper introduces a novel framework for ‘Efficient and Reliable Teleoperation through Real-to-Sim-to-Real Shared Autonomy’ that addresses this limitation by augmenting human control with a learned corrective âcopilotâ policy. Utilizing a lightweight k-nearest neighbor human surrogate trained on minimal real-world data, our system enables stable transfer from simulation to the real world, improving both task success rates and execution efficiency for operators of all skill levels. Could this approach unlock more intuitive and robust human-robot collaboration in complex manipulation tasks across diverse applications?
Deconstructing Control: The Fragility of Precision
Traditional robotic control systems frequently falter when confronted with tasks demanding delicate manipulation, particularly outside of highly structured laboratory settings. These systems typically rely on precise pre-programming and predictable environments, making them ill-equipped to handle the inherent uncertainties of the real world. Minute variations in object position, unexpected contact forces, or unforeseen obstacles can quickly derail a carefully planned sequence. The rigidity of these approaches contrasts sharply with human dexterity, which effortlessly adapts to changing conditions and imprecise inputs. Consequently, robots struggle with even seemingly simple actions-like grasping a novel object or assembling components with slight imperfections-highlighting a critical need for more robust and adaptable control strategies that can bridge the gap between programmed precision and real-world variability.
Successfully executing intricate tasks like assembling complex products or performing in-space servicing necessitates a departure from rigidly pre-programmed robotic routines. These scenarios are rarely predictable; variations in part placement, lighting conditions, or unexpected contact forces introduce substantial uncertainty. A truly capable system must therefore move beyond simply repeating memorized motions and instead demonstrate a capacity for real-time adaptation. This demands advanced sensing capabilities – allowing the robot to perceive its environment and the state of manipulated objects – coupled with intelligent algorithms that can reason about potential outcomes and adjust actions accordingly. The ability to robustly handle such uncertainty isnât merely about preventing failures; itâs about achieving the efficiency and reliability required for widespread deployment in dynamic, real-world applications.
The reliance on human teleoperation for intricate robotic tasks introduces significant bottlenecks in both speed and practicality. While effective for complex scenarios, directly controlling a robot remotely demands sustained cognitive effort and physical exertion from the operator, leading to fatigue and reduced performance over time. This limitation hinders scalability, as each robot requires dedicated human oversight, and diminishes efficiency, slowing down task completion rates. Consequently, industries aiming for widespread robotic deployment in fields like space exploration, disaster response, or precision assembly face considerable challenges in overcoming the inherent constraints of human-in-the-loop control systems, prompting research into more autonomous or semi-autonomous manipulation strategies.
Robotic systems encounter significant hurdles when performing contact-rich manipulation – tasks fundamentally reliant on sustained physical interaction with the environment. Unlike simple pick-and-place operations, successfully meshing gears, inserting a peg into a hole, or threading a nut onto a bolt demands precise force control and nuanced tactile sensing. These actions arenât solely about reaching a target position; they require the robot to actively feel for contact, adapt to slight misalignments, and apply just the right amount of force to overcome friction and achieve a secure connection. The inherent uncertainty in these interactions – variations in part geometry, surface texture, and external disturbances – necessitate advanced algorithms capable of real-time adaptation and robust error recovery, pushing the boundaries of current robotic dexterity and intelligence.

Transcending Control: Shared Autonomy and the Collaborative Machine
Shared autonomy represents a fundamental change in human-robot interaction, moving away from fully autonomous systems designed to operate independently and towards collaborative models where robots function as assistive tools. This paradigm shift prioritizes augmenting human capabilities rather than outright replacement, leading to increased efficiency in complex tasks and improved safety through the robotâs ability to handle repetitive or hazardous sub-components. By distributing workload and leveraging the strengths of both human operators – including adaptability, judgment, and intuitive problem-solving – and robotic systems – characterized by precision, repeatability, and endurance – shared autonomy unlocks performance levels exceeding those attainable by either entity alone. This approach is particularly relevant in industries requiring nuanced control and real-time decision-making, such as surgery, manufacturing, and disaster response.
Shared autonomy systems are designed to capitalize on complementary capabilities. Human operators excel in tasks requiring adaptability to unforeseen circumstances, intuitive judgment in complex environments, and high-level decision-making. Conversely, robots consistently deliver precise movements, repeatable actions, and sustained performance without fatigue. By integrating these strengths, shared autonomy aims to offload repetitive or physically demanding components of a task to the robot, while retaining human oversight and enabling flexible responses to dynamic situations. This division of labor results in increased overall efficiency, reduced error rates, and improved safety compared to purely manual or fully automated approaches.
For shared autonomy to function effectively, robotic systems must move beyond reactive responses and actively predict the operatorâs goals. This necessitates the integration of intent recognition algorithms, frequently utilizing sensor data – including biomechanical signals, gaze tracking, and environmental context – to infer the operatorâs desired actions. Proactive assistance, delivered based on these predictions, then involves the robot offering support before the operator explicitly requests it, such as pre-positioning tools, adjusting trajectory parameters, or providing force assistance. The timing and magnitude of this assistance are critical; interventions must be subtle enough to avoid hindering the operator but significant enough to demonstrably improve performance or safety.
Establishing an âExpert Priorâ involves incorporating data derived from skilled human operators to define a baseline for optimal robotic behavior within a shared autonomy system. This data, typically collected through methods like motion capture or teleoperation recordings, characterizes successful task execution in terms of kinematic trajectories, force application, and decision-making processes. The âExpert Priorâ serves as a starting point for the robotâs control algorithms, enabling it to anticipate likely operator actions and offer assistance aligned with established best practices. By learning from demonstrated expertise, the system reduces the need for extensive real-time adaptation and accelerates the development of robust and intuitive shared control strategies, particularly in complex or safety-critical scenarios.

Unlocking Assistance: Copilot Learning and the Predictive Machine
Copilot Learning facilitates the development of robotic assistance policies by directly linking robot actions to both operator commands and the current system state. This conditioning on both input and context allows the robot to provide support tailored to the specific situation and the operatorâs intent, rather than executing pre-programmed behaviors. The resulting policies are not simply reactive to commands; they are predictive and adaptive, anticipating the operator’s needs based on the observed system state and dynamically adjusting assistance accordingly. This context-aware support enhances the operator’s control and reduces cognitive load by providing timely and relevant assistance when and where it is most effective.
The Human Surrogate model functions as a behavioral cloning system, generating training data for the assistance policy. It leverages the k-Nearest Neighbor (kNN) algorithm to identify past operator actions most similar to the current system state and pilot command. Specifically, kNN retrieves a distribution of previously successful control signals – such as steering angles or actuator commands – from a dataset of human demonstrations. This distribution then serves as a target for the assistance policy to learn from, effectively mimicking the operatorâs typical behavior. The surrogate model does not attempt to optimize a reward function; instead, it provides a direct, data-driven estimate of the humanâs intended control action, enabling the assistance policy to learn from observed expert behavior.
The Real-to-Sim-to-Real pipeline addresses the challenges of robotic system training by leveraging simulation for initial policy development. This process begins with training an assistance policy within a physics-based simulation environment, allowing for rapid iteration and data collection without the risks and costs associated with real-world experimentation. Subsequently, techniques like domain randomization are employed to increase the robustness of the learned policy to discrepancies between the simulated and real environments. Finally, the policy is transferred to the physical robot for further refinement, often requiring minimal real-world data due to the prior simulation training, and enabling efficient deployment of assistance capabilities.
The assistance policy is refined through model-free reinforcement learning utilizing a residual formulation. This approach frames the learning problem as identifying corrective actions to supplement the human operatorâs existing control inputs. Rather than learning a complete control policy from scratch, the system learns a residual – the difference between the desired behavior and the humanâs baseline control. This significantly reduces the complexity of the learning task and accelerates convergence. The residual policy is trained to predict actions that minimize a defined reward function, effectively guiding the robot to provide assistance only when and where it is needed, while respecting the operatorâs overall control authority. This allows the robot to adapt to variations in human control styles and improve performance over time.

Measuring the Impact: Refining Control Through Human-Robot Synergy
Evaluating the shared autonomy systemâs success requires quantifying the cognitive demand placed on the human operator, and this is achieved, in part, through the use of the NASA Task Load Index (NASA-TLX). This established metric assesses workload across six subscales – mental demand, physical demand, temporal demand, performance, effort, and frustration – providing a comprehensive view of operator state during robotic task execution. By systematically measuring these factors, researchers can determine whether the copilot learning component effectively reduces the burden on the operator, allowing them to focus on supervisory control rather than low-level manipulation. Lower NASA-TLX scores indicate a more efficient and less stressful experience, demonstrating that the shared autonomy system is achieving its goal of augmenting human capabilities and streamlining complex robotic operations.
The developed shared autonomy system demonstrably reduces the cognitive burden on human operators while simultaneously boosting task success. Evaluations reveal a significant shift from the intensive focus required in traditional teleoperation – where constant, direct control is needed – to a more supervisory role. This is achieved by the systemâs ability to anticipate operator needs and provide adaptive assistance, allowing individuals to concentrate on high-level planning and decision-making rather than low-level motor control. Consequently, operators experience reduced mental fatigue and improved efficiency, as evidenced by increased task completion rates and a measurable lessening of workload when compared to conventional methods of remote operation. The observed gains aren’t merely incremental; the system consistently outperforms direct teleoperation, suggesting a fundamental improvement in the human-robot interaction paradigm.
This shared autonomy framework streamlines complex manipulation by offering adaptive assistance throughout task execution. Rather than requiring constant, low-level control inputs, the system intelligently anticipates operator needs and provides support tailored to the specific demands of each movement. This allows human operators to shift their focus from the intricacies of robotic control – such as precise positioning and force application – towards higher-level strategic decision-making, like planning the overall manipulation sequence and troubleshooting unexpected situations. By offloading routine tasks and providing intuitive support, the framework not only enhances performance metrics like success rates and completion times, but also significantly reduces operator cognitive load, fostering a more natural and efficient human-robot interaction.
The systemâs enhanced performance stems from the implementation of Admittance and Task Space Control, which allow the robot to move with a degree of flexibility previously unattainable in traditional teleoperation. These control methods move beyond simply dictating position; instead, they prioritize compliant interaction with the environment, enabling the robot to yield to external forces and adjust its movements accordingly. This is particularly crucial for delicate manipulation tasks, such as gear meshing and insertion, where rigid control can lead to collisions or failed attempts. By controlling not just where the robot moves, but how it moves – its stiffness, damping, and responsiveness – the system achieves a more natural and forgiving interaction, significantly boosting success rates and reducing operator workload. The robot essentially learns to âfeelâ its way through complex tasks, mirroring the nuanced movements of a human operator and enhancing overall efficiency.
A notable improvement stemming from the developed system is a 30% increase in successful gear meshing when contrasted with traditional direct teleoperation methods. This gain highlights the efficacy of the shared autonomy framework in assisting with intricate manipulation tasks. By providing adaptive support, the system effectively reduces the precision demands on the operator, allowing for more consistent and reliable task completion. This performance boost isnât merely incremental; it represents a substantial advancement in robotic manipulation, potentially enabling the completion of complex assemblies in challenging environments where precision and consistency are paramount.
Evaluations of the shared autonomy system demonstrate a marked improvement in robotic manipulation success rates. Utilizing data gathered through the developed methodology, the system consistently achieved successful grasps and insertions in 19 out of 20 attempts. This performance stands in stark contrast to direct teleoperation, where operators managed only 6 successful grasps and completed zero successful insertions from the same number of attempts. The substantial difference highlights the efficacy of the copilot learning component and its ability to significantly enhance the robotâs dexterity and precision, ultimately streamlining complex tasks and reducing the burden on the human operator.
Comparative evaluations demonstrate a substantial performance advantage for the developed system when tackling complex manipulation tasks. Specifically, the shared autonomy approach achieved success in 18 out of 20 attempted grasps and 11 out of 20 insertions, representing a marked improvement over direct teleoperation. Under identical conditions and with a matched number of attempts, direct teleoperation yielded considerably lower success rates of only 7 out of 20 grasps and a single successful insertion out of 20. These results highlight the systemâs ability to significantly enhance task completion, suggesting that adaptive assistance effectively mitigates the challenges associated with precise robotic control and improves overall operational efficiency.

The pursuit of efficient teleoperation, as demonstrated in this work, inherently involves challenging established boundaries. One might consider the system not merely as a tool for control, but as a framework for exposing the limits of current assumptions about human-robot interaction. As Paul ErdĆs once stated, âA mathematician knows a lot of things, but a physicist knows everything.â This rings true here; the team doesnât simply aim for successful manipulation, but actively probes the interface, revealing what’s genuinely necessary for intuitive shared autonomy. The residual learning approach, allowing the system to adapt to operator nuances, isnât about fixing imperfections, but understanding the signal within the ‘bug’ of individual control styles – a delightful disruption of the expected.
Beyond the Copilot
The demonstrated success of transferring a learned âcopilotâ from simulation to reality is, predictably, not the end. Itâs a well-observed phenomenon that systems initially improve performance, then reveal their brittle underpinnings. This work skirts the edge of that cliff, achieving gains in teleoperation, but the limitations inherent in any surrogate model-be it kNN or a more complex learned representation-remain. The true test will be forcing this system to truly fail – not just in predictable scenarios, but in creatively adversarial ones. Only then can one begin to understand the boundaries of its competence and, more importantly, the nature of the errors it is most likely to propagate.
Future iterations should explicitly address the question of âwhat is not known.â Current approaches focus on doing-augmenting human capability. A parallel, and arguably more crucial, line of inquiry involves quantifying uncertainty. A system that can reliably state âI donât knowâ – and, critically, why it doesnât know – is demonstrably more trustworthy than one that confidently proceeds into the unknown. Furthermore, exploration of alternative residual learning strategies, perhaps incorporating predictive models of operator intent, could unlock even greater levels of shared control.
Ultimately, the goal isn’t simply efficient manipulation. It’s the construction of a symbiotic relationship between human and machine, one built on mutual understanding and a shared acknowledgement of fallibility. The illusion of perfect control is a dangerous one. True progress lies in embracing the inherent messiness of complex systems and designing for graceful degradation, not flawless execution.
Original article: https://arxiv.org/pdf/2603.17016.pdf
Contact the author: https://www.linkedin.com/in/avetisyan/
See also:
- CookieRun: Kingdom 5th Anniversary Finale update brings Episode 15, Sugar Swan Cookie, mini-game, Legendary costumes, and more
- Gold Rate Forecast
- 3 Best Netflix Shows To Watch This Weekend (Mar 6â8, 2026)
- How to get the new MLBB hero Marcel for free in Mobile Legends
- eFootball 2026 JĂŒrgen Klopp Manager Guide: Best formations, instructions, and tactics
- American Idol vet Caleb Flynn in solitary confinement after being charged for allegedly murdering wife
- Seeing in the Dark: Event Cameras Guide Robots Through Low-Light Spaces
- eFootball 2026 is bringing the v5.3.1 update: What to expect and whatâs coming
- Kat Graham gives birth to a baby boy! Vampire Diaries star welcomes first child with husband Bryant Wood
- PUBG Mobile collaborates with Apollo Automobil to bring its Hypercars this March 2026
2026-03-19 15:49