Putting Manipulation in Human Hands

Author: Denis Avetisyan


Researchers have unveiled a comprehensive ecosystem designed to teach robots how to perform everyday tasks with human-like dexterity.

The study visualizes task variations within a cross-embodiment manipulation framework, demonstrating how subtle shifts in approach can redefine operational parameters.
The study visualizes task variations within a cross-embodiment manipulation framework, demonstrating how subtle shifts in approach can redefine operational parameters.

The World In Your Hands dataset, sensing system, and benchmarks enable large-scale research in human-centric dexterous manipulation and embodied AI.

Despite recent advances in AI, scaling dexterous manipulation to real-world scenarios remains a significant challenge due to limited and inadequately diverse training data. This work introduces World In Your Hands: A Large-Scale and Open-source Ecosystem for Learning Human-centric Manipulation in the Wild, a comprehensive resource encompassing over 1,000 hours of multi-modal data captured with a novel wearable sensing system, alongside extensive annotations and benchmarks. Our ecosystem demonstrably enhances the generalization and robustness of hand manipulation policies in tabletop tasks, offering a crucial step towards more adaptable embodied AI. Will this open-source platform catalyze a new wave of innovation in human-robot interaction and unlock the full potential of dexterous manipulation?


The Inevitable Bridge: Addressing the Reality Gap in Robotic Learning

Robotic manipulation in realistic settings presents a significant challenge due to the inherent unpredictability and complexity of the physical world. Traditional learning approaches often falter when confronted with cluttered environments, where objects occlude sensors, lighting conditions vary, and precise movements are constantly disrupted by unforeseen contact. These difficulties stem from the sheer dimensionality of the problem – robots must simultaneously account for their own configuration, the pose of multiple objects, and the forces exerted during interaction. Consequently, policies learned in simplified scenarios often fail to generalize, leading to brittle performance and a substantial gap between simulation and real-world capabilities. Overcoming this ‘reality gap’ necessitates new strategies that enable robots to robustly adapt to the messiness and uncertainty characteristic of everyday environments.

Despite advancements in robotic simulation, a significant fidelity gap often hinders the successful deployment of policies learned in virtual environments to physical robots. These discrepancies arise from imperfect modeling of real-world physics, sensor noise, and unpredictable environmental factors – elements readily present in the physical world but difficult to replicate accurately in simulation. Consequently, a robot trained solely in simulation may exhibit degraded performance or even failure when confronted with the nuances of a real-world task, necessitating costly and time-consuming adaptation procedures. Researchers are actively exploring techniques like domain randomization – deliberately varying simulation parameters – and domain adaptation – refining policies to bridge the gap – to improve the transferability of learned skills and unlock the full potential of simulation-based robotic learning.

Real-robot manipulation experiments demonstrate that co-training policies with both UMI data and annotated human demonstrations (WiYH) enhances performance across four novel task settings compared to training exclusively with UMI data.
Real-robot manipulation experiments demonstrate that co-training policies with both UMI data and annotated human demonstrations (WiYH) enhances performance across four novel task settings compared to training exclusively with UMI data.

Echoes of Expertise: Human-Centric Learning as a Path Forward

Human-Centric Learning represents a significant departure from traditional robotic learning methodologies by directly incorporating knowledge derived from human action. This approach acknowledges that humans possess an inherent understanding of physical interactions and task execution, often developed through years of experience. By leveraging this intuitive knowledge – typically captured through observation or demonstration – robots can bypass the need for extensive trial-and-error learning in complex environments. This differs from purely reinforcement learning or imitation learning techniques, as it focuses on extracting and applying the reasoning behind human actions, rather than simply replicating observed behaviors. The core principle is to transfer the efficiency and adaptability of human problem-solving to robotic systems, ultimately leading to more robust and generalizable performance.

The CoTraining method addresses limitations in robot learning by integrating data from two primary sources: autonomous robot experimentation and curated HumanVideoDemonstrations. This approach allows the robot to leverage the breadth of data achievable through self-exploration, while simultaneously benefiting from the efficiency and accuracy of human-provided examples. Specifically, the robot initially learns from its own trial-and-error interactions with the environment, then refines its understanding using the labeled data present in the human videos. This synergistic process accelerates the learning curve and improves the robot’s ability to generalize to novel situations, resulting in more robust performance compared to systems relying solely on either robot-collected or human-provided data.

Implementation of the Human-Centric Learning approach resulted in a measurable 13% improvement in task completion rates within single-object environments. More significantly, the method enabled successful operation in previously intractable multi-object cluttered scenes, where prior attempts consistently failed. This demonstrates the technique’s capacity to address challenges posed by increased environmental complexity and highlights its potential for robust performance in real-world applications characterized by variable and congested settings.

Unlike the UMI dataset, which focuses on manipulation from constrained initial states, the Human Video dataset captures a wider range of actions performed in more complex environments.
Unlike the UMI dataset, which focuses on manipulation from constrained initial states, the Human Video dataset captures a wider range of actions performed in more complex environments.

The World in Your Hands: A System for Data and Benchmarking

The WorldInYourHands (WiYH) ecosystem is designed to accelerate research in human-centric robotic learning by integrating data acquisition tools, a large-scale dataset, and standardized benchmarking procedures. This framework allows researchers to collect data from robotic platforms and human demonstrations, then evaluate the performance of new algorithms against established baselines. The ecosystem’s tools support the capture of diverse data modalities, including visual, tactile, and kinematic information, which is then compiled into the WiYHDataset. Standardized benchmarks within the WiYH framework provide a common evaluation platform, facilitating reproducible research and accelerating progress in robotic manipulation and learning.

The DexUMI and OracleSuite systems are central to data acquisition within the WorldInYourHands ecosystem. DexUMI provides a robotic platform capable of executing precise manipulation tasks, generating data related to robotic actions and object interactions. Simultaneously, OracleSuite captures a range of multi-modal sensory information, including visual data from cameras, force/torque readings from sensors, and potentially other modalities like tactile sensing. This combined data stream, generated through both robotic execution and comprehensive sensory input, allows for the creation of datasets suitable for training and benchmarking human-centric robotic learning algorithms. The systems are designed to ensure high-precision data capture, crucial for accurately modeling and replicating human manipulation skills.

The WiYHDataset is a large-scale collection of human manipulation sequences built by augmenting existing datasets with new recordings from two primary sources: HumanVideoDemonstrations and DexUMI. HumanVideoDemonstrations contributes data captured from human actors performing manipulation tasks, providing a diverse range of natural behaviors. Complementing this, DexUMI provides high-precision, robot-performed manipulations with corresponding multi-modal sensory data, including force/torque measurements and visual observations. This combination expands the scope and fidelity of available data for training and benchmarking robotic learning algorithms, exceeding the limitations of single-source datasets.

The Oracle Suite is a human-centric data collection system comprising the <span class="katex-eq" data-katex-display="false">H-FPVHive</span> (a multi-modal perception suite), <span class="katex-eq" data-katex-display="false">H-Glove</span> (a synchronized hand motion and tactile perception module), and <span class="katex-eq" data-katex-display="false">H-Backpack</span> (a portable power and data storage unit) to enable comprehensive action localization and capture in real-world environments.
The Oracle Suite is a human-centric data collection system comprising the H-FPVHive (a multi-modal perception suite), H-Glove (a synchronized hand motion and tactile perception module), and H-Backpack (a portable power and data storage unit) to enable comprehensive action localization and capture in real-world environments.

Beyond Observation: Visual-Linguistic Action Understanding for Robust Control

The development of robust robotic manipulation increasingly relies on systems that mimic human cognitive abilities, and a key component of this progress is the integration of visual, linguistic, and action data – a concept known as VisionLanguageAction. This approach moves beyond traditional robot programming by allowing machines to not only see an environment but also understand instructions expressed in natural language and connect those instructions to appropriate physical actions. By processing information from these three modalities simultaneously, robots can build a more comprehensive and nuanced understanding of manipulation tasks, enabling them to adapt to novel situations, resolve ambiguities in instructions, and ultimately perform complex tasks with greater reliability and flexibility – a significant step toward truly intelligent and versatile robotic systems.

The convergence of visual and linguistic data empowers robots with increasingly sophisticated manipulation skills through capabilities like spatial referencing, subtask prediction, and completion verification. Spatial referencing allows a robot to accurately identify objects within a scene based on natural language instructions – for instance, distinguishing “the red block” from others. Crucially, subtask prediction enables the robot to anticipate the necessary sequence of actions to achieve a goal, streamlining execution and improving efficiency. Finally, completion verification provides a mechanism for assessing whether a task has been successfully completed, either through direct observation or by confirming that expected conditions have been met, fostering greater autonomy and reliability in complex environments.

Advancements in robotic perception leverage 4D reconstruction to build a richer understanding of changing environments. By utilizing techniques such as Gaussian Splatting, robots move beyond static scene analysis to model the dynamic relationships between objects over time. This capability is crucial for complex manipulation tasks, allowing for more accurate predictions of object trajectories and improved grasp planning. Reported performance gains, as quantified by Peak Signal-to-Noise Ratio (PSNR) metrics detailed in Appendix B, demonstrate a significant increase in the fidelity of reconstructed scenes and, consequently, a boost in the robot’s ability to interact effectively with its surroundings. The system doesn’t simply ‘see’ a scene; it anticipates how that scene will evolve, leading to more robust and adaptable control strategies.

The dataset encompasses a diverse range of real-world scenarios-from industrial settings to daily life-and provides detailed task and subtask annotations that are crucial for aligning instructions with actions and enabling task decomposition in vision-language action (VLA) models.
The dataset encompasses a diverse range of real-world scenarios-from industrial settings to daily life-and provides detailed task and subtask annotations that are crucial for aligning instructions with actions and enabling task decomposition in vision-language action (VLA) models.

Toward Complex Behaviors: A Glimpse into the Future of Robotic Learning

Recent progress in robotics is increasingly fueled by a convergence of approaches centered on how humans learn and interact with the world. By prioritizing human-centric learning – where robots are taught through demonstrations and natural language – alongside the collection of extensive, diverse datasets, researchers are building systems capable of more nuanced actions. This is further enhanced by advancements in visual-linguistic understanding, allowing robots to not only ‘see’ but also interpret the meaning of objects and instructions within a scene. The result is a pathway toward generating robotic behaviors that are not merely pre-programmed, but are adaptable, complex, and capable of responding effectively to unforeseen circumstances – a crucial step toward truly versatile and helpful robotic assistants.

Recent progress in robotics envisions a future where robots seamlessly interpret and execute commands articulated in natural language. LanguageConditionedVideoGeneration represents a key step toward this goal, enabling the creation of robotic systems capable of sophisticated manipulation based solely on linguistic input. By learning from extensive datasets linking language to observed actions, these models can generate realistic video sequences depicting robots performing complex tasks-from arranging objects to utilizing tools-in response to instructions like “stack the red block on the blue one.” This capability transcends pre-programmed routines, allowing for greater adaptability and the potential for robots to address novel situations and intricate requests without explicit re-coding, ultimately paving the way for more intuitive and versatile human-robot interaction.

Evaluations utilizing the VBench benchmark reveal significant enhancements in the robotic behaviors produced by the fine-tuned models. Generated videos consistently exhibit greater fidelity in movement, appearing smoother and more natural than previous iterations. Importantly, these improvements extend beyond mere aesthetics; dynamic accuracy – the robot’s ability to precisely execute instructed actions – has demonstrably increased. This is further corroborated by gains in overall video quality, indicating that the models are not simply generating plausible motions, but rather, are producing realistic and physically-grounded robotic performances. These metrics collectively validate the effectiveness of the approach in creating more capable and believable robotic systems.

Fine-tuning with our dataset substantially reduces hallucinations in language-conditioned video prediction models, enabling them to more accurately imagine future video states from textual instructions.
Fine-tuning with our dataset substantially reduces hallucinations in language-conditioned video prediction models, enabling them to more accurately imagine future video states from textual instructions.

The WiYH ecosystem, as detailed in the study, strives to capture the nuances of human manipulation – a system inherently complex and prone to entropy. This pursuit echoes Claude Shannon’s observation that, “The most important thing in communication is to convey the meaning, not the message.” WiYH isn’t merely collecting data points; it’s attempting to encode the meaning of dexterous manipulation, the underlying principles that allow for adaptability and problem-solving in unpredictable environments. The scale of the dataset acknowledges that any simplification of this process – reducing it to a limited set of actions or scenarios – carries a future cost, inevitably leading to a degradation of performance when confronted with the truly ‘wild’ variations of real-world interaction. The system’s long-term viability rests on its ability to gracefully accommodate this decay, continuously learning and refining its understanding of human-centric manipulation.

What Lies Ahead?

The WiYH ecosystem, as presented, is less a solution and more a detailed logging of the current state. Any system built on observation-even one as ambitious as capturing human manipulation-inherently records decay. The benchmarks established will, inevitably, be surpassed. The true metric isn’t the score achieved, but the rate at which those scores improve-a measure of the field’s responsiveness to its own limitations. Deployment of this dataset is merely a moment on that timeline.

A persistent challenge lies in the inherent ambiguity of ‘human-centric’ tasks. The dataset captures actions, but not necessarily the intent behind them. Future work must address this gap, moving beyond imitation towards genuine understanding. This necessitates richer annotation, incorporating not just what was done, but why. The system’s chronicle, however extensive, remains incomplete without that contextual layer.

The long view suggests a need for systems that don’t merely react to observed behavior, but anticipate it. This requires a shift in focus: from collecting examples of dexterity to modelling the underlying principles of adaptability. The WiYH ecosystem provides a valuable foundation, but the next iteration must consider not just what hands can do, but how they learn to do it, and what that says about intelligence itself.


Original article: https://arxiv.org/pdf/2512.24310.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-01-04 02:17