Author: Denis Avetisyan
New research reveals that robots can effectively learn from human video data, but only after undergoing extensive training with diverse robotic experiences.

Diverse robotic pretraining is critical for successful cross-embodiment transfer from human demonstrations to robotic control in vision-language-action models.
Despite the promise of broadly capable embodied agents, scaling vision-language-action (VLA) models remains data-intensive, prompting exploration of readily available human video as a potential source of training data. In ‘Emergence of Human to Robot Transfer in Vision-Language-Action Models’, we demonstrate that successful transfer from human to robot relies not on bespoke engineering, but on pre-training VLAs on sufficiently diverse robotic datasets. This diverse pre-training fosters embodiment-agnostic representations, effectively unlocking the potential of human video for improved generalization in robotic systems. Could this emergent capability pave the way for foundation models capable of seamlessly learning from diverse demonstrations, regardless of the agent performing the action?
Bridging the Gap: Toward Embodied Intelligence
Conventional robotic learning methodologies frequently encounter limitations when applied to novel scenarios, hindering a robot’s ability to perform tasks outside of its specific training parameters. This fragility stems from an over-reliance on meticulously curated datasets and a difficulty in adapting to even slight variations in task requirements or environmental conditions. Robots trained in one setting often demonstrate a precipitous decline in performance when confronted with unfamiliar objects, altered lighting, or unexpected obstacles – a phenomenon known as poor generalization. Consequently, significant research efforts are directed towards developing more robust learning algorithms that enable robots to extrapolate knowledge from limited experience and operate reliably in the dynamic and unpredictable real world, mirroring the adaptability inherent in biological systems.
A significant challenge in robotics lies in the disparity between how humans and robots perceive and interact with the world – a phenomenon known as the embodiment gap. Current machine learning techniques often require robots to relearn skills with each new physical form or environment, mirroring a human having to consciously relearn to walk after switching bodies. This limitation stems from robots typically learning tasks in a manner tightly coupled to their specific morphology and sensor suite; knowledge gained by one robot cannot be readily applied to another, even if the tasks are conceptually identical. Consequently, developing robots capable of adaptable and generalized behavior requires overcoming this transfer problem, allowing them to leverage knowledge acquired through diverse embodiments – a feat humans accomplish with remarkable ease.
A novel vision-language-action (VLA) model offers a pathway toward more versatile robotic learning by establishing shared understandings between human and machine. This model doesn’t simply process visual inputs or language commands in isolation; instead, it’s engineered to concurrently interpret what an agent sees, what it is told to do, and the resulting actions performed. By training on datasets encompassing both human demonstrations and robotic executions, the VLA model learns a unified representation of tasks-a common ‘language’ expressing goals and procedures irrespective of the physical body performing them. This shared representation allows knowledge gained from human examples to be readily transferred to a robot, and vice versa, circumventing the typical embodiment gap that hinders generalization and enabling more robust performance across diverse situations.
The pursuit of adaptable robotics hinges on overcoming limitations in generalization – the ability to perform well in unseen situations. This research proposes a unified learning framework designed to address this challenge by fostering shared understandings between humans and robots. Instead of training robots for specific tasks and environments, the model learns a common language that connects visual perception, linguistic instruction, and physical action. This shared representation allows the system to leverage knowledge gained from one embodiment – be it human demonstration or robotic experience – and apply it effectively to novel scenarios and even different robotic platforms. The result is a pathway towards more robust and versatile robots capable of seamlessly operating in dynamic, real-world settings, reducing the need for extensive retraining with each new challenge.

Establishing a Foundation: Pretraining the VLA
The performance of the Visual Language Action (VLA) model is predicated on the development of representations that are both robust to variations in input data and independent of the specific robotic embodiment used for action execution. Robustness ensures consistent performance across diverse environments and sensor configurations, while embodiment agnosticism allows the model to generalize to new robots without requiring retraining or significant adaptation. This is achieved by focusing the learning process on the underlying relationship between visual inputs, linguistic instructions, and resultant actions, rather than memorizing specific robot kinematics or dynamics. Consequently, the model learns a generalized mapping applicable to a range of robotic platforms and operational scenarios, increasing its adaptability and overall efficacy.
Pretraining the Visual Language Agent (VLA) utilizes a two-stage process to establish a foundational understanding prior to adaptation for specific tasks. This strategy involves initially training the model on a large corpus of data without task-specific labels, enabling it to learn general relationships between visual inputs, linguistic instructions, and robotic actions. This initial phase aims to create a broadly capable agent, which is subsequently refined through fine-tuning on datasets tailored to particular applications, resulting in improved performance and efficiency compared to training directly on task-specific data.
The pretraining phase utilizes datasets comprising both robot teleoperation recordings and extensive human video data. Robot teleoperation data consists of paired visual observations, language instructions provided to a human operator, and the corresponding robot actions executed in response. Complementing this, the model is trained on large-scale human video datasets, exposing it to a diverse range of activities and visual scenes. This combined approach allows the VLA to learn correlations between visual inputs, linguistic descriptions, and observed behaviors across both robotic and human contexts, forming a broad base of multimodal understanding before downstream task adaptation.
The VLA model’s pretraining phase establishes correlations between multimodal data streams: visual inputs, natural language instructions, and the robot’s subsequent actions. This is achieved by exposing the model to extensive datasets containing paired observations of a scene, corresponding language commands given to a robot, and the recorded actions the robot performed in response. The model learns to predict likely actions given visual and linguistic inputs, and conversely, to anticipate visual outcomes from given actions and commands. This associative learning forms the basis for the model’s ability to generalize to novel situations and tasks during fine-tuning, as it has already established a foundational understanding of how language, vision, and action are interconnected.

Predictive Action: The VLA in Operation
The VLA model functions as a predictive system for robotic control, generating both continuous action outputs and discrete, high-level subtask designations. This dual output allows for both fine-grained motor control and strategic task planning within a robotic system. The model doesn’t simply react to stimuli; it anticipates required actions based on input, effectively forecasting the necessary sequence of movements and overarching goals. This predictive capability is central to the VLA’s functionality, enabling proactive rather than reactive robot behavior and facilitating more complex task execution.
The VLA model integrates both visual and linguistic information for action prediction through the use of dense language annotations. These annotations provide detailed descriptions of the desired robot behavior, going beyond simple keyword-based instructions. Specifically, the model processes visual input from the environment alongside these dense textual descriptions, allowing it to correlate observed states with nuanced language commands. This multimodal approach enables the VLA to understand complex requests and predict appropriate robot actions based on both what is seen and what is instructed, improving performance in scenarios requiring contextual awareness and precise execution of commands.
Action generation within the VLA model utilizes a flow matching framework coupled with discretized action representations termed FAST Action Tokens. Flow matching involves learning a continuous normalizing flow that transforms a simple distribution into the complex distribution of robot actions, enabling efficient sampling. FAST Action Tokens represent continuous action spaces as a finite set of discrete options, effectively quantizing the action space for improved computational efficiency and training stability. This discretization allows the model to predict actions as a sequence of these tokens, simplifying the output space while retaining a high degree of expressiveness in controlling robot behavior.
TSNE analysis was performed on the learned embodiment agnostic representations to evaluate the similarity of action manifolds between human and robot demonstrations. Results demonstrate that the model successfully learns a shared representation space where actions with similar kinematic properties, regardless of the executor (human or robot), are clustered together in the reduced-dimensional TSNE space. This indicates the model is not simply memorizing robot-specific trajectories but rather abstracting underlying action similarities, confirming the embodiment agnostic nature of the learned representations and suggesting potential for generalization to novel scenarios and even transfer to different robotic platforms.

Demonstrated Versatility: Benchmarking the VLA
The VLA model underwent rigorous evaluation using a suite of generalization benchmarks designed to assess its performance in previously unseen situations. This testing protocol moved beyond standard training environments, presenting the model with novel tasks, scenes, and objects to truly gauge its adaptability. Researchers specifically focused on how well the model could transfer learned skills to these unfamiliar contexts, pushing the boundaries of robotic versatility. The benchmarks weren’t simply about achieving high scores on known challenges, but rather about demonstrating a capacity for robust performance when confronted with the unexpected-a crucial step toward creating robots capable of operating effectively in the real world.
The VLA model exhibits a remarkable capacity for generalization, consistently outperforming existing approaches across a diverse range of robotic tasks, environments, and objects. Evaluations reveal substantial improvements in success rates, with gains reaching up to 71% when faced with novel scenarios not encountered during initial training. This robust performance indicates the model’s ability to adapt to previously unseen circumstances, suggesting a significant step toward creating robots capable of functioning reliably in real-world, unpredictable settings. The model doesn’t simply memorize specific actions; it learns underlying principles enabling it to apply learned skills to new challenges, marking a considerable advancement in robotic adaptability and versatility.
The VLA model’s ability to generalize to new situations is significantly improved through a co-training process, which integrates both robot-collected data and human demonstrations. This approach allows the model to learn from a broader range of experiences, enhancing its adaptability and robustness. Notably, performance on the complex ‘Spice’ task saw a dramatic increase, jumping from a 32% success rate to 71% when incorporating human data into the training regimen; this exemplifies how leveraging human insight can bridge the gap between robotic capability and real-world task completion, paving the way for more versatile and intuitive robotic systems.
Evaluations revealed substantial performance gains on complex manipulation tasks through co-training, where the model learned from both robotic experience and human demonstrations. Notably, success rates on the Dresser task-requiring precise folding and placement of clothing-increased by 25% to 50% when incorporating human data into the training process. Similarly, the challenging Bussing task, involving navigating obstacles while carrying objects, saw improvements ranging from 53% to 63% with co-training, demonstrating the model’s enhanced ability to adapt to nuanced scenarios and execute intricate procedures with greater reliability.
The VLA model demonstrated a significant performance boost on the Egg Sorting task through the incorporation of human data during training. Initial accuracy stood at 57%, indicating a considerable challenge in reliably identifying and sorting eggs. However, by leveraging insights gained from human demonstrations and corrections, the model’s ability to successfully complete the task improved dramatically, reaching an accuracy of 78%. This nearly 21 percentage point gain underscores the effectiveness of co-training and highlights how human guidance can effectively address nuanced challenges in robotic manipulation, enabling the VLA model to learn and adapt to complex real-world scenarios.
The demonstrated gains in generalization across diverse robotic tasks – from the nuanced manipulation required for the Dresser task to the precise object handling in Egg Sorting and the complex sequence of actions in Bussing – suggest a significant leap toward building robots capable of operating reliably in unstructured, real-world environments. The VLA model doesn’t simply memorize training scenarios; it learns underlying principles of interaction, enabling it to adapt to novel situations and object configurations with remarkable efficiency. This adaptability, bolstered by the synergistic use of both robotic and human data during training, points to a future where robots are less brittle and more readily deployed to assist with a wider range of everyday tasks, ultimately realizing the promise of truly versatile robotic assistants.

The study illuminates a critical juncture in embodied AI: the efficacy of transfer learning hinges not merely on the quantity of human data, but on the robustness of robotic pre-training. Diverse pre-training acts as a necessary filter, enabling robots to interpret and utilize human demonstrations effectively. This echoes Ken Thompson’s sentiment: “There’s no reason to have a complicated solution when a simple one works.” The research demonstrates that a complex model fed with inadequate robotic experience yields limited results; simplicity in foundational robotic understanding unlocks the potential of human-derived knowledge. The core concept of diverse pre-training, therefore, isn’t simply about scaling data; it’s about refining the foundation upon which transfer can occur, achieving efficiency through clarity.
What Lies Ahead?
This work clarifies a simple truth: robots learn from humans best after first learning about robots. Abstractions age, principles don’t. The finding isn’t revolutionary, yet its demonstration is valuable. It highlights the necessity of diverse robotic pretraining-a foundational step often overlooked in the rush to leverage large human datasets. Every complexity needs an alibi.
Remaining questions are stark. How much diversity is enough? Can this pretraining be automated, shifting the burden from curated datasets to continuous self-supervised learning? And critically, what constitutes ‘success’ for embodied AI? Current metrics focus on task completion. A more nuanced evaluation must address safety, adaptability, and genuine understanding-qualities not easily quantified.
The path forward isn’t about building bigger models. It’s about building smarter ones. Models that prioritize fundamental principles over superficial complexity. The focus must shift from simply transferring knowledge to grounding it. Robots must learn to see, act, and understand within their own physical reality, before they can meaningfully interact with ours.
Original article: https://arxiv.org/pdf/2512.22414.pdf
Contact the author: https://www.linkedin.com/in/avetisyan/
See also:
- Clash Royale Best Boss Bandit Champion decks
- Vampire’s Fall 2 redeem codes and how to use them (June 2025)
- Clash Royale Furnace Evolution best decks guide
- Best Hero Card Decks in Clash Royale
- Mobile Legends: Bang Bang (MLBB) Sora Guide: Best Build, Emblem and Gameplay Tips
- Best Arena 9 Decks in Clast Royale
- Clash Royale Witch Evolution best decks guide
- Wuthering Waves Mornye Build Guide
- Dawn Watch: Survival gift codes and how to use them (October 2025)
- Brawl Stars December 2025 Brawl Talk: Two New Brawlers, Buffie, Vault, New Skins, Game Modes, and more
2025-12-31 00:47