Catching on: Robots Learn Agile Manipulation from Pixels Alone

Author: Denis Avetisyan


A new framework enables robots to reliably catch dynamic objects using only visual information from a single camera, bypassing the need for complex 3D sensing.

The deployed policy successfully integrates real-world visual input-specifically, segmented object geometries derived from [latex]SAM2[/latex] analysis of camera feeds-to execute catching sequences, demonstrating an ability to process complex scenes and actuate appropriate responses.
The deployed policy successfully integrates real-world visual input-specifically, segmented object geometries derived from [latex]SAM2[/latex] analysis of camera feeds-to execute catching sequences, demonstrating an ability to process complex scenes and actuate appropriate responses.

This work presents a multi-agent reinforcement learning approach to robotic manipulation that achieves robust performance through pixel-level visual features and sim-to-real transfer.

Achieving robust robotic manipulation in dynamic scenarios remains challenging due to the complexities of visual perception and control. This paper, ‘Pixel2Catch: Multi-Agent Sim-to-Real Transfer for Agile Manipulation with a Single RGB Camera’, introduces a novel framework for catching thrown objects that bypasses explicit 3D pose estimation by leveraging pixel-level visual features. A heterogeneous multi-agent reinforcement learning approach is employed, treating the robotic arm and hand as cooperative agents, enabling stable learning and successful transfer from simulation to real-world environments. Could this pixel-based, multi-agent paradigm unlock more adaptable and efficient robotic systems for a wider range of manipulation tasks?


The Illusion of Control: Why Catching is Hard

Conventional robotic grasping systems face significant hurdles when attempting to intercept and secure objects in motion, a task demanding not only speed but also a remarkable capacity for real-time adaptation. Unlike static grasping, where a robot can carefully position its end-effector around a stationary object, dynamic catching requires anticipating the object’s trajectory, accounting for its velocity, and adjusting the grip during the interaction. This presents a complex control problem; even slight errors in timing or positioning can lead to a failed catch. The challenge lies in the inherent unpredictability of real-world motion, where objects rarely follow perfectly prescribed paths, and external disturbances are commonplace. Consequently, traditional methods, often reliant on pre-programmed movements or limited sensory input, struggle to reliably execute successful intercepts, highlighting the need for more sophisticated perception and control strategies.

Current robotic catching systems frequently falter when faced with real-world variability because they depend heavily on meticulously planned movements or insufficient environmental awareness. These approaches presume a predictable interaction, creating rigid behaviors that struggle with deviations in object speed, trajectory, or even minor disturbances. A robot programmed with a fixed catching motion will likely fail if the object arrives a fraction of a second early or late, or if it veers slightly off course. Similarly, systems relying on limited sensor input – perhaps only detecting the object’s general location – lack the nuanced information necessary to adjust the grasp in real-time. This dependence on pre-defined actions or incomplete data significantly restricts a robot’s ability to generalize its catching skills to novel, unpredictable scenarios, highlighting the need for more adaptable and perceptive robotic manipulation strategies.

Effective interception of moving objects by robotic systems hinges on the seamless integration of two critical capabilities: robust perception and agile control. A system must not only accurately perceive the trajectory, velocity, and other relevant characteristics of a dynamic target – often under conditions of visual occlusion or rapid change – but also translate this information into precisely timed and coordinated movements. This requires advanced sensing modalities, such as high-speed vision or force-torque sensors, coupled with control algorithms capable of rapidly adjusting to unforeseen perturbations or deviations from predicted paths. The challenge lies in minimizing the latency between perception and action, effectively creating a feedback loop that allows the robot to continuously refine its movements and successfully intercept the object, demanding a level of dexterity and responsiveness previously unattainable in robotic manipulation.

The persistent difficulty in robotic catching stems from a fundamental disconnect between how robots perceive the world and how they react to it. Traditional robotic manipulation prioritizes precise, pre-planned movements, ill-suited for the unpredictable nature of intercepting a moving object. A novel approach demands integration – a system where visual data concerning an object’s trajectory instantly informs and adjusts the robot’s motor controls. This necessitates algorithms capable of real-time analysis, predictive modeling, and agile adaptation, effectively creating a closed-loop system where perception and action are seamlessly intertwined. Such a development moves beyond simple grasping towards a dynamic, responsive manipulation – allowing robots to not merely hold objects, but to actively engage with them in motion, mirroring the dexterity observed in biological systems.

Training, validation, and real-world experiments utilize a diverse set of objects, and simulated trajectories demonstrate realistic object dynamics independent of robot actions.
Training, validation, and real-world experiments utilize a diverse set of objects, and simulated trajectories demonstrate realistic object dynamics independent of robot actions.

Deconstructing the Task: A Multi-Agent Approach

Pixel2Catch utilizes a multi-agent reinforcement learning (MARL) framework where the overall catching task is divided among multiple autonomous agents, each trained to perform a specific sub-task. This decomposition improves efficiency by allowing agents to specialize, reducing the complexity of the learning problem for each individual agent. Agents do not operate independently; instead, they learn to coordinate their actions through shared experience and reward signals, effectively distributing the workload and enabling more robust and scalable performance compared to a single, monolithic agent attempting the entire task. The MARL approach facilitates parallel learning and allows for the exploration of diverse strategies, ultimately leading to a more adaptable and efficient catching system.

Pixel2Catch utilizes raw RGB camera images as the primary sensory input, circumventing the need for pre-defined object representations. Visual features are extracted directly at the pixel level using convolutional neural networks, creating a high-dimensional feature space that captures detailed information about the observed scene. This approach provides a rich and readily available perceptual input, as RGB cameras are standard equipment on most robotic platforms and require minimal pre-processing. The pixel-level features represent edges, textures, and color variations, allowing the system to learn directly from visual data without relying on intermediate symbolic representations or complex state estimation techniques. This direct perception pathway facilitates adaptability to variations in object appearance and lighting conditions.

Pixel2Catch streamlines the perception process by eliminating the requirement for pre-defined object models or computationally expensive state estimation techniques. Traditional robotic systems often rely on detailed 3D models or complex algorithms to track object position, velocity, and orientation. This approach introduces significant development and computational overhead, and limits adaptability to novel objects. Pixel2Catch, conversely, operates directly on raw pixel data from RGB cameras, learning to infer necessary information implicitly through reinforcement learning. This simplification reduces the complexity of the perception pipeline, lowering both development time and computational costs, and enabling the system to operate with greater flexibility in dynamic environments.

Pixel2Catch’s capacity to learn directly from raw visual input – RGB camera images – enables adaptability to diverse object characteristics without requiring pre-programmed object models. The system achieves this by establishing correlations between visual features and successful catching actions through reinforcement learning. Consequently, Pixel2Catch demonstrates proficiency in intercepting objects exhibiting variations in geometry, dimensions, and kinematic profiles. This eliminates the need for explicit state estimation of object properties, allowing the system to generalize to previously unseen objects and trajectories without retraining or modification of internal parameters.

Object detection relies on pixel-level features extracted from bounding box coordinates-including corner and center positions, width, and height-and their temporal changes in both simulated and real-world environments.
Object detection relies on pixel-level features extracted from bounding box coordinates-including corner and center positions, width, and height-and their temporal changes in both simulated and real-world environments.

Seeing is Believing: Segmentation and Feature Extraction

Object segmentation within the system utilizes the Segment Anything Model 2 (SAM2) to define the boundaries of the target object as captured in the RGB image. SAM2 is a promptable segmentation model, meaning it requires minimal user input – typically a point or bounding box – to generate a high-quality segmentation mask. This mask then isolates the pixels belonging to the target object from the background, creating a precise representation of its shape and size for downstream processing. The model’s output is a binary mask where foreground pixels are assigned a value of 1 and background pixels are assigned a value of 0, facilitating subsequent feature extraction and robotic control.

Following object segmentation, pixel-level visual features are derived to quantify the isolated object’s characteristics. Specifically, the centroid of the segmented region is calculated to determine the object’s center coordinates within the image frame. Furthermore, bounding box dimensions are computed around the segmentation mask, providing precise measurements of the object’s width and height in pixels. These extracted features – center coordinates, width, and height – constitute a standardized, numerical representation of the object’s visual properties, facilitating its use as input for downstream processing tasks such as robotic control.

Extracted visual features, including object center coordinates, width, and height derived from image segmentation, are directly utilized as input data for multi-agent reinforcement learning (MARL) algorithms. These algorithms then process this data to generate control signals for both the robotic arm and hand. The MARL framework allows for coordinated movement, enabling the system to calculate the optimal trajectory for the arm and the precise grip configuration for the hand. This data-driven approach facilitates real-time adjustments based on the object’s position and dimensions, resulting in improved accuracy and adaptability during the catching process. The use of feature-based input eliminates the need for pre-programmed trajectories, allowing the robotic system to learn and refine its movements through interaction and feedback.

The implementation of advanced segmentation techniques directly addresses challenges inherent in robotic catching scenarios, specifically improving performance in variable lighting conditions and with partially occluded objects. By precisely delineating object boundaries at the pixel level, segmentation reduces the impact of background noise and visual clutter on feature extraction. This results in more reliable object state estimation – including position, orientation, and velocity – which are critical inputs for trajectory planning and control. Consequently, the robotic system demonstrates increased success rates in catching attempts and maintains consistent performance across a wider range of environmental and object variations, enhancing overall system robustness.

A collaborative robotic system utilizes two policies [latex]\pi_{arm}[/latex] and [latex]\pi_{hand}[/latex] operating on consecutive observations from a single overhead RGB camera (positioned 0.5 m behind and 2.2 m above) to collaboratively catch thrown objects, with privileged information used solely for value network training.
A collaborative robotic system utilizes two policies [latex]\pi_{arm}[/latex] and [latex]\pi_{hand}[/latex] operating on consecutive observations from a single overhead RGB camera (positioned 0.5 m behind and 2.2 m above) to collaboratively catch thrown objects, with privileged information used solely for value network training.

The Illusion of Reality: Bridging the Simulation Gap

Training robotic policies within simulated environments offers a compelling pathway to accelerate development and mitigate risks associated with real-world experimentation. This approach allows for rapid iteration and exploration of a vast design space without the constraints – or potential damage – inherent in physical testing. Virtual environments enable researchers to systematically manipulate variables, collect extensive datasets, and refine algorithms at a fraction of the time and cost. Furthermore, simulation provides access to ground truth data – precise measurements of position, velocity, and forces – that are often difficult or impossible to obtain in the real world, ultimately leading to more efficient and reliable learning processes. This foundation of virtual training is crucial for bridging the gap between simulated success and robust real-world performance.

To overcome the limitations of training in meticulously crafted, yet ultimately unrealistic, simulations, researchers employ Domain Randomization. This technique deliberately introduces a wide range of randomized variations within the simulation itself. Parameters such as the color and texture of objects, the intensity and direction of lighting, and even the physics governing object interactions – like friction and mass – are altered randomly during each training episode. This forces the learning agent to develop policies that are not overly specialized to any particular simulated condition, but rather generalize effectively across a distribution of possibilities. By experiencing a constantly shifting virtual world, the agent learns to focus on the essential features of the task, becoming remarkably resilient to the discrepancies inherent when deployed in the complexities of the real world.

The core of successful sim-to-real transfer lies in cultivating policies resilient to environmental discrepancies. By intentionally varying simulation parameters – encompassing factors like friction, mass, lighting conditions, and even the textures of objects – learning agents are compelled to generalize beyond the precise conditions of any single simulation. This deliberate introduction of ‘noise’ prevents the agent from over-fitting to the simulated world; instead, it encourages the development of strategies that function reliably across a spectrum of possibilities. Consequently, the resulting policies aren’t merely proficient within the simulation, but possess an inherent adaptability that enables consistent performance when deployed in the unpredictable conditions of the real world, a key factor in achieving markedly improved success rates compared to traditionally trained agents.

Real-world implementation of the system demonstrates a significant performance advantage over traditional single-agent reinforcement learning. Testing revealed a consistent 70% tracking rate, indicating the system’s ability to maintain accurate target identification throughout a task, coupled with an overall 50% success rate in completing the designated objective. This outcome represents a substantial improvement when contrasted with the baseline single-agent RL approach, which only achieved a 24% success rate under identical conditions. The marked difference highlights the efficacy of domain randomization in fostering robust policies capable of generalizing from simulated training to the complexities of a real-world environment, suggesting a viable pathway for deploying AI systems in dynamic and unpredictable settings.

Beyond the Catch: Towards Adaptive Robotic Systems

The Pixel2Catch framework establishes a crucial link between visual input and robotic control, but its true potential lies in combining this with system identification techniques. This pairing allows robots to move beyond pre-programmed motions and instead learn the dynamic properties of objects during interaction. By observing how an object responds to force – its mass, friction, and center of gravity – the robot builds an internal model enabling it to predict the outcome of different actions. This dynamic modeling capability is fundamental for grasping novel objects, adapting to changing conditions, and performing complex manipulation tasks with increased reliability and precision. Essentially, it transforms robotic arms from rigid, predictable machines into adaptable, learning systems capable of nuanced and versatile physical interaction.

Ongoing research endeavors are geared towards broadening the scope of robotic manipulation beyond simplified scenarios. Current efforts prioritize enabling robots to navigate the intricacies of complex object interactions, such as assembling multi-part objects or manipulating deformable items. Simultaneously, investigations are underway to improve robotic performance within dynamic environments-those characterized by moving obstacles, unpredictable disturbances, and real-time changes. This expansion necessitates advancements in both hardware and software, with a particular emphasis on robust perception algorithms capable of accurately tracking objects in motion and adaptive control strategies that allow robots to react effectively to unforeseen circumstances. Ultimately, the goal is to move beyond pre-programmed routines and cultivate a level of adaptability that allows robots to seamlessly integrate into and operate effectively within the complexities of the real world.

The confluence of sophisticated perception and machine learning holds substantial promise for revolutionizing robotic capabilities across critical sectors. Advanced perception systems, leveraging technologies like computer vision and tactile sensing, enable robots to gain a more nuanced understanding of their surroundings and the objects they manipulate. When paired with learning techniques – including reinforcement learning and imitation learning – robots can move beyond pre-programmed routines to adapt to novel situations and refine their performance over time. This synergy is poised to unlock new automation possibilities in manufacturing, where robots could handle increasingly complex assembly tasks; in logistics, where they could navigate dynamic warehouse environments and manage intricate picking and packing operations; and in healthcare, where they could assist surgeons with delicate procedures or provide personalized patient care. Ultimately, this integration aims to create robotic systems capable of not just executing commands, but intelligently responding to unforeseen challenges and optimizing performance in real-world scenarios.

The pursuit of genuinely adaptable robotic systems hinges on effectively uniting perception and action – moving beyond pre-programmed sequences to enable robots to understand their surroundings and respond intelligently. Current limitations often stem from a disconnect between what a robot sees and what it does; a robot may identify an object, but struggle to grasp or manipulate it successfully in a dynamic setting. Bridging this gap necessitates sophisticated algorithms that translate visual information into precise motor commands, accounting for factors like object weight, fragility, and environmental disturbances. Successful integration promises robots capable of tackling unpredictable real-world challenges – from assembling intricate products on a manufacturing line to assisting surgeons with delicate procedures, or even navigating complex logistics operations with greater efficiency and resilience.

The pursuit of robotic agility, as demonstrated by this framework, inherently demands a willingness to dismantle conventional approaches. The system eschews precise 3D modeling, opting instead for direct pixel-level interpretation – a bold rejection of established norms. This resonates with a core tenet of innovation: understanding limitations through deconstruction. As Tim Berners-Lee aptly stated, “The Web is more a social creation than a technical one.” Similarly, this research acknowledges that robust manipulation isn’t solely about technical precision, but about a system’s ability to adapt and learn from imperfect visual data, mirroring the organic, often messy, nature of real-world interaction. Each successful catch, therefore, becomes a testament to the beauty of imperfect systems working in harmony.

What Lies Beyond the Catch?

The elegance of bypassing explicit 3D reconstruction, as demonstrated in this work, should not be mistaken for a solved problem. It’s merely a shift in where the difficulty resides. The system trades geometric precision for a robust reliance on pixel-level features – a pragmatic move, certainly, but one that implicitly pushes the burden of environmental understanding onto the learning algorithm itself. Future iterations will inevitably probe the limits of this approach; how gracefully does it degrade with increasingly complex visual clutter, novel object appearances, or unpredictable lighting conditions? The true test isn’t a clean lab demo, but a chaotic real-world scenario.

Moreover, the multi-agent framework, while effective, presents a path ripe for further dissection. The current implementation treats agents as largely independent learners. However, a deeper investigation into inter-agent communication – not merely for task allocation, but for shared perceptual refinement – could yield substantial gains. Could agents collaboratively “teach” each other to interpret ambiguous visual cues, effectively building a more resilient and adaptable perceptual system? This raises a fascinating, if somewhat unsettling, question: at what point does this collective learning resemble something akin to emergent intelligence?

Ultimately, this work isn’t about catching objects; it’s about reverse-engineering the perceptual processes necessary for any agile manipulation. The ultimate goal isn’t a perfect catch rate, but a system that, when faced with an unforeseen situation, doesn’t simply fail, but actively seeks to understand-and then, adapt. That, after all, is the essence of intelligence – and a truly robust robotic system.


Original article: https://arxiv.org/pdf/2602.22733.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-02-28 18:49