Dexterous Robots Learn to Open Anything

Author: Denis Avetisyan


Researchers have developed a new framework enabling legged robots to autonomously manipulate and open a wide variety of complex, previously unseen articulated objects.

A legged manipulator demonstrates the capacity to open diverse articulated objects-including cabinets with revolute mechanisms and drawers with prismatic slides-without reliance on pre-existing object-specific models, suggesting a pathway toward generalized robotic manipulation in unstructured environments.
A legged manipulator demonstrates the capacity to open diverse articulated objects-including cabinets with revolute mechanisms and drawers with prismatic slides-without reliance on pre-existing object-specific models, suggesting a pathway toward generalized robotic manipulation in unstructured environments.

The OpenHEART system combines proprioceptive feedback with exteroceptive sensing and a learned object representation to achieve sample-efficient manipulation of heterogeneous articulated objects.

While legged robots offer promising mobility for manipulation, robustly interacting with diverse articulated objects remains a significant challenge due to complex dynamics and varying object types. This paper introduces ‘OpenHEART: Opening Heterogeneous Articulated Objects with a Legged Manipulator’, a framework designed to enable autonomous opening of previously unseen objects by learning compact, low-dimensional representations of handle and panel geometry alongside an articulation estimator that adaptively fuses proprioceptive and exteroceptive feedback. The proposed Sampling-based Abstracted Feature Extraction (SAFE) and Articulation Information Estimator (ArtIEst) improve sample efficiency and generalization across heterogeneous objects. Could this approach pave the way for more versatile and adaptable legged robots capable of complex in-hand manipulation tasks in unstructured environments?


The Inevitable Complexity of Interaction

Robotic systems designed for physical interaction with the world often face difficulty when handling objects that aren’t rigid – those with moving parts and complex geometries, termed ‘Heterogeneous Articulated Objects’. Unlike grasping a simple block, manipulating a tool like a wrench, a door handle, or even a piece of clothing presents a unique challenge because these objects possess internal degrees of freedom. Traditional robotic grasping and manipulation strategies, typically optimized for static, well-defined shapes, struggle with the infinite configurations and unpredictable movements inherent in these articulated items. This complexity arises not just from the object’s shape, but from the relationships between its parts, demanding a level of perceptive and adaptive control that surpasses the capabilities of many current robotic platforms. Successfully interacting with these objects requires a fundamental shift towards methods capable of understanding and accommodating dynamic, multi-part configurations.

Effective manipulation of heterogeneous articulated objects-those with diverse shapes and movable parts-hinges on a robot’s ability to accurately perceive ‘Articulation Information’. This encompasses not just identifying the presence of joints, but also determining their specific direction and the distance between them. A nuanced understanding of these parameters is crucial for planning successful grasps and movements; without it, a robot risks colliding with the object, applying insufficient force, or failing to maintain a stable hold. Consequently, research focuses on developing perception systems capable of extracting this critical articulation data from complex sensory input, enabling robots to interact with a far wider range of real-world objects and perform increasingly sophisticated tasks.

Robotic systems attempting to interact with the physical world are often overwhelmed by the sheer volume of incoming sensory information, a phenomenon known as the curse of dimensionality. Raw data from vision sensors, force-torque sensors, and joint encoders generates incredibly high-dimensional observations that are computationally expensive to process and difficult to interpret. This presents a substantial bottleneck for manipulation tasks, as efficiently representing the state of an object – its configuration and properties – requires distilling this complex data into a manageable and informative form. Current methodologies frequently struggle to achieve this efficient state representation without sacrificing crucial details, hindering a robot’s ability to reliably grasp, move, and manipulate objects in dynamic environments. Consequently, research focuses on developing techniques that can effectively reduce dimensionality while preserving the essential information needed for robust and adaptable robotic manipulation.

Simulation results show successful manipulation of heterogeneous articulated objects with varying handle shapes, dimensions, and opening directions.
Simulation results show successful manipulation of heterogeneous articulated objects with varying handle shapes, dimensions, and opening directions.

Adaptive Perception: Fusing Data for Articulation

The ArtIEst method addresses articulation estimation challenges by adaptively integrating data from exteroceptive sensors – specifically cameras and LiDAR – with proprioceptive data representing the robot’s internal state. This fusion isn’t a simple concatenation of data; rather, the system dynamically weights the contribution of each sensor modality based on contextual relevance and reliability. By combining external observations of the environment with the robot’s understanding of its own configuration, ArtIEst aims to create a more complete and accurate representation of object articulation than would be achievable using either exteroception or proprioception in isolation. This adaptive approach allows the system to leverage the strengths of both data sources, mitigating the weaknesses inherent in relying on a single sensory input.

The ArtIEst method achieves improved accuracy and robustness in articulation information estimation by adaptively fusing exteroceptive and proprioceptive data. Comparative ablation studies demonstrate a quantifiable reduction in estimation error both prior to and during physical contact with objects when utilizing the fused approach. Specifically, reliance on a single sensor modality-either external sensors or internal state alone-results in a higher margin of error than the combined estimation. This improvement indicates the fusion process effectively mitigates the limitations inherent in each individual data source, leading to a more reliable assessment of object articulation.

ArtIEst employs a low-dimensional representation of the object being manipulated, moving beyond processing raw sensory data directly. This is achieved by projecting high-dimensional sensor inputs – such as point clouds from LiDAR or images from cameras – into a lower-dimensional space defined by a limited set of features relevant to articulation estimation. This dimensionality reduction significantly improves computational efficiency by decreasing the number of parameters required for processing and storage. Furthermore, utilizing a lower-dimensional representation mitigates the risk of overfitting to noisy or incomplete sensor data, resulting in a more generalized and robust articulation estimation, particularly in dynamic or cluttered environments.

This hierarchical framework utilizes SAFE for efficient shape representation and ArtIEst to estimate articulation by adaptively fusing external and internal sensing, with a history encoder extracting relevant proprioceptive features.
This hierarchical framework utilizes SAFE for efficient shape representation and ArtIEst to estimate articulation by adaptively fusing external and internal sensing, with a history encoder extracting relevant proprioceptive features.

Abstraction and Robustness Through SAFE Representation

The SAFE method builds upon low-dimensional representation techniques by explicitly addressing the abstraction of object shapes. Traditional low-dimensional representations often lack a systematic approach to generalization across variations in object appearance or pose. SAFE introduces a framework that learns these abstractions through a distributional approach, aiming to capture the essential features defining an object’s shape while disregarding irrelevant details. This is achieved by learning a latent space where similar shapes are clustered together, enabling the system to effectively represent and manipulate objects even with incomplete or noisy sensory data. The core principle involves mapping high-dimensional sensory inputs to a lower-dimensional latent space that prioritizes shape characteristics, thereby improving robustness and reducing the risk of overfitting to specific instances.

The SAFE method employs Kullback-Leibler (KL) Divergence to minimize the distributional discrepancy between the learned representation and a prior distribution, thereby mitigating overfitting. KL Divergence, measured as [latex]D_{KL}(P||Q) = \in t P(x) log \frac{P(x)}{Q(x)} dx[/latex], quantifies the information lost when [latex]Q[/latex] is used to approximate probability distribution [latex]P[/latex]. By minimizing this divergence during training, SAFE encourages the system to learn representations that are closer to the prior, reducing sensitivity to noise and irrelevant details in the training data. This regularization technique improves the system’s ability to generalize to unseen object instances and variations, enhancing its robustness and manipulation performance in real-world scenarios.

Effective robotic manipulation requires systems to generalize across instances of the same object despite inherent variations in shape, pose, and appearance. The ability to abstract object characteristics – separating essential features from accidental details – is therefore critical for robust performance in unstructured environments. Without abstraction, a robot trained on a specific instance of an object may fail when presented with a slightly different one. By focusing on the underlying, consistent properties of an object category, the system can achieve reliable grasping and manipulation even when faced with significant real-world variability, improving overall system resilience and adaptability.

Exteroception-based estimation exhibits errors due to visual ambiguity in object opening direction (rightward vs. upward) which are mitigated by augmenting with proprioceptive data, as demonstrated by the improved estimation accuracy when visual ambiguity is present or absent.
Exteroception-based estimation exhibits errors due to visual ambiguity in object opening direction (rightward vs. upward) which are mitigated by augmenting with proprioceptive data, as demonstrated by the improved estimation accuracy when visual ambiguity is present or absent.

Hierarchical Control: Orchestrating Complex Manipulation

The system utilizes a hierarchical control framework to manage complex manipulation tasks. This framework decomposes the control problem into two distinct levels: a high-level planner and a low-level controller. The high-level planner is responsible for generating a sequence of waypoints or sub-goals, defining the overall strategy for the manipulation. The low-level controller then executes these sub-goals by directly controlling the robot’s actuators, focusing on precise movement and force control to achieve each individual objective. This separation of concerns allows for more efficient and robust manipulation, as the high-level planner can reason about long-term goals while the low-level controller handles the immediate physical interactions.

The high-level planner utilizes a history encoder to integrate temporal information into its decision-making process. This encoder processes a sequence of past robot states – including joint angles, end-effector positions, and object interactions – and compresses this data into a fixed-length vector representation. This vector, effectively a learned summary of the recent past, is then fed as input to the planner alongside the current state. By considering historical context, the planner can anticipate the consequences of actions more accurately, improve long-term strategy, and perform more robustly in dynamic environments where immediate sensor data may be insufficient for optimal control.

The robotic manipulation framework utilizes Reinforcement Learning (RL) for training, specifically employing the Proximal Policy Optimization (PPO) algorithm. PPO is an on-policy algorithm that iteratively refines the robot’s manipulation policies through trial-and-error interaction with the environment. This approach allows the system to learn optimal strategies by maximizing cumulative rewards received for successful task completion. The PPO implementation incorporates techniques to ensure stable policy updates, preventing drastic changes that could hinder learning and maintaining consistent performance during training. Through this process, the robot learns to adapt its actions to achieve complex manipulation goals.

Despite an initial unstable grasp, the robot autonomously adjusted to retrieve the drawer handle and successfully open the drawer, demonstrating robust manipulation in a real-world scenario.
Despite an initial unstable grasp, the robot autonomously adjusted to retrieve the drawer handle and successfully open the drawer, demonstrating robust manipulation in a real-world scenario.

Towards Robust Robotic Manipulation: A New Standard

A novel robotic system, termed the ‘Legged Manipulator,’ demonstrates an unprecedented ability to reliably interact with complex objects. This capability stems from the synergistic integration of three key techniques: SAFE abstraction, hierarchical control, and reinforcement learning. SAFE abstraction simplifies the perception of complex environments, allowing the robot to focus on essential information. Hierarchical control breaks down manipulation tasks into manageable sub-goals, improving efficiency and robustness. Finally, reinforcement learning allows the system to adapt and refine its strategies through trial and error. The resulting system achieves a remarkable 99.35% success rate when tested on ‘Heterogeneous Articulated Objects’ – items with varying shapes and moving parts – indicating a significant leap forward in robotic manipulation and a substantial improvement over existing approaches.

Traditional robotic manipulation systems often struggle when confronted with the intricacies of real-world objects and environments. These systems frequently falter due to their inability to effectively process the vast amount of sensory information – the high-dimensional observations – and to accurately model the complex, often irregular, geometries of the objects they attempt to manipulate. This research addresses these limitations through a novel approach that demonstrably outperforms existing methods. Rigorous testing, including comparisons to baseline and ablation studies, reveals a significantly improved success rate – currently reported at 99.35% – in grasping and manipulating ‘Heterogeneous Articulated Objects’. This achievement highlights the system’s robust ability to navigate complex scenarios and represents a substantial step towards more adaptable and reliable robotic manipulation capabilities.

The development of this legged manipulator signifies a substantial leap towards robotic systems capable of genuine environmental adaptability. Current robotic manipulation often falters when confronted with the unpredictable nature of real-world settings, requiring highly curated conditions for even simple tasks. However, this research demonstrates a pathway to overcome these limitations, enabling robots to reliably interact with diverse objects – even those with intricate mechanisms – within cluttered and dynamic spaces. The implications extend beyond mere automation; it suggests a future where robots can autonomously perform complex tasks in unstructured environments, such as disaster relief, in-home assistance, or remote exploration, ultimately broadening the scope of robotic applications and increasing their utility in tackling previously insurmountable challenges.

Our policy outperforms baseline methods in learning an opening reward, as demonstrated by the learning curve and confirmed by saliency maps for both revolute and prismatic joints which highlight the policy's focus on relevant object features.
Our policy outperforms baseline methods in learning an opening reward, as demonstrated by the learning curve and confirmed by saliency maps for both revolute and prismatic joints which highlight the policy’s focus on relevant object features.

The pursuit of autonomous manipulation, as demonstrated by OpenHEART, inherently involves navigating systems prone to eventual complexity. This framework, with its emphasis on low-dimensional object representation and articulation estimation, attempts to impose order on inherent disorder. As Carl Friedrich Gauss observed, “If other objects have been described by it, it may be fairly asserted that the law extends to all objects.” The paper’s approach to generalizing across heterogeneous articulated objects echoes this sentiment – seeking underlying principles applicable to a wider range of scenarios. While OpenHEART focuses on sample efficiency in the present, the accumulation of complexity in representing diverse objects presents a future cost. The system’s ‘memory’ – its ability to generalize – is constantly tested and refined, a process mirroring the inevitable decay and adaptation of all complex systems.

What Lies Ahead?

The framework detailed within represents a predictable, yet valuable, step toward robotic manipulation of the truly complex. The ability to estimate articulation, however reliant on current sensor suites, merely postpones the inevitable confrontation with perceptual ambiguity. Every bug encountered during deployment will be a moment of truth in the timeline, revealing the limits of the low-dimensional object representation. The system functions, but the elegance of its current state will degrade as it encounters objects designed by evolution, not engineers.

The pursuit of sample efficiency, while laudable, addresses a symptom, not the disease. Technical debt, in the form of simplified object models and reliance on specific sensor modalities, is the past’s mortgage paid by the present. Future work must confront the fundamental problem of generalization – not by collecting more data, but by embracing methods that allow the system to learn how little it truly knows. A shift toward actively seeking out and modeling uncertainty may prove more fruitful than striving for ever-more-accurate estimations.

Ultimately, the longevity of this approach, like all systems, will be determined not by its initial successes, but by its capacity to age gracefully. The question isn’t whether the system can open these objects, but whether it can adapt to the infinite variety of objects yet to be conceived – or discovered.


Original article: https://arxiv.org/pdf/2603.05830.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-03-09 18:33