Author: Denis Avetisyan
Researchers have developed a new representation learning framework that allows robots to better understand and interact with objects, leading to more reliable and adaptable manipulation skills.

DexRepNet++ leverages geometric and spatial features of hand-object interactions to improve generalization in both simulation and real-world robotic manipulation tasks.
Despite advances in robotic dexterity, generalizing manipulation policies remains challenging due to the complexities of hand-object interaction and high-dimensional action spaces. This work, ‘DexRepNet++: Learning Dexterous Robotic Manipulation with Geometric and Spatial Hand-Object Representations’, addresses this limitation by introducing DexRep, a novel representation that captures crucial geometric and spatial features of hand-object contact. Experimental results demonstrate that DexRep significantly improves performance across grasping, in-hand reorientation, and bimanual handover tasks, achieving state-of-the-art success rates in both simulation and real-world deployments with minimal sim-to-real transfer gap. Could this representation learning approach unlock more robust and adaptable robotic manipulation capabilities in unstructured environments?
Bridging the Reality Gap: The Core of Robotic Dexterity
Despite considerable progress in robotic grasping algorithms and hardware, a persistent disparity exists between performance in simulated environments and real-world applications – a challenge known as the ‘Sim-to-Real Gap’. This gap arises because simulations, while offering controlled conditions for development and testing, inevitably fail to fully capture the complexities of unstructured environments. Factors like imperfect lighting, variations in object textures and shapes, and unpredictable disturbances all contribute to discrepancies between simulated and actual robotic performance. Consequently, grasping strategies that function flawlessly in simulation often falter when deployed in the messiness of a real-world setting, limiting the ability of robots to reliably manipulate objects and perform tasks autonomously. Bridging this gap is therefore paramount to unlocking the full potential of robotic manipulation and enabling broader deployment in fields like manufacturing, logistics, and even domestic assistance.
Robotic manipulation often relies on sensor data to perceive and interact with objects, but real-world environments introduce significant challenges to this process. Traditional robotic systems frequently encounter ‘partial point clouds’ – incomplete or noisy data sets arising from occlusions, sensor limitations, or dynamic lighting conditions. These incomplete representations of the environment create inherent uncertainty, making it difficult for robots to accurately identify grasp points, predict object behavior, and execute precise movements. Consequently, algorithms developed in controlled simulations often fail when deployed in the messiness of the real world, as they haven’t been trained to cope with the ambiguity present in these imperfect sensory inputs. Overcoming this challenge necessitates the development of robust algorithms capable of inferring complete information from fragmented data and adapting to the inherent uncertainties of real-world perception.
The reliable execution of complex robotic tasks, such as bimanual handover – the coordinated transfer of an object between a robot and a human – fundamentally depends on overcoming the persistent ‘Sim-to-Real Gap’. This gap represents the discrepancy between performance in simulated environments and the unpredictable conditions of the real world. Successfully bridging this divide isn’t merely about incremental improvements in grasping algorithms; it necessitates a robust system capable of adapting to real-world uncertainties like incomplete sensor data and unexpected disturbances. Only when robots can consistently and safely perform bimanual handovers – a task demanding precise coordination, force control, and adaptability – will they truly become collaborative partners in complex human-robot interactions, unlocking applications in manufacturing, healthcare, and assistive robotics.

DexRep: A Comprehensive Representation of Hand-Object Interaction
DexRep is a hand-object interaction representation built upon three core features designed to comprehensively characterize the interaction state. The ‘Occupancy’ feature provides a volumetric understanding of the object’s overall shape, indicating space occupied. Complementing this is the ‘Surface’ feature, which focuses on precise contact points between the hand and the object. Finally, DexRep incorporates ‘Local-Geo’ features, capturing detailed geometric properties at the point of interaction. These three features, when combined, provide a robust and informative representation of the hand-object relationship for robotic manipulation tasks.
The DexRep representation utilizes two primary features for characterizing hand-object interaction: ‘Occupancy’ and ‘Surface’. The ‘Occupancy Feature’ provides a coarse, volumetric understanding of the object’s overall shape, effectively establishing a global context for manipulation. Complementing this is the ‘Surface Feature’, which encodes precise contact information between the hand and the object. This includes detailed data on the points of contact, normal vectors, and distances, providing a localized representation of the interaction. Combined, these features offer both a broad understanding of the object and a precise account of the hand’s engagement with it.
DexRep utilizes a pretrained PointNet architecture to generate ‘Local-Geo Features’, which capture detailed geometric information essential for dexterous manipulation. PointNet, a deep learning model designed for point cloud data, processes local regions of the object in contact with the hand to extract these features. This process focuses on subtle geometric variations – such as edges, corners, and surface curvature – that are critical for precise grip control and manipulation planning. The use of a pretrained PointNet allows DexRep to leverage existing knowledge of geometric feature extraction, improving performance and reducing the need for extensive training data specifically for hand-object interaction tasks. These local geometric details, encoded as feature vectors, provide a nuanced understanding of the object’s shape at the point of contact, enabling more robust and adaptable manipulation strategies.
DexRep’s comprehensive feature set – encompassing global occupancy, precise surface contact, and fine-grained local geometry – facilitates robust policy learning under conditions of imperfect sensor data. The representation’s ability to encode both high-level shape understanding and detailed geometric features allows for effective generalization even when sensor input is noisy, incomplete, or subject to occlusion. This resilience stems from DexRep’s capacity to infer likely hand-object interactions based on partial observations, enabling continued performance despite sensor limitations. Consequently, learned policies built upon DexRep demonstrate improved stability and adaptability in real-world scenarios where perfect sensor data is rarely available.
![DexRep's effectiveness was analyzed using a diverse set of 40 objects, comprising the GRAB dataset [50] and the 3DNet dataset [61].](https://arxiv.org/html/2602.21811v1/x3.png)
Reinforcement Learning with DexRep: Demonstrable Gains in Performance
Reinforcement learning policies utilizing DexRep demonstrate a substantial performance increase in real-world grasping tasks. Comparative analysis reveals a 30.8% improvement in grasping success rate when compared to state-of-the-art methodologies, including UniDexGrasp, ILAD, and DAPG. This performance gain indicates DexRep’s efficacy in enabling robots to reliably grasp objects in unstructured environments. The evaluation was conducted using standardized benchmarks and metrics to ensure a fair and accurate comparison of grasping performance across all tested algorithms.
DexRep demonstrates a grasping success rate of 85.0% due to its capacity to process incomplete sensory data in the form of partial point clouds. Unlike systems requiring complete object representations, DexRep’s architecture allows it to infer object affordances and plan successful grasps even with limited or occluded visual input. This robustness to data incompleteness facilitates generalization across variations in object pose, lighting conditions, and environmental disturbances, contributing significantly to its improved performance compared to methods reliant on complete data sets.
The DAgger (Dataset Aggregation) algorithm was implemented to iteratively refine the reinforcement learning policies developed with DexRep. This involved executing the current policy in the real environment, collecting the resulting state-action pairs, and then adding these samples to the training dataset. The model was subsequently retrained on this expanded dataset, effectively learning from its own interactions with the environment and correcting for any discrepancies between the simulation and real-world conditions. This iterative process of execution, data collection, and retraining significantly improved the robustness and adaptability of the learned policies, leading to enhanced performance in grasping tasks and reducing the impact of environmental disturbances.
Domain randomization was implemented during DexRep training to improve policy resilience to real-world disturbances. This technique involved varying simulation parameters such as lighting, textures, and object poses during training, forcing the policy to learn features robust to these changes. Evaluation demonstrated a 5% reduction in the sim-to-real gap for DexRep, indicating improved performance when transferring learned policies to real-world scenarios. In comparison, UniDexGrasp++ experienced an 18% performance drop when deployed in the real world, highlighting the effectiveness of domain randomization within the DexRep framework.

Towards a Future of Adaptive and Robust Robotic Dexterity
DexRep represents a substantial advancement in robotic manipulation by directly addressing the persistent challenge of transferring skills learned in simulated environments to the complexities of the real world – a problem commonly known as the ‘Sim-to-Real Gap’. Traditional robotic systems often struggle with even slight variations in object properties, lighting, or positioning, limiting their effectiveness outside carefully controlled settings. This new approach creates a more adaptable system through a robust representation of hand-object interaction, allowing robots to generalize their learned behaviors to unseen objects and unpredictable conditions. Consequently, robots equipped with DexRep demonstrate improved performance in unstructured environments, opening doors to more versatile applications in fields requiring dexterity and adaptability, such as automated assembly, flexible manufacturing, and even complex surgical procedures.
DexRep establishes a novel framework for robotic manipulation by focusing on a detailed and resilient understanding of how hands interact with objects. This representation goes beyond simple object recognition, instead modeling the complex interplay of forces, contacts, and geometries that define successful grasping and manipulation. Consequently, robots equipped with DexRep can approach tasks previously considered too delicate or unpredictable – such as assembling intricate components, skillfully wielding tools, or dynamically re-orienting objects within the hand – with significantly improved reliability. The system’s capacity to accurately model these interactions allows for more robust planning and control, enabling robots to adapt to variations in object shape, pose, and even environmental disturbances, ultimately broadening the scope of tasks achievable in real-world settings.
Demonstrating a substantial advancement in robotic manipulation, the DexRep system achieves remarkably high success rates across a range of challenging tasks. Rigorous testing on the 3DNet dataset reveals a 96.6% success rate in grasping previously unseen objects, indicating strong generalization capabilities. Beyond initial acquisition, DexRep also excels in more complex manipulations; the system achieves 76.3% success in reorienting objects within the robot’s grasp and a 72.6% success rate in performing handovers, even when relying on incomplete sensor data – specifically, partial point clouds. These results highlight DexRep’s robustness and potential for deployment in real-world scenarios where complete information is rarely available, paving the way for more adaptable and reliable robotic systems.
The advent of robust robotic dexterity, as exemplified by technologies like DexRep, promises a substantial reshaping of several key industries. Manufacturing stands to benefit from automated assembly lines capable of handling diverse and delicate parts with greater precision and adaptability, reducing errors and increasing throughput. In logistics, robots equipped with advanced grasping capabilities could efficiently sort, pack, and transport goods, addressing labor shortages and streamlining supply chains. Perhaps most profoundly, the healthcare sector anticipates a revolution in assistive robotics, where robots could aid surgeons with intricate procedures, deliver medications and supplies within hospitals, and provide personalized care to patients in their homes. Furthermore, individuals with limited mobility could regain independence through robotic assistance with daily tasks, fostering a future where technology empowers a higher quality of life for all.
![Hand posture alignment is achieved by optimizing key vectors-including finger-to-finger, finger-to-wrist, and finger-to-object-between a human hand model ([latex]MANO[/latex]) and a robotic hand ([latex]Adroit[/latex]).](https://arxiv.org/html/2602.21811v1/fig/retarget.png)
The pursuit of robust robotic manipulation, as demonstrated by DexRepNet++, often leads to intricate architectures. One observes a tendency to layer complexity upon complexity, ostensibly to address every conceivable contingency. Yet, the elegance of DexRepNet++ lies in its distillation of hand-object interaction into fundamental geometric and spatial features. This echoes Blaise Pascal’s sentiment: “The eloquence of simplicity is a sign of maturity.” The researchers didn’t seek to model all of reality, but rather to extract the essential information – occupancy, surface normals, and local geometry – needed for reliable grasping and manipulation. It’s a reminder that true progress isn’t always about adding more, but about discerning what can be removed without sacrificing performance; a principle neatly embodied in DexRepNet++’s focus on core representations.
Where to Now?
The presented work, while demonstrably effective in constructing a functional representation for hand-object interaction, merely clarifies the boundaries of the problem. It does not, of course, solve it. The reliance on geometric and spatial features, however elegantly combined, introduces a fragility inherent in any system predicated on precise measurement. Real-world entropy will always exceed the fidelity of the sensor. Future work must address not the refinement of representation, but the graceful degradation of performance as that representation inevitably fails.
A critical, and largely unaddressed, limitation lies in the implicit assumption of static object properties. The world is not composed of Platonic solids awaiting manipulation. The challenge, then, shifts from representing the object to predicting its deformation under force. This demands a move beyond purely geometric descriptions toward models incorporating material properties and dynamic response. Such complexity, naturally, invites further compression-the art lies in discerning what truly constitutes noise.
Ultimately, the pursuit of ‘generalization’ feels a misnomer. Total generality is an asymptotic ideal, forever receding. A more pragmatic goal is ‘robustness’ – the ability to maintain functionality despite the inevitable imperfections of both perception and execution. The field should prioritize systems that learn to ignore irrelevant information, rather than attempting to capture it all. Beauty, after all, resides not in completeness, but in efficient omission.
Original article: https://arxiv.org/pdf/2602.21811.pdf
Contact the author: https://www.linkedin.com/in/avetisyan/
See also:
- Clash of Clans Unleash the Duke Community Event for March 2026: Details, How to Progress, Rewards and more
- Brawl Stars February 2026 Brawl Talk: 100th Brawler, New Game Modes, Buffies, Trophy System, Skins, and more
- Gold Rate Forecast
- Kylie Jenner squirms at ‘awkward’ BAFTA host Alan Cummings’ innuendo-packed joke about ‘getting her gums around a Jammie Dodger’ while dishing out ‘very British snacks’
- Jason Statham’s Action Movie Flop Becomes Instant Netflix Hit In The United States
- Free Fire Beat Carnival event goes live with DJ Alok collab, rewards, themed battlefield changes, and more
- eFootball 2026 Jürgen Klopp Manager Guide: Best formations, instructions, and tactics
- Brent Oil Forecast
- Hailey Bieber talks motherhood, baby Jack, and future kids with Justin Bieber
- eFootball 2026 Show Time Worldwide Selection Contract: Best player to choose and Tier List
2026-02-26 16:33