Beyond Vision: Reconstructing Soft Robot Shapes with Touch

Author: Denis Avetisyan


A new method enables real-time 3D reconstruction of soft robots using only tactile sensing, eliminating the need for cameras and opening doors to more robust and adaptable robotic systems.

A camera-free framework reconstructs and renders deformable objects in real-time by translating resistance data from a flexible sensor array into cage node displacements via a Graph Attention Network, which then drives a low-dimensional cage representation to control dense surface deformation and ultimately updates Gaussian primitives for real-time rendering using Gaussian splatting.
A camera-free framework reconstructs and renders deformable objects in real-time by translating resistance data from a flexible sensor array into cage node displacements via a Graph Attention Network, which then drives a low-dimensional cage representation to control dense surface deformation and ultimately updates Gaussian primitives for real-time rendering using Gaussian splatting.

This work introduces a zero-shot learning framework leveraging flexible sensors and cage-based 3D Gaussian modeling for accurate and geometry-independent soft robot deformation reconstruction.

Reconstructing the 3D pose of soft robots remains challenging due to their inherent deformability and lack of rigid features. This is addressed in ‘Zero Shot Deformation Reconstruction for Soft Robots Using a Flexible Sensor Array and Cage Based 3D Gaussian Modeling’, which introduces a novel framework for real-time, camera-free 3D shape reconstruction. By integrating tactile sensing with a cage-based 3D Gaussian deformation model, the system achieves zero-shot generalization to previously unseen robot geometries and motions. Could this approach unlock more robust and adaptable control strategies for a wider range of soft robotic applications?


The Challenge of Representing Continuous Form

Conventional robotic modeling techniques, designed for rigid bodies with predictable movements, falter when applied to soft robots. These robots, constructed from compliant materials, exhibit continuous deformation – bending, stretching, and twisting – creating an infinite number of possible configurations. This contrasts sharply with the limited, discrete movements of traditional robots, which simplifies their mathematical representation. The high degree of freedom inherent in soft robotics-the ability to move in numerous, interconnected ways-further complicates the modeling process. Each degree of freedom introduces additional variables that must be accounted for, exponentially increasing the computational burden and making accurate prediction of the robot’s behavior exceedingly difficult. Consequently, creating reliable control systems and realistic simulations for soft robots requires entirely new approaches to capture and represent this complex, continuous motion.

The ability to precisely determine a soft robot’s configuration – its 3D shape at any given moment – is fundamental to its effective operation, yet presents a substantial hurdle for researchers. Unlike rigid robots with predictable movements, soft robots continuously deform, creating an infinite number of possible shapes during even simple tasks. This dynamic complexity makes traditional control methods, reliant on precise positional knowledge, largely ineffective. Accurate reconstruction of these shapes isn’t merely an academic exercise; it underpins everything from real-time control and reliable simulation to safe and intuitive human-robot interaction. However, capturing this continuous deformation requires enormous computational power, often necessitating either painstakingly detailed simulations – which struggle with real-world complexities – or extensive robot-specific training datasets for machine learning algorithms, limiting the broad applicability and adaptive capacity of these systems.

Current approaches to modeling soft robotic systems frequently encounter limitations stemming from substantial computational demands or a reliance on large, specialized datasets. Many techniques depend on running complex simulations to predict robot behavior, a process that becomes increasingly prohibitive as the robot’s complexity and the duration of the simulated movement increase. Alternatively, machine learning methods, while offering potential speed gains, often require extensive training data collected specifically for each unique robot design and task. This robot-specific dependency hinders the broader application of these models, making it difficult to adapt them to new robot morphologies or environments without repeating the often-laborious data collection and retraining process. Consequently, a significant challenge remains in developing computationally efficient and readily adaptable modeling techniques for the rapidly evolving field of soft robotics.

Real-time reconstruction of 3D geometry from tactile pressure, demonstrated on a wallet, extends beyond soft robotics to enable the dynamic modeling of object deformation.
Real-time reconstruction of 3D geometry from tactile pressure, demonstrated on a wallet, extends beyond soft robotics to enable the dynamic modeling of object deformation.

Gaussian Splatting: A Compact Representation of Dynamic Form

3D Gaussian Splatting is employed as a novel representation for the dynamic geometry of soft robots, facilitating both high-fidelity reconstruction and real-time rendering capabilities. This technique models the robot’s shape as a collection of 3D Gaussians, where each Gaussian possesses an ellipsoid shape, scale, and rotation. During rendering, these Gaussians are projected onto the 2D image plane and blended using alpha compositing. The method avoids traditional polygon-based meshes, resulting in a more compact representation and reduced computational demands, especially for complex deformations inherent in soft robotics. Reconstruction is achieved through optimization, minimizing the photometric error between rendered projections and captured images, allowing for accurate and detailed representation of the robot’s shape over time.

The method extends 3D Gaussian Splatting to a 4D representation by integrating temporal information, allowing for the reconstruction and rendering of dynamic scenes. This is achieved by treating each Gaussian as a time-varying entity and incorporating data from multiple frames. To ensure smooth transitions between these frames, the Real-Time Interpolation for Video (RIFE) algorithm is utilized. RIFE facilitates the generation of intermediate frames, effectively increasing the temporal resolution and reducing flickering or discontinuities in the reconstructed sequence. This approach enables the rendering of dynamically changing shapes with improved visual fidelity and temporal coherence.

Representing the soft robot as a collection of anisotropic 3D Gaussians provides a computationally efficient alternative to traditional mesh-based or voxel-based methods. Unlike isotropic Gaussians which are uniform in all directions, anisotropic Gaussians allow for deformation along specific axes, accurately capturing the elongated and varying shapes of the robot’s structure with fewer parameters. This compact representation minimizes memory usage and accelerates rendering speeds; the number of Gaussians required to represent the robot is significantly lower than the number of vertices or voxels needed for equivalent detail. Consequently, computational cost is reduced across multiple stages of the reconstruction and rendering pipeline, enabling real-time performance even with complex robot deformations.

The proposed geometry proxy initialization accurately predicts soft robot deformation-demonstrated through qualitative comparisons of predicted 3D Gaussian models, rendered views, and real-world observations for both twisting and bending-enabling accurate reconstruction of unseen soft robot configurations.
The proposed geometry proxy initialization accurately predicts soft robot deformation-demonstrated through qualitative comparisons of predicted 3D Gaussian models, rendered views, and real-world observations for both twisting and bending-enabling accurate reconstruction of unseen soft robot configurations.

Tactile Sensing: Reconstructing Form Without Sight

The reconstruction system employs a Flexible Sensor Array integrated into the robot’s exterior to measure localized surface deformations. This array consists of multiple tactile sensors distributed across the robot’s deformable body, each providing data on the magnitude and direction of physical displacement. By capturing these deformation signals, the system generates a dataset representing the robot’s changing shape without the need for external visual sensors such as cameras. The sensor array’s data serves as the primary input for subsequent processing stages, enabling the reconstruction of the 3D shape based solely on tactile information. The sensor array utilizes capacitive sensing to measure the distance between the sensor and the robot’s surface, allowing for precise detection of even minor deformations.

Tactile data processing employs Structure-Aware Deformation Propagation and Cage-Based Deformation to generate 3D reconstructions. This process utilizes a Graph Attention Network (GAT) which receives local deformation measurements from the flexible sensor array as node features. The GAT then propagates these features across a graph representing the robot’s structural connectivity, allowing information to be shared and aggregated. This propagation step effectively infers deformations in areas not directly sensed, creating a globally consistent 3D shape representation. The cage-based deformation component constrains the reconstruction to maintain a plausible robot shape, preventing unrealistic or physically impossible configurations during the propagation process.

Quantitative evaluation of the proposed reconstruction method on previously unseen soft robotic objects demonstrates high accuracy. Specifically, the system achieves an average Chamfer Distance of 3.48 mm, indicating the average distance between reconstructed and ground truth surfaces. Furthermore, the method yields an average Intersection-over-Union (IoU) score of 0.67, quantifying the volumetric overlap between the reconstructed shape and the reference geometry. These metrics confirm the system’s ability to generate accurate 3D reconstructions from tactile data alone.

Cage-based deformation networks utilize a graph attention mechanism to propagate deformations across a cage structure, adaptively weighting neighboring vertices to model spatially coherent and flexible, even anisotropic, deformations.
Cage-based deformation networks utilize a graph attention mechanism to propagate deformations across a cage structure, adaptively weighting neighboring vertices to model spatially coherent and flexible, even anisotropic, deformations.

Towards Ubiquitous Soft Robotics: Zero-Shot Generalization and Broad Impact

The developed method exhibits a remarkable capacity for zero-shot generalization, meaning it can accurately reconstruct the deformation of soft robotic bodies it has never encountered during training. This is achieved without requiring any robot-specific adaptation or fine-tuning; the system directly infers the three-dimensional pose and deformation from visual input, regardless of the robot’s design or morphology. This capability represents a significant advancement, as traditional approaches necessitate extensive data collection and training for each new robotic platform. By eliminating this dependency, the system drastically simplifies deployment and expands the potential for utilizing soft robots in dynamic and previously inaccessible environments, fostering adaptability and paving the way for more versatile robotic solutions.

The system achieves a processing speed of approximately 30 frames per second (FPS) utilizing a single NVIDIA RTX 3070 Ti GPU, a performance level critical for real-world deployment. This real-time capability allows for immediate visual feedback and control, moving beyond purely simulated environments and enabling dynamic interaction with the physical world. Such speed is essential for applications requiring responsive control, like surgical assistance or in-field repairs, where delays could compromise functionality or safety. The ability to reconstruct soft robot deformations at this rate represents a significant step toward truly adaptable and versatile robotic systems, paving the way for broader integration into diverse and unpredictable settings.

The system’s capacity for precise deformation reconstruction is demonstrated through consistently low angular errors. Evaluations reveal an average bending angle error of just 4.7°, and a twisting angle error of 4.9°. These metrics indicate the method accurately captures and reproduces the complex movements of soft robots, even during intricate deformations. Such high precision is critical for applications demanding fine motor control, like surgical assistance or delicate manipulation of objects, and confirms the system’s reliability in reconstructing a robot’s pose with minimal deviation from its actual configuration.

The ability to reconstruct soft robot deformations without prior training unlocks deployment in environments previously considered too challenging. Unlike traditional robots requiring precise mapping and pre-programming for each new setting, this system allows soft robots to operate effectively in dynamic and unpredictable spaces – from the cluttered environment of a surgical theatre to the uneven terrain of a disaster zone, or even the complex interiors of the human body. This adaptability isn’t merely a technical achievement; it fundamentally expands the scope of potential applications, moving beyond highly structured industrial tasks to enable responsive assistance in healthcare, versatile manipulation in manufacturing, and resilient exploration in previously inaccessible areas, ultimately fostering a new generation of truly versatile robotic tools.

The development of an efficient and adaptable reconstruction system promises a new generation of soft robots capable of operating with increased robustness and intelligence. This technology transcends the limitations of pre-programmed movements, allowing robots to respond dynamically to unforeseen circumstances and complex environments. Consequently, applications extend significantly into fields requiring delicate precision and adaptability, such as minimally invasive surgery and personalized prosthetics within healthcare. In manufacturing, these robots offer potential for intricate assembly tasks and inspection in unstructured settings. Furthermore, the capacity for real-time reconstruction and adaptation is crucial for deploying soft robots in challenging exploration scenarios, including search and rescue operations or planetary exploration, where pre-defined routines are insufficient to navigate unpredictable terrains and obstacles.

Pneumatically actuated soft robots exhibiting twisting and bending motions demonstrate zero-shot evaluation capabilities.
Pneumatically actuated soft robots exhibiting twisting and bending motions demonstrate zero-shot evaluation capabilities.

The presented framework elegantly addresses the challenge of reconstructing soft robot geometry without relying on visual data. It mirrors a holistic understanding of systems-recognizing that accurate deformation reconstruction necessitates a model of the robot’s inherent structure and its interaction with the environment. As Ada Lovelace observed, “The Analytical Engine has no pretensions whatever to originate anything. It can do whatever we know how to order it to perform.” This sentiment resonates with the approach taken in this work; the system doesn’t ‘guess’ the shape, but meticulously reconstructs it based on tactile input and a predefined structural model, enabling zero-shot generalization to previously unseen geometries. The cage-based deformation model provides that essential ‘order’, allowing the system to interpret sensory data and accurately represent the robot’s state.

Future Directions

The decoupling of perception from visual input, demonstrated here, is a necessary, if belated, step. Reliance on vision in robotics has long masked fundamental shortcomings in geometric understanding and control. This work, while promising, reveals the persistent challenge of transferring learned deformation models across increasingly complex topologies. The ‘zero-shot’ claim, while technically achieved, glosses over the implicit assumptions about material properties and the limited diversity of tested geometries. A truly generalizable system demands a more robust encoding of material priors, perhaps through integration with physics-informed neural networks, and a methodology for adaptive cage construction-a cage, after all, is only as good as the space it defines.

The framework’s current reliance on a pre-defined cage also hints at a broader limitation: the tendency to impose artificial structure on inherently amorphous systems. Soft robotics, at its core, seeks to abandon rigid constraints, yet much of the perception pipeline remains tethered to Cartesian coordinates and pre-defined meshes. Future investigations should explore perception modalities that embrace, rather than resist, the fluid nature of these materials – perhaps drawing inspiration from biological systems capable of proprioceptive awareness without explicit geometric mapping.

Ultimately, the true measure of this, or any, robotic system lies not in its ability to reconstruct a shape, but in its capacity to act effectively within a dynamic environment. Good architecture is invisible until it breaks, and only then is the true cost of decisions visible.


Original article: https://arxiv.org/pdf/2603.19543.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-03-23 20:32