Author: Denis Avetisyan
A new method enables real-time 3D reconstruction of soft robots using only tactile sensing, eliminating the need for cameras and opening doors to more robust and adaptable robotic systems.

This work introduces a zero-shot learning framework leveraging flexible sensors and cage-based 3D Gaussian modeling for accurate and geometry-independent soft robot deformation reconstruction.
Reconstructing the 3D pose of soft robots remains challenging due to their inherent deformability and lack of rigid features. This is addressed in ‘Zero Shot Deformation Reconstruction for Soft Robots Using a Flexible Sensor Array and Cage Based 3D Gaussian Modeling’, which introduces a novel framework for real-time, camera-free 3D shape reconstruction. By integrating tactile sensing with a cage-based 3D Gaussian deformation model, the system achieves zero-shot generalization to previously unseen robot geometries and motions. Could this approach unlock more robust and adaptable control strategies for a wider range of soft robotic applications?
The Challenge of Representing Continuous Form
Conventional robotic modeling techniques, designed for rigid bodies with predictable movements, falter when applied to soft robots. These robots, constructed from compliant materials, exhibit continuous deformation – bending, stretching, and twisting – creating an infinite number of possible configurations. This contrasts sharply with the limited, discrete movements of traditional robots, which simplifies their mathematical representation. The high degree of freedom inherent in soft robotics-the ability to move in numerous, interconnected ways-further complicates the modeling process. Each degree of freedom introduces additional variables that must be accounted for, exponentially increasing the computational burden and making accurate prediction of the robot’s behavior exceedingly difficult. Consequently, creating reliable control systems and realistic simulations for soft robots requires entirely new approaches to capture and represent this complex, continuous motion.
The ability to precisely determine a soft robot’s configuration – its 3D shape at any given moment – is fundamental to its effective operation, yet presents a substantial hurdle for researchers. Unlike rigid robots with predictable movements, soft robots continuously deform, creating an infinite number of possible shapes during even simple tasks. This dynamic complexity makes traditional control methods, reliant on precise positional knowledge, largely ineffective. Accurate reconstruction of these shapes isn’t merely an academic exercise; it underpins everything from real-time control and reliable simulation to safe and intuitive human-robot interaction. However, capturing this continuous deformation requires enormous computational power, often necessitating either painstakingly detailed simulations – which struggle with real-world complexities – or extensive robot-specific training datasets for machine learning algorithms, limiting the broad applicability and adaptive capacity of these systems.
Current approaches to modeling soft robotic systems frequently encounter limitations stemming from substantial computational demands or a reliance on large, specialized datasets. Many techniques depend on running complex simulations to predict robot behavior, a process that becomes increasingly prohibitive as the robot’s complexity and the duration of the simulated movement increase. Alternatively, machine learning methods, while offering potential speed gains, often require extensive training data collected specifically for each unique robot design and task. This robot-specific dependency hinders the broader application of these models, making it difficult to adapt them to new robot morphologies or environments without repeating the often-laborious data collection and retraining process. Consequently, a significant challenge remains in developing computationally efficient and readily adaptable modeling techniques for the rapidly evolving field of soft robotics.

Gaussian Splatting: A Compact Representation of Dynamic Form
3D Gaussian Splatting is employed as a novel representation for the dynamic geometry of soft robots, facilitating both high-fidelity reconstruction and real-time rendering capabilities. This technique models the robot’s shape as a collection of 3D Gaussians, where each Gaussian possesses an ellipsoid shape, scale, and rotation. During rendering, these Gaussians are projected onto the 2D image plane and blended using alpha compositing. The method avoids traditional polygon-based meshes, resulting in a more compact representation and reduced computational demands, especially for complex deformations inherent in soft robotics. Reconstruction is achieved through optimization, minimizing the photometric error between rendered projections and captured images, allowing for accurate and detailed representation of the robot’s shape over time.
The method extends 3D Gaussian Splatting to a 4D representation by integrating temporal information, allowing for the reconstruction and rendering of dynamic scenes. This is achieved by treating each Gaussian as a time-varying entity and incorporating data from multiple frames. To ensure smooth transitions between these frames, the Real-Time Interpolation for Video (RIFE) algorithm is utilized. RIFE facilitates the generation of intermediate frames, effectively increasing the temporal resolution and reducing flickering or discontinuities in the reconstructed sequence. This approach enables the rendering of dynamically changing shapes with improved visual fidelity and temporal coherence.
Representing the soft robot as a collection of anisotropic 3D Gaussians provides a computationally efficient alternative to traditional mesh-based or voxel-based methods. Unlike isotropic Gaussians which are uniform in all directions, anisotropic Gaussians allow for deformation along specific axes, accurately capturing the elongated and varying shapes of the robot’s structure with fewer parameters. This compact representation minimizes memory usage and accelerates rendering speeds; the number of Gaussians required to represent the robot is significantly lower than the number of vertices or voxels needed for equivalent detail. Consequently, computational cost is reduced across multiple stages of the reconstruction and rendering pipeline, enabling real-time performance even with complex robot deformations.

Tactile Sensing: Reconstructing Form Without Sight
The reconstruction system employs a Flexible Sensor Array integrated into the robot’s exterior to measure localized surface deformations. This array consists of multiple tactile sensors distributed across the robot’s deformable body, each providing data on the magnitude and direction of physical displacement. By capturing these deformation signals, the system generates a dataset representing the robot’s changing shape without the need for external visual sensors such as cameras. The sensor array’s data serves as the primary input for subsequent processing stages, enabling the reconstruction of the 3D shape based solely on tactile information. The sensor array utilizes capacitive sensing to measure the distance between the sensor and the robot’s surface, allowing for precise detection of even minor deformations.
Tactile data processing employs Structure-Aware Deformation Propagation and Cage-Based Deformation to generate 3D reconstructions. This process utilizes a Graph Attention Network (GAT) which receives local deformation measurements from the flexible sensor array as node features. The GAT then propagates these features across a graph representing the robot’s structural connectivity, allowing information to be shared and aggregated. This propagation step effectively infers deformations in areas not directly sensed, creating a globally consistent 3D shape representation. The cage-based deformation component constrains the reconstruction to maintain a plausible robot shape, preventing unrealistic or physically impossible configurations during the propagation process.
Quantitative evaluation of the proposed reconstruction method on previously unseen soft robotic objects demonstrates high accuracy. Specifically, the system achieves an average Chamfer Distance of 3.48 mm, indicating the average distance between reconstructed and ground truth surfaces. Furthermore, the method yields an average Intersection-over-Union (IoU) score of 0.67, quantifying the volumetric overlap between the reconstructed shape and the reference geometry. These metrics confirm the system’s ability to generate accurate 3D reconstructions from tactile data alone.

Towards Ubiquitous Soft Robotics: Zero-Shot Generalization and Broad Impact
The developed method exhibits a remarkable capacity for zero-shot generalization, meaning it can accurately reconstruct the deformation of soft robotic bodies it has never encountered during training. This is achieved without requiring any robot-specific adaptation or fine-tuning; the system directly infers the three-dimensional pose and deformation from visual input, regardless of the robot’s design or morphology. This capability represents a significant advancement, as traditional approaches necessitate extensive data collection and training for each new robotic platform. By eliminating this dependency, the system drastically simplifies deployment and expands the potential for utilizing soft robots in dynamic and previously inaccessible environments, fostering adaptability and paving the way for more versatile robotic solutions.
The system achieves a processing speed of approximately 30 frames per second (FPS) utilizing a single NVIDIA RTX 3070 Ti GPU, a performance level critical for real-world deployment. This real-time capability allows for immediate visual feedback and control, moving beyond purely simulated environments and enabling dynamic interaction with the physical world. Such speed is essential for applications requiring responsive control, like surgical assistance or in-field repairs, where delays could compromise functionality or safety. The ability to reconstruct soft robot deformations at this rate represents a significant step toward truly adaptable and versatile robotic systems, paving the way for broader integration into diverse and unpredictable settings.
The system’s capacity for precise deformation reconstruction is demonstrated through consistently low angular errors. Evaluations reveal an average bending angle error of just 4.7°, and a twisting angle error of 4.9°. These metrics indicate the method accurately captures and reproduces the complex movements of soft robots, even during intricate deformations. Such high precision is critical for applications demanding fine motor control, like surgical assistance or delicate manipulation of objects, and confirms the system’s reliability in reconstructing a robot’s pose with minimal deviation from its actual configuration.
The ability to reconstruct soft robot deformations without prior training unlocks deployment in environments previously considered too challenging. Unlike traditional robots requiring precise mapping and pre-programming for each new setting, this system allows soft robots to operate effectively in dynamic and unpredictable spaces – from the cluttered environment of a surgical theatre to the uneven terrain of a disaster zone, or even the complex interiors of the human body. This adaptability isn’t merely a technical achievement; it fundamentally expands the scope of potential applications, moving beyond highly structured industrial tasks to enable responsive assistance in healthcare, versatile manipulation in manufacturing, and resilient exploration in previously inaccessible areas, ultimately fostering a new generation of truly versatile robotic tools.
The development of an efficient and adaptable reconstruction system promises a new generation of soft robots capable of operating with increased robustness and intelligence. This technology transcends the limitations of pre-programmed movements, allowing robots to respond dynamically to unforeseen circumstances and complex environments. Consequently, applications extend significantly into fields requiring delicate precision and adaptability, such as minimally invasive surgery and personalized prosthetics within healthcare. In manufacturing, these robots offer potential for intricate assembly tasks and inspection in unstructured settings. Furthermore, the capacity for real-time reconstruction and adaptation is crucial for deploying soft robots in challenging exploration scenarios, including search and rescue operations or planetary exploration, where pre-defined routines are insufficient to navigate unpredictable terrains and obstacles.

The presented framework elegantly addresses the challenge of reconstructing soft robot geometry without relying on visual data. It mirrors a holistic understanding of systems-recognizing that accurate deformation reconstruction necessitates a model of the robot’s inherent structure and its interaction with the environment. As Ada Lovelace observed, “The Analytical Engine has no pretensions whatever to originate anything. It can do whatever we know how to order it to perform.” This sentiment resonates with the approach taken in this work; the system doesn’t ‘guess’ the shape, but meticulously reconstructs it based on tactile input and a predefined structural model, enabling zero-shot generalization to previously unseen geometries. The cage-based deformation model provides that essential ‘order’, allowing the system to interpret sensory data and accurately represent the robot’s state.
Future Directions
The decoupling of perception from visual input, demonstrated here, is a necessary, if belated, step. Reliance on vision in robotics has long masked fundamental shortcomings in geometric understanding and control. This work, while promising, reveals the persistent challenge of transferring learned deformation models across increasingly complex topologies. The ‘zero-shot’ claim, while technically achieved, glosses over the implicit assumptions about material properties and the limited diversity of tested geometries. A truly generalizable system demands a more robust encoding of material priors, perhaps through integration with physics-informed neural networks, and a methodology for adaptive cage construction-a cage, after all, is only as good as the space it defines.
The framework’s current reliance on a pre-defined cage also hints at a broader limitation: the tendency to impose artificial structure on inherently amorphous systems. Soft robotics, at its core, seeks to abandon rigid constraints, yet much of the perception pipeline remains tethered to Cartesian coordinates and pre-defined meshes. Future investigations should explore perception modalities that embrace, rather than resist, the fluid nature of these materials – perhaps drawing inspiration from biological systems capable of proprioceptive awareness without explicit geometric mapping.
Ultimately, the true measure of this, or any, robotic system lies not in its ability to reconstruct a shape, but in its capacity to act effectively within a dynamic environment. Good architecture is invisible until it breaks, and only then is the true cost of decisions visible.
Original article: https://arxiv.org/pdf/2603.19543.pdf
Contact the author: https://www.linkedin.com/in/avetisyan/
See also:
- Invincible Season 4 Episode 4 Release Date, Time, Where to Watch
- Physics Proved by AI: A New Era for Automated Reasoning
- American Idol vet Caleb Flynn in solitary confinement after being charged for allegedly murdering wife
- Gold Rate Forecast
- “Wild, brilliant, emotional”: 10 best dynasty drama series to watch on BBC, ITV, Netflix and more
- Total Football free codes and how to redeem them (March 2026)
- Seeing in the Dark: Event Cameras Guide Robots Through Low-Light Spaces
- Magicmon: World redeem codes and how to use them (March 2026)
- eFootball 2026 is bringing the v5.3.1 update: What to expect and what’s coming
- Hatch Dragons Beginners Guide and Tips
2026-03-23 20:32