Author: Denis Avetisyan
A new framework combines visual understanding and active perception, enabling robots to more effectively locate and manipulate objects in real-world environments.

SaPaVe decouples action spaces and leverages vision-language models to improve semantic grounding and spatial reasoning for embodied AI.
Robust robotic interaction with complex environments requires both perceiving relevant information and executing precise manipulations, yet current vision-language-action models struggle to unify these capabilities in a viewpoint-invariant manner. This paper introduces ‘SaPaVe: Towards Active Perception and Manipulation in Vision-Language-Action Models for Robotics’, a novel end-to-end framework that learns decoupled camera and manipulation actions via a bottom-up training strategy, achieving improved data efficiency and execution robustness. By introducing the ActiveViewPose-200K dataset and ActiveManip-Bench benchmark, we demonstrate that tightly coupled perception and execution-when trained with coordinated yet decoupled strategies-significantly outperforms existing models, achieving up to 31.25\% higher success rates in real-world tasks. Could this approach pave the way for more adaptable and intelligent robots capable of truly understanding and interacting with their surroundings?
Beyond Static Vision: The Robot’s Need to See More
Conventional robotic systems, designed for manipulation, often operate with a static visual perspective, a limitation that introduces significant challenges when navigating complex, cluttered spaces. This fixed viewpoint creates a perceptual bottleneck, as critical information about objects obscured from the primary camera remains inaccessible, hindering the robot’s ability to plan effective grasping or manipulation strategies. Consequently, the robot may struggle to differentiate between objects, accurately assess their pose, or even determine if a clear path for interaction exists. This reliance on a single, unchanging perspective necessitates pre-programmed trajectories or extensive re-planning, dramatically reducing efficiency and adaptability in dynamic environments where object positions are constantly shifting or partially hidden.
Robotic systems relying solely on passive observation frequently encounter limitations in complex environments. A static viewpoint provides an incomplete picture of the surroundings, as objects can be obscured or critical features hidden from view. This restricted information access directly impacts task completion rates; a robot unable to fully perceive its workspace struggles to accurately plan movements, grasp objects securely, or adapt to unforeseen circumstances. The resulting uncertainty necessitates cautious, often inefficient, strategies, hindering performance and scalability. Consequently, a reliance on purely passive sensing creates a significant bottleneck, preventing robots from achieving the dexterity and robustness necessary for real-world applications demanding complete situational awareness.
Robotic systems are increasingly challenged by real-world complexity, necessitating a shift from static observation to active perception. This emerging paradigm proposes that robots should not simply receive visual data, but rather proactively control their viewpoints to maximize information gain. By strategically moving and orienting sensors, a robot can resolve ambiguities, overcome occlusions, and build a more complete understanding of its environment. This active exploration isn’t merely about acquiring more data; it’s about intelligently seeking the right data – the views that most effectively support task completion and robust manipulation in cluttered, dynamic scenes. Such an approach promises to unlock more adaptable and reliable robotic systems capable of navigating and interacting with the world much as humans do.

SaPaVe: A Unified Framework for Perception and Action
SaPaVe establishes a unified framework integrating semantic perception and action execution, departing from traditional pipelines that treat these as separate modules. This system directly links visual-language understanding with robotic control, enabling the robot to not only interpret the semantic content of a scene but also to actively manipulate its viewpoint to improve that understanding. The architecture processes visual input and language prompts to generate action commands, closing the loop between perception and execution within a single, cohesive system. This contrasts with methods requiring discrete perception and planning stages, allowing for more efficient and adaptable behavior in dynamic environments.
SaPaVe utilizes Vision-Language Models (VLMs) to process visual input and associated textual descriptions of a scene, enabling the system to understand object relationships and contextual information. This understanding is then used to predict the informational value of different camera viewpoints. Specifically, the VLM analyzes the current scene representation and estimates the expected reduction in uncertainty regarding object poses and scene semantics achievable by moving to a particular camera pose. The system then selects camera movements that maximize this expected information gain, effectively directing the camera to focus on the most informative regions of the scene for subsequent robotic manipulation.
The SaPaVe framework employs a two-stage training methodology to optimize performance. Initially, the system undergoes training specifically focused on achieving robust semantic understanding of visual scenes, utilizing datasets designed to enhance its ability to interpret and categorize objects and relationships. Following this initial phase, a second stage integrates the semantically-trained model with the robotic manipulation pipeline; this allows the system to translate its understanding of the scene into actionable commands for camera movement and object interaction. This staged approach facilitates improved generalization and allows for focused optimization of each component before their integration, resulting in a more effective end-to-end system.
Evaluations on the ActiveViewPose-200K dataset demonstrate SaPaVe’s improved performance in semantic active perception, achieving a 16% increase over Gemini 2.5 Pro. This improvement is quantified by metrics assessing the accuracy of both viewpoint prediction and pose estimation within complex scenes. The results indicate SaPaVe’s enhanced capability to interpret visual information and strategically select camera movements for more effective object recognition and scene understanding, representing a significant advancement over existing Vision-Language Models in this domain.

Architectural Innovations: Dissecting Control for Dynamic Responses
Decoupled action heads facilitate independent control of both the camera and robotic arm, addressing limitations inherent in systems where these actions are jointly predicted. This decoupling allows for more precise and coordinated manipulation by enabling the model to optimize each action separately based on the current state and desired outcome. Specifically, separate output heads are used to predict camera movements and robotic arm trajectories, avoiding constraints imposed by joint prediction and improving the system’s ability to handle complex tasks requiring independent control of visual observation and physical interaction. This architecture results in smoother, more efficient motion planning and enhanced performance in dynamic environments.
The Diffusion Transformer (DiT) architecture forms the core of the action prediction module, leveraging a transformer-based approach to model sequential dependencies in robotic manipulation. DiT utilizes a diffusion process to generate action sequences, iteratively refining predictions from noise based on observed states and goals. This differs from autoregressive models by enabling parallel decoding and improving robustness to long-horizon predictions. The framework employs a discrete diffusion process, quantizing the action space to facilitate stable and efficient learning. Experimental results demonstrate that DiT outperforms standard autoregressive models in terms of both prediction accuracy and sample efficiency for complex robotic tasks, achieving state-of-the-art performance on benchmark datasets.
The Universal Spatial Encoder (USE) is a core component designed to ingest and process 3D geometric information from the environment, specifically point clouds, to provide enhanced spatial awareness for the robotic system. This encoder transforms raw 3D data into a latent spatial representation, which is then integrated with visual and language embeddings. The resulting combined embedding allows the system to reason about object positions, shapes, and relationships, improving the accuracy and efficiency of action planning and execution. The USE operates independently of the specific robotic arm or camera configuration, facilitating generalization across different setups and tasks. It leverages a convolutional neural network architecture optimized for point cloud processing, outputting a fixed-length vector representing the spatial scene understanding.
Parameter-efficient fine-tuning utilizing Low-Rank Adaptation (LoRA) addresses the computational cost associated with adapting large Vision-Language Models (VLMs) to specialized robotic manipulation tasks. Instead of updating all VLM parameters, LoRA introduces trainable low-rank matrices that represent the parameter updates; these matrices are significantly smaller than the original model weights. This approach drastically reduces the number of trainable parameters – often by over 90% – and the associated memory requirements, enabling effective adaptation on limited computational resources. LoRA maintains the pre-trained knowledge of the VLM while allowing it to learn task-specific behaviors without catastrophic forgetting, and allows for easy swapping and combination of task-specific adaptations.

Benchmarking and Validation: Pushing the Limits in Complex Worlds
SaPaVe’s capabilities are thoroughly assessed using ActiveManip-Bench, a sophisticated simulation environment specifically engineered to push the boundaries of active manipulation research. Unlike traditional benchmarks that rely on static camera perspectives, ActiveManip-Bench demands performance across dynamically changing viewpoints – mirroring the complexities of real-world robotic tasks. This benchmark presents a series of challenges requiring the agent to actively adjust its viewpoint to successfully locate and manipulate objects, thereby evaluating not only the manipulation skill itself but also the ability to plan and execute effective visual exploration strategies. Rigorous testing on ActiveManip-Bench provides a robust and standardized measure of SaPaVe’s performance in scenarios that extend beyond the limitations of fixed-viewpoint evaluations, offering a more realistic and comprehensive assessment of its practical applicability.
The foundation of SaPaVe’s capabilities lies in its training regimen utilizing the ActiveViewPose-200K dataset, a substantial collection designed to bridge the gap between visual perception and robotic action. This dataset uniquely pairs images with corresponding natural language instructions and the necessary camera movements to execute those instructions – a crucial element for active manipulation tasks. By learning from this diverse range of image-language-motion triplets, the framework develops a robust understanding of how visual goals translate into coordinated robotic movements. The scale of ActiveViewPose-200K – encompassing 200,000 examples – enables SaPaVe to generalize effectively to novel scenes and instructions, allowing it to navigate complex environments and successfully complete manipulation tasks with greater reliability.
To bolster the framework’s resilience and adaptability to unseen scenarios, a suite of data augmentation techniques was strategically implemented. These methods artificially expanded the training dataset by introducing variations in lighting conditions, object textures, and background clutter, effectively exposing the system to a wider range of potential real-world disturbances. Furthermore, the framework underwent training with randomized viewpoints and simulated sensor noise, enhancing its ability to generalize beyond the precise conditions present in the original training data. This deliberate diversification of the training environment proved crucial in equipping the system to perform reliably in complex and unpredictable settings, ultimately contributing to its improved performance across both simulated and real-world evaluations.
Evaluations reveal a substantial advancement in robotic task completion through active visual learning. The SaPaVe framework achieves a remarkable 58% absolute success rate improvement on a challenging simulated active manipulation benchmark when contrasted with traditional, fixed-viewpoint visual localization and action (VLA) approaches. This significant leap in performance underscores the benefits of dynamically adjusting viewpoints to gather more informative visual data, enabling the robot to effectively plan and execute complex manipulation tasks. By actively seeking optimal perspectives, the system overcomes limitations inherent in static observation, resulting in considerably more reliable and successful task outcomes in simulated environments.
Real-world testing demonstrates SaPaVe’s substantial advancement in robotic manipulation success. Evaluations reveal a significant performance leap, with SaPaVe achieving a 40% higher success rate when contrasted with the π0 framework and a 31.25% improvement over GR00T-N1. These results, obtained in complex, non-simulated environments, highlight the framework’s capacity to reliably execute active manipulation tasks, suggesting a marked step forward in the development of adaptable and robust robotic systems capable of functioning effectively in unstructured settings. The observed gains underscore the value of SaPaVe’s approach to visual language alignment and active perception in bridging the reality gap for robotic manipulation.

The pursuit within SaPaVe-deconstructing robotic action into semantic understanding and active perception-mirrors a fundamental drive to decipher underlying systems. It’s a process of reverse-engineering reality, much like attempting to read source code for a program one hasn’t yet fully grasped. Blaise Pascal observed, “The eloquence of angels is a song, and the eloquence of men is a syllogism.” This resonates with the framework’s approach; SaPaVe doesn’t rely on pre-programmed responses, but actively seeks information, constructing its own ‘syllogisms’ through visual and linguistic cues. The model’s decoupled action space isn’t simply about manipulating objects; it’s about building a functional understanding of the environment – reading the code, as it were – to achieve a desired outcome.
What’s Next?
The pursuit of embodied AI, as exemplified by frameworks like SaPaVe, inevitably circles back to the fundamental problem of representation. The decoupling of action spaces, while pragmatic, merely shifts the complexity. It doesn’t resolve the inherent ambiguity in translating semantic grounding – a linguistic construct – into the physics of interaction. Every exploit starts with a question, not with intent; future work must aggressively probe the limits of these decoupled spaces, deliberately introducing contradictions to expose the fault lines between perception and execution.
A critical, largely unaddressed limitation lies in the model’s reliance on pre-defined object affordances. Current systems excel at manipulating objects known to be manipulable in expected ways. But reality rarely conforms. The next iteration demands a framework capable of discovering affordances – inferring how an object can be interacted with based on observation, not prior knowledge. This requires moving beyond correlation to a form of causal reasoning, a particularly thorny problem when dealing with noisy sensory data and imperfect models of the physical world.
Ultimately, the success of such systems will not be measured by their ability to mimic human manipulation, but by their capacity to surprise it. To operate effectively in unstructured environments, robots must move beyond predictable behaviors and develop a degree of creative problem-solving. This demands a shift in focus from optimizing for success to tolerating – and learning from – failure, a proposition that requires a fundamental rethinking of both learning algorithms and reward functions.
Original article: https://arxiv.org/pdf/2603.12193.pdf
Contact the author: https://www.linkedin.com/in/avetisyan/
See also:
- CookieRun: Kingdom 5th Anniversary Finale update brings Episode 15, Sugar Swan Cookie, mini-game, Legendary costumes, and more
- Call the Midwife season 16 is confirmed – but what happens next, after that end-of-an-era finale?
- Taimanin Squad coupon codes and how to use them (March 2026)
- Robots That React: Teaching Machines to Hear and Act
- Heeseung is leaving Enhypen to go solo. K-pop group will continue with six members
- Gold Rate Forecast
- PUBG Mobile collaborates with Apollo Automobil to bring its Hypercars this March 2026
- Alan Ritchson’s ‘War Machine’ Netflix Thriller Breaks Military Action Norms
- Marilyn Manson walks the runway during Enfants Riches Paris Fashion Week show after judge reopened sexual assault case against him
- How to get the new MLBB hero Marcel for free in Mobile Legends
2026-03-14 06:25