Building Worlds with Words: A New Framework for AI Navigation

Author: Denis Avetisyan


Researchers have developed a novel system that generates complex, multi-story 3D environments from natural language descriptions, enabling more realistic and challenging tests for embodied AI agents.

The MANSION framework constructs multi-story 3D buildings from natural language descriptions through a sequential pipeline-beginning with whole building planning, refining to per-floor layouts, synthesizing floorplans, and ultimately instantiating the complete scene-demonstrating a system capable of translating linguistic intent into architectural form.
The MANSION framework constructs multi-story 3D buildings from natural language descriptions through a sequential pipeline-beginning with whole building planning, refining to per-floor layouts, synthesizing floorplans, and ultimately instantiating the complete scene-demonstrating a system capable of translating linguistic intent into architectural form.

This paper introduces MANSION, a language-driven framework and dataset for building-scale environment generation and long-horizon embodied AI research.

Existing embodied AI benchmarks largely fail to capture the complexity of real-world tasks requiring multi-floor spatial reasoning and long-horizon planning. To address this, we introduce ‘MANSION: Multi-floor lANguage-to-3D Scene generatIOn for loNg-horizon tasks’, a language-driven framework capable of generating building-scale, navigable 3D environments and accompanying dataset, MansionWorld. This work demonstrates the creation of over 1,000 diverse buildings, alongside an agent for task-specific scene editing via open-vocabulary commands. Will these more realistic and challenging environments catalyze advancements in long-horizon embodied AI and spatial planning capabilities?


The Illusion of Progress: Why AI Still Can’t Plan Beyond the Next Room

Conventional artificial intelligence often falters when confronted with tasks demanding extended sequences of coordinated actions – known as long-horizon tasks. These challenges arise because most AI systems are trained on limited datasets or in simplified simulations, hindering their ability to anticipate consequences far into the future. Unlike humans who intuitively grasp the unfolding of events over time, AI frequently struggles with the compounding uncertainty inherent in prolonged interactions with a dynamic environment. This limitation manifests as difficulty in maintaining coherent plans, adapting to unexpected changes, and effectively allocating resources over extended periods, particularly when the task requires navigating complex, unpredictable scenarios – essentially, a failure to reason effectively about the distant consequences of present actions.

The development of truly versatile embodied artificial intelligence is frequently hampered by a significant gap between training grounds and real-world conditions. Current simulated environments, while useful for initial development, often fall short in replicating the sheer scale and intricate details present in everyday spaces. These limitations extend beyond mere size; a lack of varied object interactions, unpredictable lighting conditions, and the presence of numerous dynamic elements – such as moving people or changing layouts – restrict an agent’s ability to generalize learned behaviors. Consequently, an AI proficient within a simplified simulation may struggle dramatically when deployed into a more complex, unscripted environment, highlighting the critical need for training platforms that accurately mirror the richness and unpredictability of the physical world to foster robust navigation and manipulation skills.

Truly effective embodied artificial intelligence necessitates more than just task completion; it requires systems exhibiting robust reasoning and continuous adaptation within genuinely expansive and interactive environments. Current AI often excels in controlled, limited scenarios, but struggles when confronted with the unpredictable nuances of real-world complexity. To navigate and manipulate such spaces, an agent must not simply react to immediate stimuli, but proactively anticipate consequences, learn from failures, and refine its strategies over extended periods. This demands computational architectures capable of processing vast amounts of sensory information, building internal models of the environment, and generating flexible plans that can be dynamically adjusted as conditions change. The capacity for such reasoning and adaptation isn’t merely a performance enhancement-it’s a fundamental prerequisite for deploying AI agents that can reliably operate and collaborate within the complex, open-ended landscapes of the physical world.

The Task-Semantic Scene Editing Agent utilizes a ReAct controller to iteratively plan and execute tasks via a tool invoker, bridging static semantic understanding with an on-demand physics engine and maintaining consistency through hybrid state management that synchronizes simulation results with the static scene [latex]	ext{JSON}[/latex].
The Task-Semantic Scene Editing Agent utilizes a ReAct controller to iteratively plan and execute tasks via a tool invoker, bridging static semantic understanding with an on-demand physics engine and maintaining consistency through hybrid state management that synchronizes simulation results with the static scene [latex] ext{JSON}[/latex].

MansionWorld: A Larger Sandbox, But Still Just Sand

MansionWorld constitutes a large-scale simulation environment comprising over 1,000 uniquely designed, interactive buildings. These buildings are not static scenes, but rather fully traversable spaces intended to facilitate comprehensive evaluation of artificial intelligence agents. The scale of the environment allows for testing of AI performance across a diverse range of architectural layouts and object arrangements. Each building features multiple floors, enabling investigation of AI capabilities in vertically-oriented navigation and multi-story interaction scenarios, and providing a benchmark for agents operating in complex, real-world environments.

MansionWorld addresses limitations in prior AI simulation environments by enabling agents to perform complex, multi-floor navigation tasks within buildings extending up to 10 floors. Previous simulated environments often restricted agent movement to single-floor layouts, hindering the development of robust navigation and planning capabilities. The inclusion of verticality in MansionWorld necessitates more sophisticated pathfinding, object manipulation across floors, and spatial reasoning, providing a more challenging and realistic benchmark for AI agent evaluation. This focus on cross-floor navigation pushes the boundaries of current AI algorithms and promotes research into more generalized and adaptable agents.

MansionWorld leverages the AI2-THOR simulation platform, providing a physically realistic environment for AI agent training and evaluation. This foundation enables robust interaction with objects and accurate modeling of agent movement. The environment’s composition is specifically designed to represent a diverse range of indoor spaces, comprising 50% residential buildings, 30% office spaces, and 20% public buildings; this distribution supports evaluation across a variety of common navigational and interaction scenarios. Navigation capabilities within MansionWorld have been enhanced beyond those of the base AI2-THOR simulator to facilitate more complex agent behaviors and path planning within the multi-story buildings.

The MansionWorld ecosystem integrates various components, including a blockchain for ownership, an AI-powered agent for interaction, and a procedural content generation system for dynamic environments.
The MansionWorld ecosystem integrates various components, including a blockchain for ownership, an AI-powered agent for interaction, and a procedural content generation system for dynamic environments.

MANSION: Automating the Illusion of Intelligence

MANSION employs a language-driven methodology for both the creation of large-scale environments and the subsequent evaluation of agent performance within those environments. This approach utilizes natural language instructions to procedurally generate building layouts and populate them with objects, allowing for dynamic and customizable scenes. The framework interprets linguistic prompts to define spatial relationships, object properties, and task objectives, translating these into concrete environmental configurations. This enables automated generation of diverse scenarios for testing AI agents, and also facilitates evaluation based on language-defined success criteria, effectively linking task performance to linguistic descriptions of the environment and desired outcomes.

The MANSION framework employs a Task-Semantic Scene Editing Agent to procedurally alter building interiors based on specified tasks, effectively generating diverse and customizable environments for AI agent training and evaluation. This agent doesn’t simply rearrange existing objects; it understands the semantic meaning of rooms and objects to make contextually appropriate modifications – for example, adding a stove to a kitchen or creating obstacles for a navigation task. The dynamic nature of this editing process allows for the creation of an almost limitless number of unique building layouts, differing in size, complexity, and the challenges they present to an AI agent, thus enabling robust and scalable environment generation without manual design.

The Task-Semantic Scene Editing Agent employs a ReAct controller, a reasoning and acting framework, to determine appropriate modifications to the building environment based on task requirements and semantic understanding of the scene. This controller interleaves reasoning steps – analyzing the current state and formulating a plan – with action steps, which involve executing commands to alter the environment. Complementing this is a Hybrid State Management system which combines a differentiable rendering approach for visual state with a symbolic representation for object properties and relationships; this allows for efficient simulation by reducing computational demands while maintaining the fidelity needed for realistic agent interaction and evaluation. The hybrid approach facilitates both gradient-based optimization and symbolic reasoning within the environment.

Holodeck demonstrates superior qualitative performance compared to MANSION when generating scenes from high-level semantic building prompts.
Holodeck demonstrates superior qualitative performance compared to MANSION when generating scenes from high-level semantic building prompts.

The Long Road Ahead: Why Robots Still Can’t Quite Grasp Reality

Recent advancements in embodied artificial intelligence leverage sophisticated frameworks such as MANSION and BUMBLE to facilitate complex, long-horizon task completion for robotic agents. These systems move beyond simple reactive behaviors, enabling agents – exemplified by the COME-robot – to engage in extended planning and execution over prolonged periods. MANSION and BUMBLE accomplish this by providing a robust architecture for hierarchical decision-making, allowing the agent to decompose overarching goals into manageable sub-tasks and adapt its strategy based on environmental feedback. This capability is particularly crucial in dynamic and unpredictable environments where pre-programmed sequences are insufficient, and the agent must continuously reason about its actions and anticipate future consequences to successfully navigate and achieve its objectives.

The development of navigable environments is crucial for embodied artificial intelligence, and the ‘Constrained Growth Solver’ offers a novel approach to efficient floorplan generation. This system doesn’t simply create random layouts; instead, it prioritizes the creation of structured spaces specifically designed to facilitate agent interaction and task completion. By intelligently connecting rooms and defining pathways, the solver ensures that generated floorplans are not only logically consistent but also conducive to an agent’s ability to navigate and achieve goals within the environment. This focus on ‘growable’ and constrained spaces allows for the creation of complex, multi-floor layouts with a significantly reduced computational burden compared to traditional methods, paving the way for more realistic and challenging testbeds for embodied AI research.

Despite promising advancements in embodied artificial intelligence, performance metrics reveal a noticeable decline when agents are tasked with navigating and operating within complex, four-floor environments. This reduction in success rates underscores the significant challenges inherent in scaling AI agents to more realistic and demanding scenarios, emphasizing the need for continued research into robust planning and execution strategies. However, accompanying user studies offer a positive counterpoint, demonstrating a clear preference for the scenes generated by the underlying framework – suggesting that, even with performance limitations, the created environments are perceived as plausible and engaging by human observers, potentially paving the way for more intuitive and effective human-agent interaction in the future.

Our Task-Semantic Scene Editing Agent utilizes a “Check-and-Provision” workflow-involving path connectivity, object availability, and scene editing-to verify task executability before fulfilling high-level instructions like bringing items to a specific location.
Our Task-Semantic Scene Editing Agent utilizes a “Check-and-Provision” workflow-involving path connectivity, object availability, and scene editing-to verify task executability before fulfilling high-level instructions like bringing items to a specific location.

The elegance of MANSION, with its multi-floor language-to-3D scene generation, feels
familiar. It’s a beautiful system for building complex, navigable environments, meticulously designed for long-horizon tasks. One anticipates the inevitable. As Fei-Fei Li once noted, “AI is not about replacing humans; it’s about augmenting human capabilities.” This framework certainly augments simulation, but it’s a safe bet that production will rapidly expose limitations in the generated floorplans, requiring endless, bespoke adjustments. The system may handle cross-floor navigation beautifully in simulation, yet the real world – or even a sufficiently chaotic virtual one – will demand constant patching. It’s the cycle of life for any ambitious AI project.

The Road Ahead

The creation of MANSION and MansionWorld offers a predictably complex environment for embodied agents. One anticipates a flurry of papers demonstrating increasingly sophisticated navigation
until production demands hit. The real challenge, as always, won’t be generating floorplans, but accounting for the arbitrary choices of building managers and the inevitable “quirks” of real-world spaces. The dataset, while expansive, will quickly reveal its limitations; agents trained within its confines will undoubtedly struggle with the unexpected-a misplaced fire extinguisher, a temporary obstacle, a rogue office plant. It’s the usual story.

Future work will likely focus on procedural generation techniques to increase environmental diversity. Expect research into ‘realistic’ human behavior within these simulated buildings – until someone realizes that modeling truly unpredictable human actions is a fool’s errand. The holy grail, of course, remains long-horizon reasoning. But one suspects that ‘long-horizon’ will simply become ‘slightly less short-sighted’ with each iteration.

Ultimately, MANSION represents another step in the endless cycle of abstraction. A more elaborate wrapper around the fundamental difficulties of perception, planning, and execution. It’s a useful step, certainly. But everything new is just the old thing with worse docs.


Original article: https://arxiv.org/pdf/2603.11554.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-03-15 12:29