Can AI Truly Navigate Like Humans?

Goal-oriented navigation in urban environments demands an agent translate linguistic instructions into progressive actions based on continuous observation, yet current large multimodal models demonstrate a significant disparity in spatial reasoning and action execution compared to human capabilities.

A new benchmark assesses the ability of advanced AI models to perform goal-oriented navigation in complex urban environments, revealing critical limitations in spatial understanding.

Robots Learn to Walk and Work

The study contrasts standard dynamics rollouts-where actions directly control a multi-body dynamics model at the joint level-with a network-policy-augmented approach, demonstrating how actions can instead serve as inputs to a low-level locomotion policy, effectively shifting the control paradigm.

A new control framework empowers robots to seamlessly combine locomotion and manipulation tasks, opening doors for more versatile and adaptable robotic systems.