Building Reliable Scientific Agents with Structured Execution
A new framework leverages typed execution graphs to enhance the reproducibility and scalability of complex scientific workflows powered by large language model agents.
A new framework leverages typed execution graphs to enhance the reproducibility and scalability of complex scientific workflows powered by large language model agents.
Overbroad discussions of ‘artificial intelligence’ obscure critical details and impede effective oversight, particularly when considering its application in sensitive areas.

Researchers have developed a novel soft robotic actuator and quadruped capable of dynamically switching between walking and jumping gaits.

As artificial intelligence moves beyond isolated models to complex, autonomous agents, rigorous evaluation is no longer a supporting task, but a critical necessity.

New research evaluates the best methods for equipping AI chatbots with the knowledge they need to answer questions about complex scientific literature.

New research reveals that while AI agents are quickly adopted for code contributions, they often introduce type-related issues in TypeScript projects.

Researchers have demonstrated that a surprisingly simple vision-language-action model can achieve state-of-the-art performance in robotic manipulation tasks.
![ModelSMC automatically discovers models from textual problem formulations and context data by iteratively refining an initial model-inspired by Sequential Monte Carlo methods that approximate distributions via weighted particles-to sample high-density regions of the model posterior [latex] p(m|\bm{x}_o) [/latex], effectively approaching the underlying data-generating process through model propagation via large language model sampling and weighting based on likelihood evaluation.](https://arxiv.org/html/2602.18266v1/x1.png)
A new framework harnesses the power of large language models to automate the process of scientific model discovery, moving beyond traditional methods.
![The system iteratively refines its understanding of a scene through interactive perception, where a robot evaluates observations and queries-leveraging a memory [latex]\mathcal{M}_{t}[/latex] to avoid repetition-and, when necessary, annotates images with segmented elements like push lines, keypoints, or grid patterns to inform subsequent actions [latex]a_{t}=\pi(\mathcal{M}_{t},\textbf{x}_{t},z_{t},\tilde{o}_{t})[/latex], ultimately building a contextual history [latex]S_{t}[/latex] of images and scene descriptions to enhance task efficiency.](https://arxiv.org/html/2602.18374v1/Figures/rss25/InterPer2.png)
A new framework empowers robots to manipulate objects and answer questions about their environment, even when vision is limited, by combining language understanding with enhanced visual perception.
A new framework seeks to pinpoint human accountability for the actions of increasingly complex artificial intelligence systems.