Stanford AI Lab Papers and Talks at NeurIPS 2025
- •Stanford AI Lab presents diverse research at NeurIPS 2025 covering agentic frameworks and diffusion models.
- •SWE-smith introduces large-scale data scaling to improve software engineering agents' performance on real-world tasks.
- •New benchmarks like SATBench and CodeARC test logical reasoning and program synthesis in LLM agents.
The Stanford AI Lab (SAIL) is set to showcase a formidable range of research at the NeurIPS 2025 conference in San Diego. This year’s contributions span the entire spectrum of modern AI, from enhancing the efficiency of automated software assistants (Agentic AI) to refining how models understand physical movement through diffusion-based policies. One standout is the Agentic Bridge Framework, which aims to bridge the gap between a model's raw capability and its actual performance on complex benchmarks.
A significant portion of the work focuses on making Large Language Models more reliable through better training methods. For instance, researchers are exploring how to use the time it takes a model to respond as a signal for training (Preference Learning with Response Time), building upon Reinforcement Learning from Human Feedback (RLHF). Additionally, the SWE-smith project addresses the data scarcity in software engineering, providing a way to scale data to train more capable Coding Agent systems.
Beyond text, SAIL is pushing the boundaries of generative models in the physical world. Projects like DynaGuide utilize a Diffusion Model to steer robotic policies, while other papers apply these same generative principles to biological engineering for protein structure design. By developing new ways to benchmark logical reasoning through SATBench and evaluate how models navigate social interactions, Stanford continues to define the Evaluation Metrics used to measure the next generation of artificial intelligence.