New Benchmark Challenges AI Capability to Use Tools
- •Agentic-MME introduces a process-verified benchmark for assessing how multimodal models use external tools
- •Dataset includes 418 tasks with over 2,000 human-annotated checkpoints for step-by-step evaluation
- •Gemini3-pro scores 56.3% overall, dropping to 23.0% on the most complex real-world tasks
The shift from passive AI to active agents marks a new chapter in how machines interact with the world. We are moving beyond simple chatbots into an era of 'agentic' intelligence, where systems actively solve multi-step problems by using external tools like search engines or coding environments to complete complex tasks.
However, a fundamental gap exists in how we measure this. Most evaluations are simplistic, checking only if the final answer is correct while ignoring the steps taken to arrive there. Agentic-MME addresses this by verifying the entire process, from the first tool invocation to the final conclusion. It introduces a rigorous benchmark comprising 418 real-world tasks across six domains.
By comparing a model’s trajectory against human-verified steps, the benchmark identifies if an AI is overthinking or failing to utilize resources efficiently. The results offer a sobering reality check. Even leading models like Gemini3-pro struggle as tasks become more complex, with performance dropping to just 23% in the most difficult scenarios. This underscores a persistent weakness: current multimodal models may excel at conversation, but they lack the robustness required for reliable, real-world problem solving.