Alibaba Develops AI Agent That Reduces Tool Usage by 96% While Improving Accuracy
Alibaba researchers created Metis, an AI agent that cuts redundant tool calls from 98% to 2% using a new training framework called Hierarchical Decoupled Policy Optimization.

Alibaba researchers have introduced a new AI training framework that addresses a key challenge in AI agent development: teaching models when to use external tools versus relying on internal knowledge. The framework, called Hierarchical Decoupled Policy Optimization (HDPO), was used to create Metis, a multimodal AI agent that dramatically reduces unnecessary tool usage while maintaining high accuracy.
Current AI agents suffer from what researchers describe as a "profound metacognitive deficit," where models invoke external tools like web search or code execution even when the information needed is already available in the user's prompt. This behavior creates latency bottlenecks, increases API costs, and can degrade reasoning performance by introducing environmental noise.
The HDPO framework solves this problem by separating accuracy and efficiency into two independent optimization channels during training. The accuracy channel focuses on maximizing task correctness, while the efficiency channel optimizes for execution economy. Importantly, the efficiency signal is conditional upon accuracy, meaning incorrect responses are never rewarded simply for being fast or using fewer tools.
Metis, built on the Qwen3-VL-8B-Instruct vision-language model, was tested against various open-source and state-of-the-art models on visual perception, document understanding, and mathematical reasoning tasks. The agent achieved state-of-the-art or highly competitive performance while reducing redundant tool invocations from 98% to just 2%. In practical examples, Metis demonstrated strategic thinking by recognizing when text in images was clearly readable without processing tools, or when fine-grained visual analysis required precision cropping.
The research addresses broader trends in AI development, where the scaffolding layers that developers once needed to build LLM applications are becoming less necessary as models become more capable. Industry experts note that context is becoming increasingly important as a differentiator, while the technical complexity of building AI applications continues to decrease as models improve at reasoning and self-correction.