AI Agents Gain Performance Boost with Dynamic Computing Allocation

Researchers are increasingly focused on optimising model performance through test-time scaling, yet its effectiveness in complex, agentic tasks involving multiple sequential steps requires further investigation. Nicholas Lee, Lutfi Eren Erdogan, and Chris Joseph John from UC Berkeley, working with colleagues including Surya Krishnapillai, Michael W. Mahoney from UC Berkeley, ICSI, and LBNL, and Kurt Keutzer and Amir Gholami, present a new technique called CATTS to address this challenge. Their study reveals that simply increasing computational effort at each step yields diminishing returns in long-horizon web agent environments, and introduces a dynamic compute allocation strategy based on an agent’s own uncertainty. This work is significant because CATTS not only improves performance on benchmarks like WebArena-Lite and GoBrowse by up to 9.1% compared to existing methods, but also achieves these gains with substantial reductions in computational cost, using up to 2.3times fewer tokens, offering a pathway to more efficient and reliable agentic systems.

Researchers have developed a new technique called Confidence-Aware Test-Time Scaling (CATTS) to enhance the performance of artificial intelligence agents undertaking complex, multi-step tasks. This innovation addresses a critical challenge in agentic AI, how to allocate computational resources effectively when small errors at each step can accumulate and derail long-term goals.
Current test-time scaling methods, which increase the number of attempts a model makes to solve a problem, often show diminishing returns in these scenarios, wasting processing power on straightforward decisions. The work demonstrates that simply generating more options does not guarantee improved outcomes, particularly when the model faces genuinely difficult choices.

This study began with a detailed empirical analysis of how inference-time scaling affects web-based agents, revealing that uniformly increasing computational effort quickly plateaus in complex environments. Investigators then explored more sophisticated aggregation strategies, including employing a secondary large language model as an ‘arbiter’ to refine decisions, but found this approach could sometimes override strong consensus among initial model outputs.

Crucially, the research identified that internal signals, specifically, the uncertainty statistics derived from the agent’s own voting distribution, such as entropy and top-vote margins, strongly correlate with the likelihood of downstream success. Building on these findings, CATTS dynamically allocates compute only when the agent is genuinely uncertain, concentrating resources on contentious decisions rather than squandering them on easy ones.

Evaluations on benchmark tasks, WebArena-Lite and GoBrowse, show that CATTS improves performance by up to 9.1% over the standard React approach, while simultaneously reducing token usage by as much as 2.3times. This represents a significant step towards more efficient and reliable agentic AI systems, offering both performance gains and a transparent, interpretable decision-making process. The technique promises to improve the robustness of AI agents operating in real-world scenarios where consistent, accurate performance is paramount.

Confidence and computational efficiency gains from adaptive scaling in agentic environments

Improvements of up to 9.1% in performance on WebArena-Lite and GoBrowse were achieved through the implementation of Confidence-Aware Test-Time Scaling, or CATTS. This gain was measured relative to the React baseline, demonstrating a substantial advancement in agentic task completion. Simultaneously, CATTS reduced token usage by as much as 2.3x compared to uniform scaling methods, indicating a significant increase in computational efficiency.

The research details how uniformly increasing compute per step quickly plateaus in long-horizon environments, highlighting the inefficiency of this approach. Empirical study revealed that simply adding more samples does not consistently improve outcomes, particularly when votes are highly variable and lack a clear consensus.

Analysis of agent vote distributions showed a strong correlation between uncertainty statistics, specifically entropy and top-1/top-2 margin, and downstream task success. Entropy, a measure of randomness, and top-1/top-2 margin, indicating the confidence in the most likely choices, served as practical signals for dynamically allocating compute.

These uncertainty metrics provided insight into when additional computation was most likely to influence the decision-making process. The study found that an LLM-based Arbiter, while capable of outperforming naive voting, was prone to overruling high-consensus decisions, demonstrating the potential for overthinking and harmful overrides.

CATTS leverages these findings by allocating compute only when decisions are genuinely contentious, concentrating resources on the most uncertain steps. This dynamic allocation strategy resulted in consistent performance improvements while minimising unnecessary computation. The work demonstrates that wasted computation occurs on easy steps where a majority of actions are obvious, and that focusing compute on difficult, high-variance decisions is more effective. This approach provides both efficiency gains and an interpretable decision rule, offering a clear rationale for resource allocation.

Empirical scaling analysis of inference techniques for long-horizon web agents

A systematic empirical study of inference-time scaling underpinned this work, beginning with adaptation of established techniques for long-horizon web agents. Best-of-N sampling and voting, reranking via additional rollouts, and confidence-aware filtering methods were all implemented and analysed within the context of complex, multi-step web-based tasks.

This initial phase aimed to determine where these existing methods provided benefit and, crucially, where they exhibited limitations in agentic settings. The research deliberately moved beyond simple performance gains to investigate the underlying reasons for success or failure, identifying signals that could predict the effectiveness of each technique.

Following this initial analysis, the study focused on quantifying uncertainty within the agent’s decision-making process. The distribution of answers generated at each step was examined, revealing a correlation between this distribution and the likelihood of overall task success. Specifically, entropy and top-1/top-2 margin, measures derived from the agent’s vote distribution, proved to be reliable indicators of potential difficulties.

These statistics were then leveraged to create Confidence-Aware Test-Time Scaling (CATTS), a novel technique for dynamically allocating computational resources. CATTS operates by selectively increasing compute only when decisions are genuinely contentious, as determined by the uncertainty metrics. An LLM-based Arbiter was integrated to provide a more sophisticated aggregation strategy, capable of outperforming simple voting, but crucially, CATTS controls its use, preventing overrulings of high-consensus decisions.

This conditional invocation of the Arbiter, guided by vote-derived uncertainty, distinguishes CATTS from uniform scaling approaches that expend tokens indiscriminately. The methodology was evaluated on the WebArena-Lite and GoBrowse benchmarks, allowing for a direct comparison of CATTS against baseline methods like React.

The Bigger Picture

The relentless pursuit of more capable artificial intelligence often focuses on model size, but increasingly, attention is turning to how those models think , or rather, how they make decisions during operation. This work on dynamic compute allocation represents a subtle but significant shift, acknowledging that simply throwing more processing power at a problem isn’t always the answer, particularly when agents are tasked with complex, multi-step reasoning.

For years, the field has grappled with the challenge of compounding errors in long-horizon tasks; a small mistake early on can derail an entire process. What distinguishes this research is the elegant simplicity of its core insight: not all decisions require the same level of scrutiny. By monitoring the agent’s own internal confidence, measured through the distribution of its ‘votes’ , researchers have created a system that intelligently allocates computational resources only when genuine uncertainty exists.

This is a departure from uniform scaling, where every step receives the same boost, and it demonstrably improves performance while simultaneously reducing costs. The discovery of a bimodal entropy distribution, with a large proportion of steps exhibiting strong consensus, is particularly compelling. However, the limitations of relying solely on internal confidence signals should not be overlooked.

The example of the arbiter overriding a near-unanimous decision highlights the risk of introducing errors when intervention isn’t warranted. Furthermore, the current work focuses on web-based agents; extending these findings to other domains, such as robotics or game playing, will require further investigation.

Looking ahead, we can anticipate a growing emphasis on ‘selective attention’ mechanisms within AI systems. This isn’t just about efficiency; it’s about building agents that are more robust, more interpretable, and ultimately, more trustworthy. The next step may involve combining these internal confidence signals with external sources of information, creating a hybrid approach that leverages both self-assessment and environmental feedback.

👉 More information
🗞 Agentic Test-Time Scaling for WebAgents
🧠 ArXiv: https://arxiv.org/abs/2602.12276

Rohail T.

Rohail T.

As a quantum scientist exploring the frontiers of physics and technology. My work focuses on uncovering how quantum mechanics, computing, and emerging technologies are transforming our understanding of reality. I share research-driven insights that make complex ideas in quantum science clear, engaging, and relevant to the modern world.

Latest Posts by Rohail T.:

Integrated Quantum Systems Combine Sensing and Computation with Indefinite Order

Integrated Quantum Systems Combine Sensing and Computation with Indefinite Order

February 17, 2026
AI Swiftly Solves Complex Control Problems Previously Needing Intensive Computation

AI Swiftly Solves Complex Control Problems Previously Needing Intensive Computation

February 17, 2026
AI Learns from Raw Data Using Adaptable Noise Levels for Better Results

AI Learns from Raw Data Using Adaptable Noise Levels for Better Results

February 17, 2026