AI ‘quorum’ Speeds up Decisions and Boosts Prediction Accuracy to 70.60%

Researchers are tackling the challenge of reducing processing time in artificial intelligence systems, an issue particularly relevant for real-time applications. Matteo Gambella, Fabrizio Pittorino, and Giuliano Casale, from Politecnico di Milano and Imperial College London, alongside Manuel Roveri et al., present SQUAD (Scalable Quorum Adaptive Decisions), a novel inference scheme that combines early-exit networks with distributed ensemble learning. This work is significant because it moves beyond reliance on potentially flawed single-model confidence thresholds, instead employing a quorum-based voting system to achieve statistically robust early exits. By incrementally collecting predictions and halting computation when a consensus is reached, SQUAD demonstrably improves accuracy, by as much as 5.95% over current dynamic solutions, and reduces inference latency by up to 70.60% compared to static ensembles, offering a substantial advancement in efficient AI decision-making.

Scalable inference via quorum-based consensus and diverse early-exit learners enables efficient and robust predictions

Scientists have developed a new inference scheme called SQUAD (Scalable Quorum Adaptive Decisions) that integrates early-exit mechanisms with distributed ensemble learning, significantly improving uncertainty estimation and reducing inference time. The research addresses the unreliability of standard early-exit networks which typically rely on single-model confidence thresholds prone to calibration issues.
Unlike traditional methods, SQUAD employs a quorum-based stopping criterion, collecting intermediate predictions from early-exit learners incrementally based on computational complexity. Computation halts when a statistically significant consensus is reached, providing a more robust and efficient approach to early prediction.

To maximise the effectiveness of this voting mechanism, the team also introduced QUEST (Quorum Technique), a Neural Architecture Search method designed to select early-exit learners with optimised hierarchical diversity. This ensures that learners are complementary at each intermediate layer, enhancing the reliability of the quorum-based decision-making process.

By searching for diversity not just at the final output but at every early exit gate, QUEST avoids reinforcing incorrect decisions that might occur with biased learners. This innovative approach yields statistically robust early exits, improving test accuracy by up to 5.95% compared to state-of-the-art dynamic solutions with a comparable computational cost.

Experiments demonstrate that SQUAD reduces inference latency by up to 70.60% compared to static ensembles, while maintaining good accuracy. The study was conducted using datasets including CIFAR-10, CIFAR-100, and ImageNet16-120, showcasing the method’s versatility and performance across different image recognition tasks.

This consensus-driven approach offers a unique solution to the accuracy-efficiency dilemma, paving the way for deployment in resource-constrained environments like IoT sensors and edge computing platforms where low latency is paramount. The source code for QUEST has been released to the scientific community to facilitate further research and reproducibility.

Scalable Quorum Adaptive Decisions and Neural Architecture Search for Diverse Ensemble Learning enable robust and efficient model creation

Scientists introduced SQUAD, a Scalable Quorum Adaptive Decisions system, to integrate early-exit mechanisms with distributed ensemble learning and improve uncertainty estimation during inference. The study pioneered a quorum-based stopping criterion, collecting intermediate predictions from early-exit learners incrementally based on computational complexity.

Computation halts at an exit layer when a statistically significant consensus is reached amongst the learners, unlike conventional methods reliant on single-model confidence thresholds. Researchers developed QUEST, a Neural Architecture Search method, to select early-exit learners exhibiting optimized hierarchical diversity at each intermediate layer.

This technique ensures complementary predictions, preventing reinforcement of errors at early layers and maximizing the efficacy of the voting mechanism. The QUEST method actively searches for a learner set that maximizes diversity, differing from traditional NAS approaches focused solely on individual accuracy or output diversity.

Experiments employed a dynamic committee of early-exit learners, querying them incrementally in order of computational cost until a majority consensus was achieved. A t-test was then performed to ensure the statistical robustness of the consensus before halting computation vertically, providing a statistically sound early exit decision.

This consensus-driven approach yielded a test accuracy improvement of up to 5.95% compared to state-of-the-art dynamic solutions, while maintaining comparable computational cost. The system delivers a reduction in inference latency of up to 70.60% compared to static ensembles, demonstrating a significant performance gain.

This method achieves robust early exits by mitigating the calibration issues inherent in single-model early-exit neural networks, offering a substantial advancement in efficient deep learning inference. The work demonstrates how collective confidence, derived from an ensemble, enhances the reliability of early-exit decisions and unlocks substantial performance improvements.

Quorum-based consensus and hierarchical diversity enhance efficient inference accuracy in complex systems

Scientists have developed SQUAD, a new inference scheme integrating early-exit mechanisms with distributed ensemble learning to improve uncertainty estimation and reduce inference time. Experiments revealed that SQUAD employs a quorum-based stopping criterion, collecting intermediate predictions incrementally by computational complexity until a statistically significant consensus is reached.

The team measured a substantial improvement in test accuracy, achieving up to a 5.95% increase compared to state-of-the-art dynamic solutions with comparable computational cost. Results demonstrate a reduction in inference latency of up to 70.60% compared to static ensembles, while maintaining good accuracy.

To maximise the efficacy of the voting mechanism, researchers introduced QUEST, a method for selecting early-exit learners with optimised hierarchical diversity, ensuring complementary learning at each intermediate layer. Measurements confirm that this consensus-driven approach yields statistically robust early exits, enhancing the reliability of predictions.

Data shows that QUEST effectively combines the efficiency of early exits with the robustness of ensembles through a novel Neural Architecture Search (NAS) strategy. The work on CIFAR-10, CIFAR-100, and ImageNet16-120 demonstrates that QUEST achieves a superior balance between accuracy and computational cost.

Specifically, the team recorded a 70.60% improvement in inference time compared to static ensembles. The study also measured a 5.95% accuracy increase over standard confidence-based Early Exit Neural Networks (EENNs) at comparable latency. Researchers utilised Pairwise Predictive Disagreement (PPD) to quantify the diversity necessary for effective ensembles, with higher PPD values indicating greater diversity in predictive behaviour.

The ECE score, used to measure calibration reliability, was significantly improved through the ensemble approach. The breakthrough delivers a unified framework, termed QUEST, which effectively combines efficiency with robustness. Tests prove that the SVGD-RD mechanism successfully updates architectural configurations, balancing exploitation of high-accuracy regions with repulsive forces to ensure diversity. The source code for QUEST has been released to the scientific community to facilitate comparisons and reproducibility.

Quorum-based consensus and hierarchical diversity optimise early-exit network performance significantly

Researchers have developed SQUAD, a novel framework integrating early-exit networks with distributed ensemble learning to improve inference speed and accuracy. Traditional early-exit methods rely on single-model confidence thresholds, which can be unreliable due to calibration issues. SQUAD addresses this by employing a quorum-based stopping criterion, collecting intermediate predictions from early-exit learners incrementally, and halting computation when a statistically significant consensus is reached.

To further enhance performance, the authors also introduced QUEST, a technique for selecting early-exit learners that prioritises hierarchical diversity. This ensures the learners are complementary at each layer, maximising the effectiveness of the voting mechanism. Experiments conducted on CIFAR-10, CIFAR-100, and ImageNet16-120 demonstrate that SQUAD achieves state-of-the-art results, improving test accuracy by up to 5.95% compared to dynamic solutions while reducing inference latency by up to 70.60% relative to static ensembles.

The authors acknowledge that the t-test used for early exiting is more stringent than relying on mean confidence, potentially impacting accuracy and false-positive rates. Future research could explore alternative statistical tests or adaptive thresholding methods to optimise the balance between speed and accuracy.

👉 More information
🗞 SQUAD: Scalable Quorum Adaptive Decisions via ensemble of early exit neural networks
🧠 ArXiv: https://arxiv.org/abs/2601.22711

Rohail T.

Rohail T.

As a quantum scientist exploring the frontiers of physics and technology. My work focuses on uncovering how quantum mechanics, computing, and emerging technologies are transforming our understanding of reality. I share research-driven insights that make complex ideas in quantum science clear, engaging, and relevant to the modern world.

Latest Posts by Rohail T.:

Abstract geometric manifold with quantum gate symbols represented as smooth curved paths across surface

Quantum Gates Mapped to Predictable Geometric Space

February 26, 2026
Logical qubit core protected by layered stabilizer shields, error ripples absorbed at outer boundary, central state glowing steadily with high coherence

Quantum Error Framework Boosts Logical State Fidelity

February 26, 2026
Qubit array with fewer measurement beams scanning across it, bright efficient readout node at center, minimal energy pulses compared to background

Quantum Computers Cut Measurement Costs with New Method

February 25, 2026