The increasing demand for real-time data processing fuels the growth of Edge AI, but deploying these systems presents considerable challenges in managing energy consumption and operational time. Zhiyuan Zhai, Wei Ni, and Xin Wang investigate methods to optimise learning performance within Edge AI systems, considering the interplay between data acquisition, computation, and communication. Their work models the energy and time demands of each process, rigorously analysing how factors like data volume and training duration impact overall learning. By formulating a system-wide optimisation problem, the researchers develop algorithms that maximise learning performance while adhering to strict time and energy constraints, offering valuable insights for designing efficient Edge AI systems suitable for real-world applications.
The research team modeled time and energy consumption across these processes, enabling a rigorous convergence analysis to quantify the impact of key parameters on learning performance. This analysis revealed how the amount of collected data and the number of training rounds directly affect the system’s ability to learn effectively.
Communication Efficiency in Federated Learning
Federated Learning has emerged as a powerful technique for training machine learning models on decentralized data sources, such as mobile devices and sensors. A significant body of research focuses on improving the communication efficiency of this process, which is critical for bandwidth-constrained wireless environments. Scientists are actively exploring techniques to reduce the amount of data exchanged between devices and a central server, including model compression, quantization, and sparsification. These methods reduce model size and communication overhead, enabling faster and more efficient training.
Researchers also investigate over-the-air computation, which leverages the superposition property of wireless channels to perform model aggregation directly in the air, further reducing communication latency. Addressing data heterogeneity is another key research area. Real-world data is often non-independent and non-identically distributed (non-IID) across different clients, meaning that each device may have a unique data distribution. Scientists are developing algorithms that can effectively handle this heterogeneity, ensuring that the model learns effectively from all available data. Privacy and security are also paramount concerns, with researchers exploring techniques such as differential privacy to protect user data and developing methods to defend against potential attacks.
Incentive mechanisms are being designed to motivate clients to participate in federated learning, ensuring fair compensation for their contributions. Furthermore, scientists are investigating asynchronous and decentralized federated learning, improving scalability and robustness. Recent advances explore the application of federated learning to large language models (LLMs), enabling distributed and privacy-preserving training of these powerful models. They then explored solutions to this problem in both homogeneous and heterogeneous device scenarios, developing low-complexity algorithms based on one-dimensional search and alternating optimization. These algorithms jointly optimize data collection time and the number of training rounds, seeking the most efficient balance between data acquisition and model training. The research provides a detailed representation of the system, enabling more effective resource allocation and reduced energy footprint.
Balanced Learning Under Resource Constraints
Through convergence analysis, the team quantified these relationships, revealing that increased data volume and, initially, more training rounds generally improve model performance. However, the analysis also demonstrates diminishing returns as training progresses, highlighting the need for balanced resource allocation. The work establishes a principled performance surrogate, guiding decisions about data collection and training schedules. Researchers validated their convergence analysis through simulations, demonstrating its accuracy in predicting learning performance under realistic conditions. While acknowledging that strict assumptions of Lipschitz continuity, strong convexity, and bounded gradients may not always hold for complex deep learning models, the team showed the derived convergence bound remains a reliable indicator of achievable performance even with relaxed assumptions.
👉 More information
🗞 Learning Performance Optimization for Edge AI System with Time and Energy Constraints
🧠 ArXiv: https://arxiv.org/abs/2511.06806
