Deep learning (DL) systems have transformed various industries, but their high resource demands and security concerns hinder widespread adoption. Researchers propose a new approach called SGX-O-MPSS, combining task-based programming with trusted execution environments to improve performance and security. This innovative solution leverages asynchronous task parallelism and Intel Software Guard Extensions (SGX), achieving a 94% gain in speedup while ensuring sensitive data protection.
By utilizing multiple CPU cores efficiently, SGX-O-MPSS reduces training time and energy consumption, making it an attractive solution for large-scale DL applications. The benefits of this approach include improved performance, enhanced security, and increased scalability. Intel SGX provides a secure environment for executing sensitive code, protecting data from unauthorized access.
As researchers continue to explore the potential of SGX-O-MPSS, we can expect significant advancements in deep learning. This technology has vast applications, including natural language processing, computer vision, speech recognition, and self-driving vehicles. With promising results from early evaluations and a growing community of researchers working towards improving performance and security, the future of DL systems looks bright.
Deep learning (DL) systems have revolutionized various domains, including speech recognition, face detection, self-driving vehicles, genetic sequence modeling, and natural language processing. These systems achieve high accuracy on non-trivial prediction problems at the cost of high resource demands, particularly during training. The ability to efficiently execute applications while offering security guarantees is critical for several domains.
DL systems are complex and challenging to optimize due to three main factors: performance, security guarantees, and energy requirements. Training DL models requires iterating over many input data sets in batches, which can be time-consuming and energy-intensive. For example, training natural language understanding models on a cluster of commodity hardware requires 115 hours. Inference, on the other hand, has lower resource demands when executed on edge devices, where inference latency is critical.
DL systems must also respect strong security requirements, particularly regarding data access. Under certain circumstances, these systems must resist compromised cloud providers. The energy footprint is a major concern due to the long training time and its direct relation to the minimum accuracy required by the system.
To address these challenges, researchers have proposed combining asynchronous task parallelism with hardware-assisted protection mechanisms, such as trusted execution environments (TEEs). TEEs offer a practical privacy-preserving solution available in both private and public data centers. One such approach is SGX-O-MPSS, which combines a task-based programming model with TEEs.
The Need for Secure Deep Learning
DL systems are particularly sensitive to data access, as they rely on users’ sensitive information. Preserving privacy is critical when training or tuning DL parameters. Hardware-assisted protection mechanisms, such as TEEs, offer a practical solution to this problem. These mechanisms provide a secure environment for executing applications, ensuring that sensitive data remains protected.
The use of TEEs in DL systems has gained significant attention in recent years. Researchers have proposed various approaches to combine asynchronous task parallelism with TEEs, aiming to improve performance while maintaining security guarantees. One such approach is SGX-O-MPSS, which supports asynchronous task parallelism and hardware heterogeneity by using data dependencies between tasks.
The SGX-O-MPSS Approach
SGX-O-MPSS combines a task-based programming model with TEEs, offering a new approach to secure DL. This approach enables asynchronous task parallelism and hardware heterogeneity, making it an attractive solution for DL applications. By using code annotations to specify data dependencies between tasks, SGX-O-MPSS simplifies the process of exploiting parallelism in DL systems.
The evaluation of SGX-O-MPSS via several microbenchmarks and state-of-the-art DL applications and datasets has shown promising results. The approach achieves a 94% gain in speedup while offering additional security guarantees. This makes SGX-O-MPSS an attractive solution for DL applications that require both performance and security.
Evaluating SGX-O-MPSS
The evaluation of SGX-O-MPSS involved several microbenchmarks and state-of-the-art DL applications and datasets, including YOLO and MNIST. The results showed a significant gain in speedup, with an average improvement of 94%. This demonstrates the effectiveness of SGX-O-MPSS in improving performance while maintaining security guarantees.
The evaluation also highlighted the benefits of using TEEs in DL systems. By providing a secure environment for executing applications, TEEs ensure that sensitive data remains protected. This is particularly important in DL systems, where users’ sensitive information is used to train models.
Conclusion
Combining asynchronous task parallelism with hardware-assisted protection mechanisms, such as TEEs, offers a promising solution for secure DL. SGX-O-MPSS is one such approach that supports asynchronous task parallelism and hardware heterogeneity by using data dependencies between tasks. The evaluation of SGX-O-MPSS via several microbenchmarks and state-of-the-art DL applications and datasets has shown promising results, with an average improvement of 94% in speedup.
Publication details: “Combining Asynchronous Task Parallelism and Intel SGX for Secure Deep Learning : (Practical Experience Report)”
Publication Date: 2024-04-08
Authors: Isabelly Rocha, Pascal Felber, Xavier Martorel, Marcelo Pasin, et al.
Source:
DOI: https://doi.org/10.1109/edcc61798.2024.00029
