Research demonstrates optimisation techniques impacting large language model tuning, assessed on H100 GPUs. Parallelisation strategies—including distributed data and hardware processing—were analysed for image recognition. Results reveal task-dependent iteration times, VRAM usage, and memory transfer rates during multiprocessor operation. Performance was evaluated using Direct Preference Optimisation (DPO), Low-Rank Adaptation (LoRA), Quantised LoRA (QLoRA) and Quantisation Aware Training (QAT).
The increasing computational demands of machine learning necessitate careful optimisation of both algorithms and hardware. Achieving peak performance requires a nuanced understanding of how parallel processing strategies interact with specific architectures and model parameters. Researchers at the Poznan Supercomputing and Networking Centre and the Czestochowa University of Technology have investigated these interactions, focusing on multi-GPU configurations and the impact of various optimisation techniques. In their paper, “Profiling and optimisation of multi-card GPU machine learning jobs”, Marcin Lawenda, Łukasz Szustak, Kyrylo Khloponin, and Krzesimir Samborski present a detailed analysis of key performance indicators, parallelisation strategies – including distributed data and hardware processing – and the effects of techniques such as DPO (Direct Preference Optimisation), LoRA (Low-Rank Adaptation), QLoRA (Quantised LoRA), and QAT (Quantisation Aware Training) on large language model tuning, utilising the modern H100 GPU architecture for their experiments.
Optimizing Machine Learning Models for Performance and Efficiency
The escalating demand for artificial intelligence necessitates a concurrent focus on computational efficiency alongside model performance. This research investigates diverse optimization techniques, evaluating their impact on key performance indicators across varied hardware and software configurations to provide a comprehensive understanding of their trade-offs. Our investigations encompass parallelization strategies for image recognition and detailed examinations of performance-enhancing methods including Direct Preference Optimization (DPO), Low-Rank Adaptation (LoRA), Quantized LoRA (QLoRA), and Quantization-Aware Training (QAT) as applied to large language model tuning.
We initiated our investigations by exploring parallelization strategies, recognising their potential to reduce training times for computationally intensive tasks such as image recognition. We implemented and compared data parallelism – distributing data across multiple processing units – and model parallelism – distributing the model itself. Performance was carefully measured across different hardware configurations and dataset sizes. Results demonstrate that the optimal strategy is heavily dependent on task characteristics and available resources, necessitating careful consideration and tuning.
We then focused on performance improvement techniques specifically designed for large language models, acknowledging the challenges posed by their scale and complexity. We meticulously analysed DPO, LoRA, QLoRA, and QAT, evaluating their impact on model accuracy, training speed, and memory footprint. Each technique offers distinct advantages and disadvantages, and the optimal choice depends on the specific application requirements. LoRA and QLoRA effectively reduce the number of trainable parameters, decreasing memory consumption and accelerating training without substantial loss of accuracy.
Detailed experiments quantified the impact of these techniques on key performance indicators, including iteration time, VRAM utilization, and memory transfer rates. QLoRA consistently achieved the lowest memory footprint, making it particularly suitable for resource-constrained environments. DPO demonstrated effectiveness in improving model alignment with human preferences, leading to more natural and coherent text generation. These trade-offs were carefully measured to provide valuable insights into the strengths and weaknesses of each technique.
We rigorously evaluated our findings on the NVIDIA H100 GPU architecture, a state-of-the-art platform for deep learning research and deployment. Initial testing utilised the Kaggle MNIST dataset, providing a robust baseline for evaluating generalizability. Experiments were then extended to more complex datasets, including ImageNet, to further validate our results.
Our research demonstrates that optimizing machine learning models is not merely a matter of improving performance but also of reducing computational cost and environmental impact. By leveraging techniques like parallelization, quantization – reducing the precision of numerical representations – and parameter reduction, we can train and deploy models more efficiently, making AI more accessible and sustainable.
Looking ahead, we plan to explore adaptive optimization strategies that dynamically adjust to the evolving demands of the training process. We envision systems that automatically select the optimal techniques and parameters based on data characteristics and available hardware. We also plan to investigate novel quantization methods and hardware-aware optimization algorithms that can further improve efficiency.
Our findings have significant implications for a wide range of applications, including natural language processing, computer vision, and robotics. By reducing the computational cost of training and deploying machine learning models, we can broaden access to these technologies and enable new and innovative applications.
We are committed to disseminating our findings through publications, presentations, and open-source software. Sharing our knowledge and resources with the community is essential for accelerating progress in the field of machine learning.
Our research underscores the importance of a holistic approach to machine learning optimization, considering not only performance but also efficiency, sustainability, and accessibility. By embracing these principles, we can unlock the full potential of AI and create a more equitable and sustainable future. We are committed to continuing our research in this area and contributing to the advancement of the field.
👉 More information
🗞 Profiling and optimization of multi-card GPU machine learning jobs
🧠 DOI: https://doi.org/10.48550/arXiv.2505.22905
