Analyzing the Impact of DVFS on Performance and Energy of Parallel Applications on GPUs

  • Thiago dos S. Gonçalves UFRGS
  • Arthur F. Lorenzon UFRGS

Abstract


The parallel processing capability of Graphics Processing Units (GPUs) has made their use essential in accelerating artificial intelligence applications. With a strong presence of matrix multiplication in these applications, new strategies are needed to achieve better energy efficiency. In this way, we analyze the impact of cache metrics on a GPU and show a 19.87% difference in energy consumption with small gains in performance.
Keywords: Parallel and Distributed Algorithms, Evaluation, Measurement, and Performance Prediction, Heterogeneous Computing

References

Anzt, H., Tomov, S., and Dongarra, J. (2015). Energy efficiency and performance frontiers for sparse computations on gpu supercomputers. In Proceedings of the Sixth International Workshop on Programming Models and Applications for Multicores and Manycores, PMAM ’15, page 1–10, New York, NY, USA. Association for Computing Machinery.

Baji, T. (2017). GPU: the biggest key processor for AI and parallel processing. In Takehisa, K., editor, Symposium on Photomask and Next-Generation Lithography Mask Technology, volume 10454, page 1045406. International Society for Optics and Photonics, SPIE.

Fei, X., Li, K., Yang, W., and Li, K. (2020). Analysis of energy efficiency of a parallel aes algorithm for cpu-gpu heterogeneous platforms. Parallel Computing, 94-95:102621.

Jahanshahi, A., Sabzi, H. Z., Lau, C., and Wong, D. (2020). Gpu-nest: Characterizing energy efficiency of multi-gpu inference servers. IEEE Computer Architecture Letters, 19(2):139–142.

Jin, Z. and Vetter, J. S. (2023). A benchmark suite for improving performance portability of the sycl programming model. In 2023 IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS), pages 325–327.

Li, D., Chen, X., Becchi, M., and Zong, Z. (2016). Evaluating the energy efficiency of deep convolutional neural networks on cpus and gpus. In IEEE international conferences on big data and cloud computing, pages 477–484. IEEE.

Sharma, R., M, V., and Moharir, M. (2016). Revolutionizing machine learning algorithms using gpus. In CSITSS, pages 318–323.

Wang, Y., Karimi, M., Xiang, Y., and Kim, H. (2021). Balancing energy efficiency and real-time performance in gpu scheduling. In 2021 IEEE Real-Time Systems Symposium (RTSS), pages 110–122.
Published
2025-04-23
GONÇALVES, Thiago dos S.; LORENZON, Arthur F.. Analyzing the Impact of DVFS on Performance and Energy of Parallel Applications on GPUs. In: REGIONAL SCHOOL OF HIGH PERFORMANCE COMPUTING FROM SOUTHERN BRAZIL (ERAD-RS), 25. , 2025, Foz do Iguaçu/PR. Anais [...]. Porto Alegre: Sociedade Brasileira de Computação, 2025 . p. 17-20. ISSN 2595-4164. DOI: https://doi.org/10.5753/eradrs.2025.6824.