Exploring the Diversity of Multiple Job Deployments over GPUs for Efficient Resource Sharing
Citations

WEB OF SCIENCE

4
Citations

SCOPUS

5

초록

Graphic Processing Units (GPUs) are gradually becoming mainstream computing resource for efficient execution of applications both on-premises and in the cloud. Currently however, most HPC applications are unable to leverage the large computing capabilities they provide leading to issues of resource under-utilization. Various GPU sharing approaches have been proposed which leverage either software or hardware level mechanisms like MPS or MIG in NVIDIA GPUs. However, combining both the software and hardware level technologies in an effort to mitigate resource under-utilization issues is yet to be fully explored. In this paper, we conduct a case study on scheduling memory intensive and compute intensive applications on an NVIDIA A30 GPU. We compare the performance when using only hardware level sharing mechanisms and when using both hardware and software level mechanisms. We observed that by combining both mechanisms, we improved total execution times by up to 14% for a single run whilst improving peak bandwidth utilization by about 39% for SCAN application. © 2024 IEEE.

키워드

ConcurrencyMIGMPSResource under-utilizationSpatial Sharing
제목
Exploring the Diversity of Multiple Job Deployments over GPUs for Efficient Resource Sharing
저자
Adufu, TheodoraHa, JiwonKim, Yoonhee
DOI
10.1109/ICOIN59985.2024.10572198
발행일
2024-01
유형
Proceedings Paper
저널명
International Conference on Information Networking
페이지
777 ~ 782