Topology-aware GPU scheduling for learning workloads in cloud environments
Recent advances in hardware, such as systems with multiple GPUs and their availability in the cloud, are enabling deep learning in various domains including health care, autonomous vehicles, and Internet of Things. Multi-GPU systems exhibit complex connectivity among GPUs and between GPUs and CPUs....
Gespeichert in:
| Veröffentlicht in: | International Conference for High Performance Computing, Networking, Storage and Analysis (Online) S. 1 - 12 |
|---|---|
| Hauptverfasser: | , , , , |
| Format: | Tagungsbericht Verlag |
| Sprache: | Englisch |
| Veröffentlicht: |
New York, NY, USA
ACM
12.11.2017
Association for Computing Machinery (ACM) |
| Schriftenreihe: | ACM Conferences |
| Schlagworte: | |
| ISBN: | 9781450351140, 145035114X |
| ISSN: | 2167-4337 |
| Online-Zugang: | Volltext |
| Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
| Zusammenfassung: | Recent advances in hardware, such as systems with multiple GPUs and their availability in the cloud, are enabling deep learning in various domains including health care, autonomous vehicles, and Internet of Things. Multi-GPU systems exhibit complex connectivity among GPUs and between GPUs and CPUs. Workload schedulers must consider hardware topology and workload communication requirements in order to allocate CPU and GPU resources for optimal execution time and improved utilization in shared cloud environments.
This paper presents a new topology-aware workload placement strategy to schedule deep learning jobs on multi-GPU systems. The placement strategy is evaluated with a prototype on a Power8 machine with Tesla P100 cards, showing speedups of up to ≈1.30x compared to state-of-the-art strategies; the proposed algorithm achieves this result by allocating GPUs that satisfy workload requirements while preventing interference. Additionally, a large-scale simulation shows that the proposed strategy provides higher resource utilization and performance in cloud systems. |
|---|---|
| ISBN: | 9781450351140 145035114X |
| ISSN: | 2167-4337 |
| DOI: | 10.1145/3126908.3126933 |

