Abstract
The process of training and serving deep learning (DL) models is computationally expensive, mandating the use of powerful and expensive accelerators such as GPUs and TPUs. Furthermore, the prevalence of GPUs in data centers today motivate developing database systems that can leverage the available GPU resources. Both the latency of DL tasks and database queries and high utilization of these accelerators depend on how efficiently we can move the data to the accelerators. Given today’s dataset sizes, fitting everything in GPU or even CPU memory is not always feasible or can be expensive. The I/O path while fetching the data from disks, however, still dominantly relies on CPUs.
In this work, we take a step toward understanding today’s landscape for optimizing the I/O path for reading data to GPUs from disks, with a focus on SSDs. First, we review the prominent technologies that target GPU-centric storage accesses. Then, we dive deeper into BaM, as the state-of-the-art method for GPU-centric storage, and evaluate its performance in comparison to the state-of-theart CPU-centric storage interface SPDK. Our results demonstrate that while BaM is able to match the performance of SPDK without involving CPUs on the I/O path, this comes at the cost of a very high GPU use. Finally, we highlight future research directions to enable an I/O path that is both efficient and easy-to-adopt for data-intensive systems that use GPUs.
In this work, we take a step toward understanding today’s landscape for optimizing the I/O path for reading data to GPUs from disks, with a focus on SSDs. First, we review the prominent technologies that target GPU-centric storage accesses. Then, we dive deeper into BaM, as the state-of-the-art method for GPU-centric storage, and evaluate its performance in comparison to the state-of-theart CPU-centric storage interface SPDK. Our results demonstrate that while BaM is able to match the performance of SPDK without involving CPUs on the I/O path, this comes at the cost of a very high GPU use. Finally, we highlight future research directions to enable an I/O path that is both efficient and easy-to-adopt for data-intensive systems that use GPUs.
| Originalsprog | Engelsk |
|---|---|
| Titel | Proceedings of the 21st International Workshop on Data Management on New Hardware, DaMoN 2025, Berlin, Germany, June 22-27, 2025 |
| Antal sider | 9 |
| Udgivelsessted | New York |
| Forlag | Association for Computing Machinery |
| Publikationsdato | 22 jun. 2025 |
| Sider | 1-9 |
| Artikelnummer | 3 |
| ISBN (Trykt) | 979-8-4007-1940-0 |
| DOI | |
| Status | Udgivet - 22 jun. 2025 |
| Begivenhed | Management of Data - Berlin, Berlin, Tyskland Varighed: 22 jun. 2025 → 27 jun. 2025 Konferencens nummer: 21 https://2025.sigmod.org/ |
Konference
| Konference | Management of Data |
|---|---|
| Nummer | 21 |
| Lokation | Berlin |
| Land/Område | Tyskland |
| By | Berlin |
| Periode | 22/06/2025 → 27/06/2025 |
| Internetadresse |
Emneord
- Deep Learning
- GPU Databases
- GPU-Initiated I/O
- NVMe SSDs