MGPU-TSM: A Multi-GPU System with Truly Shared Memory (2008.02300v2)
Abstract: The sizes of GPU applications are rapidly growing. They are exhausting the compute and memory resources of a single GPU, and are demanding the move to multiple GPUs. However, the performance of these applications scales sub-linearly with GPU count because of the overhead of data movement across multiple GPUs. Moreover, a lack of hardware support for coherency exacerbates the problem because a programmer must either replicate the data across GPUs or fetch the remote data using high-overhead off-chip links. To address these problems, we propose a multi-GPU system with truly shared memory (MGPU-TSM), where the main memory is physically shared across all the GPUs. We eliminate remote accesses and avoid data replication using an MGPU-TSM system, which simplifies the memory hierarchy. Our preliminary analysis shows that MGPU-TSM with 4 GPUs performs, on average, 3.9x? better than the current best performing multi-GPU configuration for standard application benchmarks.
- Saiful A. Mojumder (3 papers)
- Yifan Sun (183 papers)
- Leila Delshadtehrani (4 papers)
- Yenai Ma (3 papers)
- Trinayan Baruah (3 papers)
- José L. Abellán (10 papers)
- John Kim (23 papers)
- David Kaeli (25 papers)
- Ajay Joshi (25 papers)