HIVE: A High-Priority Victim Cache for Accelerating GPU Memory Accesses

The victim cache was originally designed as a secondary cache to handle misses in the L1 data (L1D) cache in CPUs. However, this design is often sub-optimal for GPUs. Accessing the high-latency L1D cache and its victim cache can lead to significant latency overhead, severely degrading the performanc...

Full description

Saved in:
Bibliographic Details
Published in:2025 62nd ACM/IEEE Design Automation Conference (DAC) pp. 1 - 7
Main Authors: Tang, Yuhan, Zhang, Jianmin, Ma, Sheng, Li, Tiejun, Li, Hanqing, Luo, Shengbai, Tang, Jixuan, Wu, Lizhou
Format: Conference Proceeding
Language:English
Published: IEEE 22.06.2025
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:The victim cache was originally designed as a secondary cache to handle misses in the L1 data (L1D) cache in CPUs. However, this design is often sub-optimal for GPUs. Accessing the high-latency L1D cache and its victim cache can lead to significant latency overhead, severely degrading the performance of certain applications. We introduce HIVE, a high-priority victim cache designed to accelerate GPU memory accesses. HIVE handles memory requests first, before they reach the L1D cache. Our experimental results show that HIVE achieves an average performance improvement of \mathbf{7 7. 1 \%} and \mathbf{2 1. 7 \%} compared to the baseline and the state-of-the-art architecture, respectively.
DOI:10.1109/DAC63849.2025.11133338