FRUGAL: Pushing GPU Applications beyond Memory Limits
GPUs power modern scientific and AI applications, but their limited memory capacity restricts scalability. Buying GPUs with larger HBM is prohibitively expensive and still bounded by market limits. Existing solutions either exploit application-specific knowledge through out-of-core techniques, which lack generality, or rely on system-level page faulting, which is transparent but inefficient. We propose FRUGAL, an application-agnostic framework and methodology that reduces GPU memory footprint while sustaining high performance. FRUGAL formulates memory management as an optimization over an application’s execution graph, encompassing prefetching, kernel execution, and offloading. Using static analysis and profiling, FRUGAL applies a two-phase scheduling and migration strategy, solving an otherwise intractable optimization efficiently. Evaluations on Tiled Cholesky Decomposition, Tiled LU Decomposition, Tiny-CUDA-NN, and QuEST show that FRUGAL significantly reduces maximum GPU memory usage by 80.21%, 80.20%, 64.75% and 60.86% with only a geometric mean of 28.31% slowdown. FRUGAL allows applications to exceed hardware-imposed limits, and maintains strong performance scalability beyond existing GPU memory constraints, without additional hardware cost.
Mon 2 FebDisplayed time zone: Hobart change
14:10 - 15:30 | |||
14:10 20mTalk | Flow-Graph-Aware Tiling and Rescheduling for Memory-Efficient On-Device Inference Main Conference Pre-print | ||
14:30 20mTalk | VFlatten: Selective Value-Object Flattening using Hybrid Static and Dynamic Analysis Main Conference Arjun H. Kumar IIT Mandi, Bhavya Hirani SVNIT, Surat, Hang Shao IBM, Tobi Ajila IBM, Vijay Sundaresan IBM Canada, Daryl Maier IBM Canada, Manas Thakur IIT Bombay Pre-print Media Attached | ||
14:50 20mTalk | FRUGAL: Pushing GPU Applications beyond Memory Limits Main Conference Lingqi Zhang RIKEN RCCS, Tengfei Wang Google Cloud, Jiajun Huang University of California, Riverside, Chen Zhuang Tokyo Institute of Technology, Riken Center for Computational Science, Ivan Ivanov Institute of Science Tokyo, Peng Chen RIKEN RCCS, Toshio Endo , Mohamed Wahib RIKEN Center for Computational Science Pre-print | ||
15:10 20mTalk | Automatic Data Enumeration for Fast Collections Main Conference Pre-print Media Attached | ||