Timezone: »
DNNs have revolutionized across a wide range of applications, such as image classification, speech recognition and robotics control. As DNN models become more computationally expensive to train, parallel execution with multiple accelerators (e.g. GPUs) is adopted. System efficiency is a big issue when scaling out. However, as computation power increases, GPUs are under-utilized mainly due to limited local memory size. To address this memory bound, we present Wavelet, an efficient and generic approach that can fully utilize all the available on-device memory among GPUs involved in the distributed training job. Wavelet achieves near optimal on-device memory usage by adopting a simple scheduling scheme called Tick-Tock, which interleaves waves of peak memory usage among the accelerators. Evaluations on a variety of DNN models and tasks show that, Wavelet trains models up to 6.7x faster than commonly used parallelism techniques.
Author Information
Guanhua Wang (UC Berkeley)
I am a Ph.D. student in the AMPLab / RISELab, at UC Berkeley, advised by Prof. Ion Stoica.
Kehan Wang (University of California, Berkeley)
Kenan Jiang (University of California, Berkeley)
XIANGJUN LI (UC Berkeley)
Ion Stoica (UC Berkeley)
Related Events (a corresponding poster, oral, or spotlight)
-
2021 Oral: Wavelet: Efficient DNN Training with Tick-Tock Scheduling »
Thu. Apr 8th 11:00 -- 11:20 PM Room
More from the Same Authors
-
2021 Poster: sensAI: ConvNets Decomposition via Class Parallelism for Fast Inference on Live Data »
Guanhua Wang · Zhuang Liu · Brandon Hsieh · Siyuan Zhuang · Joseph Gonzalez · Trevor Darrell · Ion Stoica -
2021 Oral: sensAI: ConvNets Decomposition via Class Parallelism for Fast Inference on Live Data »
Guanhua Wang · Zhuang Liu · Brandon Hsieh · Siyuan Zhuang · Joseph Gonzalez · Trevor Darrell · Ion Stoica -
2020 Oral: Blink: Fast and Generic Collectives for Distributed ML »
Guanhua Wang · Shivaram Venkataraman · Amar Phanishayee · Nikhil Devanur · Jorgen Thelin · Ion Stoica -
2020 Poster: Blink: Fast and Generic Collectives for Distributed ML »
Guanhua Wang · Shivaram Venkataraman · Amar Phanishayee · Nikhil Devanur · Jorgen Thelin · Ion Stoica