TY - GEN
T1 - Huffman Coding Based Encoding Techniques for Fast Distributed Deep Learning
AU - Gajjala, Rishikesh R.
AU - Banchhor, Shashwat
AU - Abdelmoniem, Ahmed M.
AU - Dutta, Aritra
AU - Canini, Marco
AU - Kalnis, Panos
N1 - KAUST Repository Item: Exported on 2020-12-02
PY - 2020/12
Y1 - 2020/12
N2 - Distributed stochastic algorithms, equipped with gradient compression techniques, such as codebook quantization, are becoming increasingly popular and considered state-of-the-art in training large deep neural network (DNN) models. However, communicating the quantized gradients in a network requires efficient encoding techniques. For this, practitioners generally use Elias encoding-based techniques without considering their computational overhead or data-volume. In this paper, based on Huffman coding, we propose several lossless encoding techniques that exploit different characteristics of the quantized gradients during distributed DNN training. Then, we show their effectiveness on 5 different DNN models across three different data-sets, and compare them with classic state-of-the-art Elias-based encoding techniques. Our results show that the proposed Huffman-based encoders (i.e., RLH, SH, and SHS) can reduce the encoded data-volume by up to 5.1×, 4.32×, and 3.8×, respectively, compared to the Elias-based encoders.
AB - Distributed stochastic algorithms, equipped with gradient compression techniques, such as codebook quantization, are becoming increasingly popular and considered state-of-the-art in training large deep neural network (DNN) models. However, communicating the quantized gradients in a network requires efficient encoding techniques. For this, practitioners generally use Elias encoding-based techniques without considering their computational overhead or data-volume. In this paper, based on Huffman coding, we propose several lossless encoding techniques that exploit different characteristics of the quantized gradients during distributed DNN training. Then, we show their effectiveness on 5 different DNN models across three different data-sets, and compare them with classic state-of-the-art Elias-based encoding techniques. Our results show that the proposed Huffman-based encoders (i.e., RLH, SH, and SHS) can reduce the encoded data-volume by up to 5.1×, 4.32×, and 3.8×, respectively, compared to the Elias-based encoders.
UR - http://hdl.handle.net/10754/666175
UR - https://dl.acm.org/doi/10.1145/3426745.3431334
U2 - 10.1145/3426745.3431334
DO - 10.1145/3426745.3431334
M3 - Conference contribution
SN - 9781450381826
BT - Proceedings of the 1st Workshop on Distributed Machine Learning
PB - ACM
ER -