Deferred Dropout: An Algorithm-Hardware Co-Design DNN Training Method Provisioning Consistent High Activation Sparsity

Cited 2 time in webofscience Cited 0 time in scopus
  • Hit : 94
  • Download : 0
This paper proposes a deep neural network training method that provisions consistent high activation sparsity and the ability to adjust the sparsity. To improve training performance, prior work reduces the memory footprint for training by exploiting input activation sparsity which is observed due to the ReLU function. However, the previous approach relies solely on the inherent sparsity caused by the function, and thus the footprint reduction is not guaranteed. In particular, models for natural language processing tasks like BERT do not use the function, so the models have almost zero activation sparsity and the previous approach loses its efficiency. In this paper, a new training method, Deferred Dropout, and its hardware architecture are proposed. With the proposed method, input activations are dropped out after the conventional forward-pass computation. In contrast to the conventional dropout where activations are zeroed before forward-pass computation, the dropping timing is deferred until the completion of the computation. Then, the sparsified activations are compressed and stashed in memory. This approach is based on our observation that networks preserve training quality even if only a few high magnitude activations are used in the backward pass. The hardware architecture enables designers to exploit the trade-off between training quality and activation sparsity. Evaluation results demonstrate that the proposed method achieves 1.21-3.60x memory footprint reduction and 1.06-1.43x speedup on the TPUv3 architecture, compared to the prior work.
Publisher
IEEE/ACM
Issue Date
2021-11
Language
English
Citation

40th IEEE/ACM International Conference on Computer Aided Design (ICCAD)

ISSN
1933-7760
DOI
10.1109/ICCAD51958.2021.9643433
URI
http://hdl.handle.net/10203/290335
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 2 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0