AdaBlock: SGD with Practical Block Diagonal Matrix Adaptation for Deep Learning

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 79
  • Download : 0
We introduce ADABLOCK, a class of adaptive gradient methods that extends popular approaches such as ADAM by adopting the simple and natural idea of using block-diagonal matrix adaption to effectively utilize structural characteristics of deep learning architectures. Unlike other quadratic or blockdiagonal approaches, ADABLOCK has complete freedom to select block-diagonal groups, providing a wider trade-off applicable even to extremely high-dimensional problems. We provide convergence and generalization error bounds for ADABLOCK, and study both theoretically and empirically the impact of the block size on the bounds and advantages over usual diagonal approaches. In addition, we propose a randomized layer-wise variant of ADABLOCK to further reduce computations and memory footprint, and devise an efficient spectrum-clipping scheme for ADABLOCK to benefit from SGD's superior generalization performance. Extensive experiments on several deep learning tasks demonstrate the benefits of block diagonal adaptation compared to adaptive diagonal methods, vanilla SGD, as well as modified versions of full-matrix adaptation.
Publisher
JMLR-JOURNAL MACHINE LEARNING RESEARCH
Issue Date
2022-03
Language
English
Citation

International Conference on Artificial Intelligence and Statistics

ISSN
2640-3498
URI
http://hdl.handle.net/10203/298302
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0