Tighter Lower Bounds for Shuffling SGD: Random Permutations and Beyond

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 46
  • Download : 0
We study convergence lower bounds of without-replacement stochastic gradient descent (SGD) for solving smooth (strongly-)convex finite-sum minimization problems. Unlike most existing results focusing on final iterate lower bounds in terms of the number of components 𝑛 and the number of epochs 𝐾, we seek bounds for arbitrary weighted average iterates that are tight in all factors including the condition number πœ…. For SGD with Random Reshuffling, we present lower bounds that have tighter πœ… dependencies than existing bounds. Our results are the first to perfectly close the gap between lower and upper bounds for weighted average iterates in both strongly-convex and convex cases. We also prove weighted average iterate lower bounds for arbitrary permutation-based SGD, which apply to all variants that carefully choose the best permutation. Our bounds improve the existing bounds in factors of 𝑛 and πœ… and thereby match the upper bounds shown for a recently proposed algorithm called GraB.
Publisher
International Conference on Machine Learning
Issue Date
2023-07-25
Language
English
Citation

40th International Conference on Machine Learning, ICML 2023, pp.3855 - 3912

ISSN
2640-3498
URI
http://hdl.handle.net/10203/316023
Appears in Collection
AI-Conference Papers(ν•™μˆ λŒ€νšŒλ…Όλ¬Έ)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0