The theory and practice of stochastic optimization has focused on stochastic gradient descent (SGD) in recent years, retaining the basic first-order stochastic nature of SGD while aiming to improve it via mechanisms such as averaging, momentum, and variance reduction. Improvement can be measured along various dimensions, however, and it has proved difficult to achieve improvements both in terms of nonasymptotic measures of convergence rate and asymptotic measures of distributional tightness. In this work, we consider first-order stochastic optimization from a general statistical point of view, motivating a specific form of recursive averaging of past stochastic gradients. The resulting algorithm, which we refer to as Recursive One-Over-T SGD (ROOT-SGD), matches the state-of-the-art convergence rate among online variance-reduced stochastic approximation methods. Moreover, under slightly stronger distributional assumptions, the rescaled last-iterate of ROOT-SGD converges to a zero-mean Gaussian distribution that achieves near-optimal covariance. This is a joint work with Wenlong Mou, Martin Wainwright, and Michael Jordan.
August 2020
11am - 12pm
Department of Mathematics
Payment Details
Alumni, Faculty and Staff, PG Students, UG Students
Other Events
Seminar, Lecture, Talk
Department of Chemistry Seminar - Deep Learning in Protein Folding: Trajectory Reconstruction from Experimental Data and Ultra-fast Latent Space Simulators
Speaker: Professor Andrew FERGUSON Institution: Pritzker School of Molecular Engineering, University of Chicago, USA Hosted by: Professor Xuhui HUANG Zoom Link:
Jun 2021
Seminar, Lecture, Talk
Seminar of Physics Department - Deep Learning Enhanced Quantum Information Experiments in Semiconductors
May 2021