Uniformly Faster Gradient Descent of Varying Step Sizes for Smooth Convex Functions
Program:
Applied Mathematics and Statistics
Project Description:
In using gradient descent method to optimize smooth convex functions, the conventional approach chooses a constant step size smaller than two for every iteration. Recent works have shown using stepsizes larger than two enables better final guarantees but at the cost of intermediate iterates performing poorly. We seek to find such longer stepsize patterns improving performance uniformly, not just at the last iteration. Using the recent computer-assisted analysis software, Performance Estimation Problems (PEP) in Python, we can construct the worst smooth convex function for given stepsizes. Numerical search over all small stepsize sequences (e.g. 3-steps) via a grid-search method and optimizing stepsizes via gradient descent produced new uniformly faster stepsize sequences. Continued work will formalize a proof for our small pattern’s performance and produce uniformly good stepsize patterns beyond small settings we have numerically explored.
Team Members
-
[foreach 357]
[if 397 not_equal=””][/if 397][395]
[/foreach 357]
Project Mentors, Sponsors, and Partners
Benjamin Grimmer
Course Faculty
-
[foreach 429]
[if 433 not_equal=””][/if 433][431]
[/foreach 429]
Project Links
Additional Project Information
Project Photo
Project Photo Caption:
Gradient Descent