Blog

I like playing with interesting ideas and do some analysis on toy problems (like quadratics).

  1. Online Learning to Precondition I. Space dilation methods for linear systems.
    Wenzhi Gao. 03/16/2026.
    HTML  |  PDF
    This is the beginning of a series of posts on Online Learning to Precondition (OL2P): learning to improve the optimization landscape through the behavior of algorithms. As the starting post, we derive a new algorithm for solving linear systems that achieves superlinear convergence, using a space dilation principle motivated by online learning.
  2. Hypergradient acceleration.
    Wenzhi Gao (with Yifa Yu). 03/06/2026.
    HTML  |  PDF  |  Code
    This post studies hypergradient descent (HDM), a stepsize adaptation heuristic for gradient-based methods. We show that when the objective function becomes flat around the optimum, HDM can automatically achieve accelerated convergence rate.
  3. Negative stepsizes also make multi-block ADMM converge.
    Wenzhi Gao. 03/02/2026.
    HTML  |  PDF  |  Code (divergence)  |  Code (negative stepsize)
    Inspired by a recent result that negative stepsizes make gradient descent ascent converge on bilinear counterexamples, this post shows that periodically using a negative dual stepsize also fixes the divergence of multi-block ADMM on classical quadratic counterexamples.