Anchored gradient descent achieves O(1/sqrt(T)) last-iterate convergence for monotone inclusions 0 in F(z) + A(z) by extending prior unconstrained proofs.
An Improved Last-Iterate Convergence Rate for Anchored Gradient Descent Ascent
1 Pith paper cite this work. Polarity classification is still indexing.
abstract
We analyze the last-iterate convergence of the Anchored Gradient Descent Ascent algorithm for smooth convex-concave min-max problems. While previous work established a last-iterate rate of $\mathcal{O}(1/t^{2-2p})$ for the squared gradient norm, where $p \in (1/2, 1)$, it remained an open problem whether the improved exact $\mathcal{O}(1/t)$ rate is achievable. In this work, we resolve this question in the affirmative. This result was discovered autonomously by an AI system capable of writing formal proofs in Lean. The Lean proof can be accessed at https://github.com/google-deepmind/formal-conjectures/pull/3675/commits/a13226b49fd3b897f4c409194f3bcbeb96a08515
fields
math.OC 1years
2026 1verdicts
UNVERDICTED 1representative citing papers
citing papers explorer
-
Last-Iterate Convergence of Anchored Gradient Descent
Anchored gradient descent achieves O(1/sqrt(T)) last-iterate convergence for monotone inclusions 0 in F(z) + A(z) by extending prior unconstrained proofs.