Part of Advances in Neural Information Processing Systems 37 (NeurIPS 2024) Main Conference Track
Sebastian Allmeier, Nicolas Gast
We study stochastic approximation algorithms with Markovian noise and constant step-size α. We develop a method based on infinitesimal generator comparisons to study the bias of the algorithm, which is the expected difference between θn ---the value at iteration n--- and θ∗ ---the unique equilibrium of the corresponding ODE. We show that, under some smoothness conditions, this bias is of order O(α). Furthermore, we show that the time-averaged bias is equal to αV+O(α2), where V is a constant characterized by a Lyapunov equation, showing that E[ˉθn]≈θ∗+Vα+O(α2), where ˉθn is the Polyak-Ruppert average. We also show that ˉθn converges with high probability around θ∗+αV. We illustrate how to combine this with Richardson-Romberg extrapolation to derive an iterative scheme with a bias of order O(α2).