A Class of Short-term Recurrence Anderson Mixing Methods and Their Applications

Fuchao Wei Department of Computer Science and Technology, Tsinghua University Chenglong Bao Yau Mathematical Sciences Center, Tsinghua University; Yanqi Lake Beijing Institute of Mathematical Sciences and Applications Yang Liu Department of Computer Science and Technology, Tsinghua University; Institute for AI Industry Research, Tsinghua University

TBD mathscidoc:2206.43011

ICLR, 2022.4
Anderson mixing (AM) is a powerful acceleration method for fixed-point iterations, but its computation requires storing many historical iterations. The extra memory footprint can be prohibitive when solving high-dimensional problems in a resource-limited machine. To reduce the memory overhead, we propose a novel class of short-term recurrence AM methods (ST-AM). The ST-AM methods only store two previous iterations with cheap corrections. We prove that the basic version of ST-AM is equivalent to the full-memory AM in strongly convex quadratic optimization, and with minor changes it has local linear convergence for solving general nonlinear fixed-point problems. We further analyze the convergence properties of the regularized ST-AM for nonconvex (stochastic) optimization. Finally, we apply ST-AM to several applications including solving root-finding problems and training neural networks. Experimental results show that ST-AM is competitive with the long-memory AM and outperforms many existing optimizers.
No keywords uploaded!
[ Download ] [ 2022-06-15 21:34:18 uploaded by Baocl ] [ 366 downloads ] [ 0 comments ]
@inproceedings{fuchao2022a,
  title={A Class of Short-term Recurrence Anderson Mixing Methods and Their Applications},
  author={Fuchao Wei, Chenglong Bao, and Yang Liu},
  url={http://archive.ymsc.tsinghua.edu.cn/pacm_paperurl/20220615213418514405367},
  booktitle={ICLR},
  year={2022},
}
Fuchao Wei, Chenglong Bao, and Yang Liu. A Class of Short-term Recurrence Anderson Mixing Methods and Their Applications. 2022. In ICLR. http://archive.ymsc.tsinghua.edu.cn/pacm_paperurl/20220615213418514405367.
Please log in for comment!
 
 
Contact us: office-iccm@tsinghua.edu.cn | Copyright Reserved