Accelerating First-Order Algorithms for High-Dimensional Minimax Optimization

Yi Zhang

Article ID: 10139
Vol 7, Issue 10, 2024

VIEWS - 2 (Abstract) 10 (PDF)

Abstract


This study introduces two first-order algorithms for high-dimensional minimax optimization: Accelerated Momentum Descent Ascent (AMDA) and Accelerated Variance-Reduced Gradient Descent Ascent (AVRGDA). These methods aim to address common challenges
in nonconvex optimization, such as slow convergence and computational complexity. AMDA leverages momentum-driven techniques to
smooth the optimization path, reducing oscillations and improving convergence speed, particularly in nonconvex-strongly-concave problems.
AVRGDA incorporates adaptive learning rates that dynamically adjust according to gradient norms, enhancing the efficiency of variance
reduction and handling complex optimization tasks in high-dimensional spaces. Through experiments in adversarial training and large-scale
logistic regression, these methods demonstrate superior performance in terms of training time, robustness, and computational cost compared
to traditional first-order methods. Theoretical analysis shows that AMDA and AVRGDA achieve convergence rates of O(ϵ−3)and O(ϵ−2.5)
respectively in high-dimensional, nonconvex minimax problems, confirming their efficiency and robustness in practical applications.

Keywords


First-Order Methods; High-Dimensional Optimization; Minimax Optimization

Full Text:

PDF


References


1. [1]Chris Junchi Li. “Accelerated Fully First-Order Methods for Bilevel and Minimax Optimization.” ArXiv(2024).

2. [2]Huang, F., Gao, S., Pei, J., & Huang, H. (2020). Accelerated Zeroth-Order and First-Order Momentum Methods from Mini to Minimax Optimization. J. Mach. Learn. Res., 23, 36:1-36:70.

3. [3]Michael Muehlebach, Michael I. Jordan. “Accelerated First-Order Optimization under Nonlinear Constraints.” ArXiv(2023).

4. [4]Kaiwen Zhou, A. M. So et al. “Boosting First-order Methods by Shifting Objective: New Schemes with Faster Worst Case Rates.”

5. ArXiv(2020).

6. [5]Ahmet Alacaoglu, Donghwan Kim et al. “Extending the Reach of First-Order Algorithms for Nonconvex Min-Max Problems with

7. Cohypomonotonicity.” ArXiv(2024).




DOI: https://doi.org/10.18686/ijmss.v7i10.10139

Refbacks

  • There are currently no refbacks.




Creative Commons License

This site is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.