Accelerating First-Order Algorithms for High-Dimensional Minimax Optimization
Vol 7, Issue 10, 2024
VIEWS - 2 (Abstract) 10 (PDF)
Abstract
in nonconvex optimization, such as slow convergence and computational complexity. AMDA leverages momentum-driven techniques to
smooth the optimization path, reducing oscillations and improving convergence speed, particularly in nonconvex-strongly-concave problems.
AVRGDA incorporates adaptive learning rates that dynamically adjust according to gradient norms, enhancing the efficiency of variance
reduction and handling complex optimization tasks in high-dimensional spaces. Through experiments in adversarial training and large-scale
logistic regression, these methods demonstrate superior performance in terms of training time, robustness, and computational cost compared
to traditional first-order methods. Theoretical analysis shows that AMDA and AVRGDA achieve convergence rates of O(ϵ−3)and O(ϵ−2.5)
respectively in high-dimensional, nonconvex minimax problems, confirming their efficiency and robustness in practical applications.
Keywords
Full Text:
PDFReferences
1. [1]Chris Junchi Li. “Accelerated Fully First-Order Methods for Bilevel and Minimax Optimization.” ArXiv(2024).
2. [2]Huang, F., Gao, S., Pei, J., & Huang, H. (2020). Accelerated Zeroth-Order and First-Order Momentum Methods from Mini to Minimax Optimization. J. Mach. Learn. Res., 23, 36:1-36:70.
3. [3]Michael Muehlebach, Michael I. Jordan. “Accelerated First-Order Optimization under Nonlinear Constraints.” ArXiv(2023).
4. [4]Kaiwen Zhou, A. M. So et al. “Boosting First-order Methods by Shifting Objective: New Schemes with Faster Worst Case Rates.”
5. ArXiv(2020).
6. [5]Ahmet Alacaoglu, Donghwan Kim et al. “Extending the Reach of First-Order Algorithms for Nonconvex Min-Max Problems with
7. Cohypomonotonicity.” ArXiv(2024).
DOI: https://doi.org/10.18686/ijmss.v7i10.10139
Refbacks
- There are currently no refbacks.
This site is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.