arXiv Analytics

Sign in

arXiv:2305.03938 [math.OC]AbstractReferencesReviewsResources

Adam-family Methods for Nonsmooth Optimization with Convergence Guarantees

Nachuan Xiao, Xiaoyin Hu, Xin Liu, Kim-Chuan Toh

Published 2023-05-06Version 1

In this paper, we present a comprehensive study on the convergence properties of Adam-family methods for nonsmooth optimization, especially in the training of nonsmooth neural networks. We introduce a novel two-timescale framework that adopts a two-timescale updating scheme, and prove its convergence properties under mild assumptions. Our proposed framework encompasses various popular Adam-family methods, providing convergence guarantees for these methods in training nonsmooth neural networks. Furthermore, we develop stochastic subgradient methods that incorporate gradient clipping techniques for training nonsmooth neural networks with heavy-tailed noise. Through our framework, we show that our proposed methods converge even when the evaluation noises are only assumed to be integrable. Extensive numerical experiments demonstrate the high efficiency and robustness of our proposed methods.

Related articles: Most relevant | Search more
arXiv:1610.03446 [math.OC] (Published 2016-10-11)
Nonsmooth optimization using Taylor-like models: error bounds, convergence, and termination criteria
arXiv:2505.07143 [math.OC] (Published 2025-05-11)
Subgradient Regularization: A Descent-Oriented Subgradient Method for Nonsmooth Optimization
arXiv:2407.02146 [math.OC] (Published 2024-07-02)
Coderivative-Based Newton Methods with Wolfe Linesearch for Nonsmooth Optimization