|国家预印本平台
首页|Zeroth-order Gradient and Quasi-Newton Methods for Nonsmooth Nonconvex Stochastic Optimization

Zeroth-order Gradient and Quasi-Newton Methods for Nonsmooth Nonconvex Stochastic Optimization

Zeroth-order Gradient and Quasi-Newton Methods for Nonsmooth Nonconvex Stochastic Optimization

来源:Arxiv_logoArxiv
英文摘要

We consider the minimization of a Lipschitz continuous and expectation-valued function defined as $f(\mathbf{x}) \triangleq \mathbb{E}[{\tilde f}(\mathbf{x}, \boldsymbolξ)]$, over a closed and convex set. Our focus lies on obtaining both asymptotics as well as rate and complexity guarantees for computing an approximate stationary point (in a Clarke sense) via zeroth-order schemes. We adopt a smoothing-based approach reliant on minimizing $f_η$ where $f_η(\mathbf{x}) = \mathbb{E}_{\mathbf{u}}[f(\mathbf{x}+η\mathbf{u})]$, $\mathbf{u}$ is a random variable defined on a unit sphere, and $η> 0$. It has been observed that a stationary point of the $η$-smoothed problem is a $2η$-stationary point for the original problem in the Clarke sense. In such a setting, we develop two sets of schemes with promising empirical behavior. (I) We develop a smoothing-enabled variance-reduced zeroth-order gradient framework (VRG-ZO) and make two sets of contributions for the sequence generated by the proposed zeroth-order gradient scheme. (a) The residual function of the smoothed problem tends to zero almost surely along the generated sequence, allowing for making guarantees for $η$-Clarke stationary solutions of the original problem; (b) To compute an $\mathbf{x}$ that ensures that the expected norm of the residual of the $η$-smoothed problem is within $ε$ requires no greater than $O(η^{-1} ε^{-2})$ projection steps and $ O\left(η^{-2} ε^{-4}\right)$ function evaluations. (II) Our second scheme is a zeroth-order stochastic quasi-Newton scheme (VRSQN-ZO) reliant on a combination of randomized and Moreau smoothing; the corresponding iteration and sample complexities for this scheme are $ O\left(η^{-5}ε^{-2}\right)$ and $ O\left(η^{-7}ε^{-4}\right)$, respectively

Luke Marrinan、Uday V. Shanbhag、Farzad Yousefian

数学

Luke Marrinan,Uday V. Shanbhag,Farzad Yousefian.Zeroth-order Gradient and Quasi-Newton Methods for Nonsmooth Nonconvex Stochastic Optimization[EB/OL].(2025-06-25)[2025-07-25].https://arxiv.org/abs/2401.08665.点此复制

评论