|国家预印本平台
首页|In almost all shallow analytic neural network optimization landscapes, efficient minimizers have strongly convex neighborhoods

In almost all shallow analytic neural network optimization landscapes, efficient minimizers have strongly convex neighborhoods

In almost all shallow analytic neural network optimization landscapes, efficient minimizers have strongly convex neighborhoods

来源:Arxiv_logoArxiv
英文摘要

Whether or not a local minimum of a cost function has a strongly convex neighborhood greatly influences the asymptotic convergence rate of optimizers. In this article, we rigorously analyze the prevalence of this property for the mean squared error induced by shallow, 1-hidden layer neural networks with analytic activation functions when applied to regression problems. The parameter space is divided into two domains: the 'efficient domain' (all parameters for which the respective realization function cannot be generated by a network having a smaller number of neurons) and the 'redundant domain' (the remaining parameters). In almost all regression problems on the efficient domain the optimization landscape only features local minima that are strongly convex. Formally, we will show that for certain randomly picked regression problems the optimization landscape is almost surely a Morse function on the efficient domain. The redundant domain has significantly smaller dimension than the efficient domain and on this domain, potential local minima are never isolated.

Felix Benning、Steffen Dereich

计算技术、计算机技术

Felix Benning,Steffen Dereich.In almost all shallow analytic neural network optimization landscapes, efficient minimizers have strongly convex neighborhoods[EB/OL].(2025-04-11)[2025-04-27].https://arxiv.org/abs/2504.08867.点此复制

评论