Convergence Rates of Training Deep Neural Networks via Alternating Minimization Methods

From MaRDI portal
Publication:6409258

DOI10.1007/S11590-023-02026-4arXiv2208.14318MaRDI QIDQ6409258FDOQ6409258


Authors: Jintao Xu, Chenglong Bao, Wenxun Xing Edit this on Wikidata


Publication date: 30 August 2022

Abstract: Training deep neural networks (DNNs) is an important and challenging optimization problem in machine learning due to its non-convexity and non-separable structure. The alternating minimization (AM) approaches split the composition structure of DNNs and have drawn great interest in the deep learning and optimization communities. In this paper, we propose a unified framework for analyzing the convergence rate of AM-type network training methods. Our analysis is based on the non-monotone j-step sufficient decrease conditions and the Kurdyka-Lojasiewicz (KL) property, which relaxes the requirement of designing descent algorithms. We show the detailed local convergence rate if the KL exponent heta varies in [0,1). Moreover, the local R-linear convergence is discussed under a stronger j-step sufficient decrease condition.













This page was built for publication: Convergence Rates of Training Deep Neural Networks via Alternating Minimization Methods

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6409258)