A homotopy training algorithm for fully connected neural networks

From MaRDI portal
Publication:5160822

DOI10.1098/RSPA.2019.0662zbMATH Open1472.68131arXiv1903.09872OpenAlexW3103609633WikidataQ91866738 ScholiaQ91866738MaRDI QIDQ5160822FDOQ5160822


Authors: Qipin Chen, Wenrui Hao Edit this on Wikidata


Publication date: 29 October 2021

Published in: Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences (Search for Journal in Brave)

Abstract: In this paper, we present a Homotopy Training Algorithm (HTA) to solve optimization problems arising from fully connected neural networks with complicated structures. The HTA dynamically builds the neural network starting from a simplified version to the fully connected network via adding layers and nodes adaptively. Therefore, the corresponding optimization problem is easy to solve at the beginning and connects to the original model via a continuous path guided by the HTA, which provides a high probability to get a global minimum. By gradually increasing the complexity of model along the continuous path, the HTA gets a rather good solution to the original loss function. This is confirmed by various numerical results including VGG models on CIFAR-10. For example, on the VGG13 model with batch normalization, HTA reduces the error rate by 11.86% on test dataset comparing with the traditional method. Moreover, the HTA also allows us to find the optimal structure for a fully connected neural network by building the neutral network adaptively.


Full work available at URL: https://arxiv.org/abs/1903.09872




Recommendations




Cites Work


Cited In (9)





This page was built for publication: A homotopy training algorithm for fully connected neural networks

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5160822)