Dynamical Systems–Based Neural Networks

From MaRDI portal
Publication:6181900

DOI10.1137/22M1527337arXiv2210.02373MaRDI QIDQ6181900FDOQ6181900

Elena Celledoni, Carola-Bibiane Schönlieb, Davide Murari, Ferdia Sherry, Brynjulf Owren

Publication date: 20 December 2023

Published in: SIAM Journal on Scientific Computing (Search for Journal in Brave)

Abstract: Neural networks have gained much interest because of their effectiveness in many applications. However, their mathematical properties are generally not well understood. If there is some underlying geometric structure inherent to the data or to the function to approximate, it is often desirable to take this into account in the design of the neural network. In this work, we start with a non-autonomous ODE and build neural networks using a suitable, structure-preserving, numerical time-discretisation. The structure of the neural network is then inferred from the properties of the ODE vector field. Besides injecting more structure into the network architectures, this modelling procedure allows a better theoretical understanding of their behaviour. We present two universal approximation results and demonstrate how to impose some particular properties on the neural networks. A particular focus is on 1-Lipschitz architectures including layers that are not 1-Lipschitz. These networks are expressive and robust against adversarial attacks, as shown for the CIFAR-10 dataset.


Full work available at URL: https://arxiv.org/abs/2210.02373







Cites Work


Cited In (9)





This page was built for publication: Dynamical Systems–Based Neural Networks

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6181900)