Deep learning: an introduction for applied mathematicians
From MaRDI portal
Publication:5243183
Abstract: Multilayered artificial neural networks are becoming a pervasive tool in a host of application fields. At the heart of this deep learning revolution are familiar concepts from applied and computational mathematics; notably, in calculus, approximation theory, optimization and linear algebra. This article provides a very brief introduction to the basic ideas that underlie deep learning from an applied mathematics perspective. Our target audience includes postgraduate and final year undergraduate students in mathematics who are keen to learn about the area. The article may also be useful for instructors in mathematics who wish to enliven their classes with references to the application of deep learning techniques. We focus on three fundamental questions: what is a deep neural network? how is a network trained? what is the stochastic gradient method? We illustrate the ideas with a short MATLAB code that sets up and trains a network. We also show the use of state-of-the art software on a large scale image classification problem. We finish with references to the current literature.
Recommendations
Cites work
- scientific article; zbMATH DE number 1243473 (Why is no real title available?)
- scientific article; zbMATH DE number 5060482 (Why is no real title available?)
- Deep learning
- Evaluating Derivatives
- Matlab guide
- Optimization methods for large-scale machine learning
- Stochastic gradient descent in continuous time
- Stochastic separation theorems
- Trust Region Algorithms and Timestep Selection
Cited in
(59)- Can neural networks learn finite elements?
- A hybrid Sobolev gradient method for learning NODEs
- MODNO: multi-operator learning with distributed neural operators
- Forecasting natural gas prices with spatio-temporal copula-based time series models
- Deep learning methods for limited data problems in X-ray tomography
- Recent developments in machine learning methods for stochastic control and games
- DeepBND: a machine learning approach to enhance multiscale solid mechanics
- Classification with Runge-Kutta networks and feature space augmentation
- Uniformly convex neural networks and non-stationary iterated network Tikhonov (iNETT) method
- Physics-informed deep learning for simultaneous surrogate modeling and PDE-constrained optimization of an airfoil geometry
- PFNN: a penalty-free neural network method for solving a class of second-order boundary-value problems on complex geometries
- Deep unfitted Nitsche method for elliptic interface problems
- Machine learning architectures for price formation models
- Book Reviews
- Supervised time series classification for anomaly detection in subsea engineering
- A new decision making method for selection of optimal data using the von Neumann-Morgenstern theorem
- An accelerated inexact Newton regularization scheme with a learned feature-selection rule for non-linear inverse problems
- A data-driven shock capturing approach for discontinuous Galerkin methods
- Adaptive sampling points based multi-scale residual network for solving partial differential equations
- Deep learning and geometric deep learning: An introduction for mathematicians and physicists
- Uncertainty quantification in scientific machine learning: methods, metrics, and comparisons
- Machine learning and reduced order computation of a friction stir welding model
- Physics-informed neural networks for the shallow-water equations on the sphere
- \(S\)-frame discrepancy correction models for data-informed Reynolds stress closure
- On a multilevel Levenberg-Marquardt method for the training of artificial neural networks and its application to the solution of partial differential equations
- Adaptive non-intrusive reduced order modeling for compressible flows
- Solving inverse problems using data-driven models
- Neural ODEs as the deep limit of ResNets with constant weights
- Mathematical methods in deep learning
- Sublinear convergence of a tamed stochastic gradient descent method in Hilbert space
- Generalization Error Analysis of Neural Networks with Gradient Based Regularization
- Asymptotic-preserving neural networks for hyperbolic systems with diffusive scaling
- Bilevel optimization, deep learning and fractional Laplacian regularization with applications in tomography
- Machine learning and computational mathematics
- An introduction to deep generative modeling
- Neural control of discrete weak formulations: Galerkin, least squares \& minimal-residual methods with quasi-optimal weights
- Deep CNNs as universal predictors of elasticity tensors in homogenization
- A machine-learning minimal-residual (ML-MRes) framework for goal-oriented finite element discretizations
- Deep learning: a Bayesian perspective
- A deep learning approach to the inversion of borehole resistivity measurements
- LSPIA, (stochastic) gradient descent, and parameter correction
- Machine learning moment closure models for the radiative transfer equation. I: Directly learning a gradient based closure
- Mathematics of deep learning. An introduction
- Solving Allen-Cahn and Cahn-Hilliard Equations using the Adaptive Physics Informed Neural Networks
- Using deep neural networks for detecting spurious oscillations in discontinuous Galerkin solutions of convection-dominated convection-diffusion equations
- A literature survey of matrix methods for data science
- Discovering phase field models from image data with the pseudo-spectral physics informed neural networks
- The train of artificial intelligence
- Time discretization in the solution of parabolic PDEs with ANNs
- Computational methods for deep learning. Theoretic, practice and applications
- Deep learning as optimal control problems: models and numerical methods
- scientific article; zbMATH DE number 7451142 (Why is no real title available?)
- Matching component analysis for transfer learning
- Long term dynamics of the subgradient method for Lipschitz path differentiable functions
- Mathematical Aspects of Deep Learning
- Control on the manifolds of mappings with a view to the deep learning
- Knowledge-informed neuro-integrators for aggregation kinetics
- Deep limits of residual neural networks
- Computational Methods for Deep Learning
Describes a project that uses
Uses Software
This page was built for publication: Deep learning: an introduction for applied mathematicians
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5243183)