Ensemble Kalman inversion: a derivative-free technique for machine learning tasks

From MaRDI portal
Publication:5197869

DOI10.1088/1361-6420/AB1C3AzbMATH Open1430.68266arXiv1808.03620OpenAlexW2949400887WikidataQ127968194 ScholiaQ127968194MaRDI QIDQ5197869FDOQ5197869


Authors: Nikola B. Kovachki, A. M. Stuart Edit this on Wikidata


Publication date: 20 September 2019

Published in: Inverse Problems (Search for Journal in Brave)

Abstract: The standard probabilistic perspective on machine learning gives rise to empirical risk-minimization tasks that are frequently solved by stochastic gradient descent (SGD) and variants thereof. We present a formulation of these tasks as classical inverse or filtering problems and, furthermore, we propose an efficient, gradient-free algorithm for finding a solution to these problems using ensemble Kalman inversion (EKI). Applications of our approach include offline and online supervised learning with deep neural networks, as well as graph-based semi-supervised learning. The essence of the EKI procedure is an ensemble based approximate gradient descent in which derivatives are replaced by differences from within the ensemble. We suggest several modifications to the basic method, derived from empirically successful heuristics developed in the context of SGD. Numerical results demonstrate wide applicability and robustness of the proposed algorithm.


Full work available at URL: https://arxiv.org/abs/1808.03620




Recommendations




Cites Work


Cited In (42)

Uses Software





This page was built for publication: Ensemble Kalman inversion: a derivative-free technique for machine learning tasks

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5197869)