Overparameterized neural networks implement associative memory

From MaRDI portal
Publication:5073192

DOI10.1073/PNAS.2005013117zbMATH Open1485.68236arXiv1909.12362OpenAlexW3093151833WikidataQ100693711 ScholiaQ100693711MaRDI QIDQ5073192FDOQ5073192


Authors: Adityanarayanan Radhakrishnan, Mikhail Belkin, Caroline Uhler Edit this on Wikidata


Publication date: 5 May 2022

Published in: Proceedings of the National Academy of Sciences (Search for Journal in Brave)

Abstract: Identifying computational mechanisms for memorization and retrieval of data is a long-standing problem at the intersection of machine learning and neuroscience. Our main finding is that standard overparameterized deep neural networks trained using standard optimization methods implement such a mechanism for real-valued data. Empirically, we show that: (1) overparameterized autoencoders store training samples as attractors, and thus, iterating the learned map leads to sample recovery; (2) the same mechanism allows for encoding sequences of examples, and serves as an even more efficient mechanism for memory than autoencoding. Theoretically, we prove that when trained on a single example, autoencoders store the example as an attractor. Lastly, by treating a sequence encoder as a composition of maps, we prove that sequence encoding provides a more efficient mechanism for memory than autoencoding.


Full work available at URL: https://arxiv.org/abs/1909.12362




Recommendations



Cites Work


Cited In (4)





This page was built for publication: Overparameterized neural networks implement associative memory

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5073192)