A network of spiking neurons for computing sparse representations in an energy-efficient way

From MaRDI portal
Publication:2840871

DOI10.1162/NECO_A_00353zbMATH Open1268.92024arXiv1210.1530OpenAlexW2005918037WikidataQ42039320 ScholiaQ42039320MaRDI QIDQ2840871FDOQ2840871


Authors: Tao Hu, Dmitri B. Chklovskii, Alexander Genkin Edit this on Wikidata


Publication date: 23 July 2013

Published in: Neural Computation (Search for Journal in Brave)

Abstract: Computing sparse redundant representations is an important problem both in applied mathematics and neuroscience. In many applications, this problem must be solved in an energy efficient way. Here, we propose a hybrid distributed algorithm (HDA), which solves this problem on a network of simple nodes communicating via low-bandwidth channels. HDA nodes perform both gradient-descent-like steps on analog internal variables and coordinate-descent-like steps via quantized external variables communicated to each other. Interestingly, such operation is equivalent to a network of integrate-and-fire neurons, suggesting that HDA may serve as a model of neural computation. We show that the numerical performance of HDA is on par with existing algorithms. In the asymptotic regime the representation error of HDA decays with time, t, as 1/t. HDA is stable against time-varying noise, specifically, the representation error decays as 1/sqrt(t) for Gaussian white noise.


Full work available at URL: https://arxiv.org/abs/1210.1530




Recommendations



Cites Work


Cited In (4)





This page was built for publication: A network of spiking neurons for computing sparse representations in an energy-efficient way

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2840871)