Active online learning in the binary perceptron problem

From MaRDI portal
Publication:3387706

DOI10.1088/0253-6102/71/2/243zbMATH Open1452.68166arXiv1902.08043OpenAlexW3103323995WikidataQ128394376 ScholiaQ128394376MaRDI QIDQ3387706FDOQ3387706


Authors: Haijun Zhou Edit this on Wikidata


Publication date: 13 January 2021

Published in: Communications in Theoretical Physics (Search for Journal in Brave)

Abstract: The binary perceptron is the simplest artificial neural network formed by N input units and one output unit, with the neural states and the synaptic weights all restricted to pm1 values. The task in the teacher--student scenario is to infer the hidden weight vector by training on a set of labeled patterns. Previous efforts on the passive learning mode have shown that learning from independent random patterns is quite inefficient. Here we consider the active online learning mode in which the student designs every new Ising training pattern. We demonstrate that it is mathematically possible to achieve perfect (error-free) inference using only N designed training patterns, but this is computationally unfeasible for large systems. We then investigate two Bayesian statistical designing protocols, which require 2.3N and 1.9N training patterns, respectively, to achieve error-free inference. If the training patterns are instead designed through deductive reasoning, perfect inference is achieved using N!+!log2!N samples. The performance gap between Bayesian and deductive designing strategies may be shortened in future work by taking into account the possibility of ergodicity breaking in the version space of the binary perceptron.


Full work available at URL: https://arxiv.org/abs/1902.08043




Recommendations




Cites Work


Cited In (3)





This page was built for publication: Active online learning in the binary perceptron problem

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q3387706)