Learnability with respect to fixed distributions (Q809614): Difference between revisions

From MaRDI portal
Import240304020342 (talk | contribs)
Set profile property.
ReferenceBot (talk | contribs)
Changed an Item
 
Property / cites work
 
Property / cites work: Fast probabilistic algorithms for Hamiltonian circuits and matchings / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q3795256 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Learnability and the Vapnik-Chervonenkis dimension / rank
 
Normal rank
Property / cites work
 
Property / cites work: Occam's razor / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q5798359 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Probably Approximate Learning over Classes of Distributions / rank
 
Normal rank
Property / cites work
 
Property / cites work: Computational limitations on learning from examples / rank
 
Normal rank
Property / cites work
 
Property / cites work: On the Uniform Convergence of Relative Frequencies of Events to Their Probabilities / rank
 
Normal rank
Property / cites work
 
Property / cites work: A theory of the learnable / rank
 
Normal rank
Property / cites work
 
Property / cites work: Deductive learning / rank
 
Normal rank

Latest revision as of 10:06, 24 June 2024

scientific article
Language Label Description Also known as
English
Learnability with respect to fixed distributions
scientific article

    Statements

    Learnability with respect to fixed distributions (English)
    0 references
    0 references
    0 references
    1991
    0 references
    Let D be a distribution over a set X (X is finite, countable or \(\subseteq E^ r)\), \(C\subseteq 2^ X-a\) concept class. For \(x\in X^ 1\), \(1\geq 1\) and a concept \(c\in C\) the 1-sample of c is \(sam_ c(x)=(<x_ 1,I(x_ 1)>,...,<x_ 1,I(x_ 1)>)\), where \(I_ c(x_ j)=1\) if \(x_ j\in C\) and \(I_ c(x_ j)=0\) otherwise. The sample space \(S_ c\) is the set of all l-samples of c over all \(c\in C\) and \(1\geq 1\). For an algebra H of Borel sets over X let \(F_{CH}\) be the set of all functions \(f:S_ c\to H\). The concept class C is learnable with respect (w.r.) to distribution D (over X) in terms of H if there exists a function \(f\in F_{CH}\) such that for all \(\epsilon,\delta >0\) there exists \(1>0\) such that for every \(c\in C\) and every \(x\in X^ 1\) selected according to D, \(f(sam_ c(x))\) and c are \(\epsilon\)-close w.r. to D with probability at least 1-\(\delta\) \((Y_ 1,Y_ 2\) are \(\epsilon\)-close w.r. to D if \(\Pr_ D(Y_ 1\oplus Y_ 2)<\epsilon\), where \(\oplus\) denotes the symmetric difference). A set \(H_{\epsilon}\subseteq 2^ X\) is an \(\epsilon\)-cover of C w.r. to D if for every \(c\in C\) there is an \(h\in H_{\epsilon}\) which is \(\epsilon\)- close to c. C is finitely coverable w.r. to D if for every \(\epsilon >0\) there is a finite \(\epsilon\)-cover of C. It is proved that C is learnable w.r. to D iff C is finitely coverable w.r. to D. A connection is established between the size of the cover and the Vapnik-Chervonenkis dimension of the concept class. Learning with errors (i.e. if labels in samples are independently correct with probability \(\zeta >0.5)\) is considered also.
    0 references
    0 references
    machine-learning
    0 references
    0 references