Depth separations in neural networks: what is actually being separated? (Q2117335): Difference between revisions

From MaRDI portal
RedirectionBot (talk | contribs)
Removed claim: reviewed by (P1447): Item:Q448742
ReferenceBot (talk | contribs)
Changed an Item
 
(4 intermediate revisions by 4 users not shown)
Property / reviewed by
 
Property / reviewed by: Alexey L. Lukashov / rank
 
Normal rank
Property / MaRDI profile type
 
Property / MaRDI profile type: MaRDI publication profile / rank
 
Normal rank
Property / OpenAlex ID
 
Property / OpenAlex ID: W2966686587 / rank
 
Normal rank
Property / arXiv ID
 
Property / arXiv ID: 1904.06984 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Proof of the Achievability Conjectures for the General Stochastic Block Model / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4222737 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Universal approximation bounds for superpositions of a sigmoidal function / rank
 
Normal rank
Property / cites work
 
Property / cites work: Theory of Classification: a Survey of Some Recent Advances / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4828422 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Approximation and learning of convex superpositions / rank
 
Normal rank
Property / cites work
 
Property / cites work: Agnostically Learning Halfspaces / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q2969663 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Approximation by Combinations of ReLU and Squared ReLU Ridge Functions With <inline-formula> <tex-math notation="LaTeX">$\ell^1$ </tex-math> </inline-formula> and <inline-formula> <tex-math notation="LaTeX">$\ell^0$ </tex-math> </inline-formula> Controls / rank
 
Normal rank
Property / cites work
 
Property / cites work: A note on approximation of a ball by polytopes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Theory of probability and random processes. / rank
 
Normal rank
Property / cites work
 
Property / cites work: Inverses of Vandermonde Matrices / rank
 
Normal rank
Property / cites work
 
Property / cites work: Provable approximation properties for deep neural networks / rank
 
Normal rank
Property / cites work
 
Property / cites work: Understanding Machine Learning / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4558174 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Error bounds for approximations with deep ReLU networks / rank
 
Normal rank

Latest revision as of 09:11, 28 July 2024

scientific article
Language Label Description Also known as
English
Depth separations in neural networks: what is actually being separated?
scientific article

    Statements

    Depth separations in neural networks: what is actually being separated? (English)
    0 references
    0 references
    0 references
    0 references
    21 March 2022
    0 references
    The authors consider approximation properties of depth 2 networks \[ N_2(\mathbf{x})=\sum_{i=1}^wu_i\sigma(\mathbf{w}_i^{\mathsf{T}}\mathbf{x}+b_i). \] The main results are given in three subsections of Section 2. Subsection 2.1 contains a formal result implying that radial functions can be approximated with depth 2, width (parameter \( w \)) poly(\( d \)) (\( \mathbf{x},\mathbf{w}_i\in\mathbb{R}^d \)) networks, to any constant accuracy \( \epsilon. \) This result is proved for networks employing any activation function \( \sigma \) satisfying mild assumption, which implies that the activation can be used to approximate univariate functions well. This assumption is satisfied for all standard activations such as ReLU and sigmoidal functions. In Subsection 2.2 the authors show how Lipschitz radial functions can be approximated by width poly(\( 1/\epsilon) \) depth 2 ReLU networks. Results of Subsection 2.3 complement previous positive approximation results with negative results. Section 3 contains proofs.
    0 references
    deep learning
    0 references
    neural network
    0 references
    approximation theory
    0 references
    depth separation
    0 references

    Identifiers

    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references