Linearized two-layers neural networks in high dimension (Q2039801): Difference between revisions

From MaRDI portal
Importer (talk | contribs)
Changed an Item
ReferenceBot (talk | contribs)
Changed an Item
 
Property / cites work
 
Property / cites work: Neural Network Learning / rank
 
Normal rank
Property / cites work
 
Property / cites work: On the Equivalence between Kernel Quadrature Rules and Random Feature Expansions / rank
 
Normal rank
Property / cites work
 
Property / cites work: Breaking the Curse of Dimensionality with Convex Neural Networks / rank
 
Normal rank
Property / cites work
 
Property / cites work: Universal approximation bounds for superpositions of a sigmoidal function / rank
 
Normal rank
Property / cites work
 
Property / cites work: Reconciling modern machine-learning practice and the classical bias–variance trade-off / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q5488485 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Optimal rates for the regularized least-squares algorithm / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q2755103 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Approximation by superpositions of a sigmoidal function / rank
 
Normal rank
Property / cites work
 
Property / cites work: Optimal nonlinear approximation / rank
 
Normal rank
Property / cites work
 
Property / cites work: Projection-based approximation and a duality with kernel methods / rank
 
Normal rank
Property / cites work
 
Property / cites work: Spherical Harmonics in p Dimensions / rank
 
Normal rank
Property / cites work
 
Property / cites work: The spectrum of kernel random matrices / rank
 
Normal rank
Property / cites work
 
Property / cites work: On information plus noise kernel random matrices / rank
 
Normal rank
Property / cites work
 
Property / cites work: Disentangling feature and lazy training in deep neural networks / rank
 
Normal rank
Property / cites work
 
Property / cites work: A distribution-free theory of nonparametric regression / rank
 
Normal rank
Property / cites work
 
Property / cites work: Wide neural networks of any depth evolve as linear models under gradient descent <sup>*</sup> / rank
 
Normal rank
Property / cites work
 
Property / cites work: Just interpolate: kernel ``ridgeless'' regression can generalize / rank
 
Normal rank
Property / cites work
 
Property / cites work: On best approximation by ridge functions / rank
 
Normal rank
Property / cites work
 
Property / cites work: On the near optimality of the stochastic approximation of smooth functions by neural networks / rank
 
Normal rank
Property / cites work
 
Property / cites work: The landscape of empirical risk for nonconvex losses / rank
 
Normal rank
Property / cites work
 
Property / cites work: A mean field view of the landscape of two-layer neural networks / rank
 
Normal rank
Property / cites work
 
Property / cites work: Dimension-independent bounds on the degree of approximation by neural networks / rank
 
Normal rank
Property / cites work
 
Property / cites work: Bayesian learning for neural networks / rank
 
Normal rank
Property / cites work
 
Property / cites work: Approximation by Ridge Functions and Neural Networks / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4938227 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Mean field analysis of neural networks: a central limit theorem / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4614113 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Introduction to nonparametric estimation / rank
 
Normal rank
Property / cites work
 
Property / cites work: Gradient descent optimizes over-parameterized deep ReLU networks / rank
 
Normal rank

Latest revision as of 04:04, 26 July 2024

scientific article
Language Label Description Also known as
English
Linearized two-layers neural networks in high dimension
scientific article

    Statements

    Linearized two-layers neural networks in high dimension (English)
    0 references
    0 references
    5 July 2021
    0 references
    The authors study nonparametric regression problems for univariate responses \(y_1, \ldots, y_n\) and \(\mathbb{R}^d\)-valued feature vectors \(\mathbf{x}_1, \hdots, \mathbf{x}_n\), where the tuples \((y_i, \mathbf{x}_i)_{1 \leq i \leq n}\) are assumed to be stochastically independent and identically distributed. Their goal is to construct a function \(f: \mathbb{R}^d \to \mathbb{R}\) which predicts future responses. The quality of such an \(f\) is assessed via its square prediction risk. In particular, the authors consider choosing \(f\) from the class \(\mathcal{F}_{\text{NN}}\) of two-layer neural networks. An approximation (based on a first-order Taylor expansion) of \(f \in \mathcal{F}_{\text{NN}}\) by a part belonging to a random features model and a part belonging to a neural tangent class is studied. The approximation errors of both parts are analyzed under different asymptotic regimes in which \(n\) and/or \(d\) tend to infinity. Furthermore, the generalization error of certain kernel methods is analyzed. Besides these theoretical contributions, the authors also present some numerical results.
    0 references
    approximation bounds
    0 references
    kernel ridge regression
    0 references
    neural tangent class
    0 references
    random features
    0 references
    0 references
    0 references
    0 references
    0 references

    Identifiers

    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references