Non-convergence of stochastic gradient descent in the training of deep neural networks (Q2034567): Difference between revisions

From MaRDI portal
Added link to MaRDI item.
Normalize DOI.
 
(4 intermediate revisions by 4 users not shown)
Property / DOI
 
Property / DOI: 10.1016/j.jco.2020.101540 / rank
Normal rank
 
Property / MaRDI profile type
 
Property / MaRDI profile type: MaRDI publication profile / rank
 
Normal rank
Property / OpenAlex ID
 
Property / OpenAlex ID: W3107512664 / rank
 
Normal rank
Property / arXiv ID
 
Property / arXiv ID: 2006.07075 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Optimization Methods for Large-Scale Machine Learning / rank
 
Normal rank
Property / cites work
 
Property / cites work: Theoretical Insights Into the Optimization Landscape of Over-Parameterized Shallow Neural Networks / rank
 
Normal rank
Property / cites work
 
Property / cites work: A comparative analysis of optimization and generalization properties of two-layer neural network and random feature models under gradient descent dynamics / rank
 
Normal rank
Property / cites work
 
Property / cites work: Gradient descent optimizes over-parameterized deep ReLU networks / rank
 
Normal rank
Property / DOI
 
Property / DOI: 10.1016/J.JCO.2020.101540 / rank
 
Normal rank

Latest revision as of 20:20, 16 December 2024

scientific article
Language Label Description Also known as
English
Non-convergence of stochastic gradient descent in the training of deep neural networks
scientific article

    Statements

    Non-convergence of stochastic gradient descent in the training of deep neural networks (English)
    0 references
    0 references
    0 references
    0 references
    22 June 2021
    0 references
    machine learning
    0 references
    deep neural networks
    0 references
    stochastic gradient descend
    0 references
    empirical risk minimization
    0 references
    non-convergence
    0 references

    Identifiers