Incremental gradient algorithms with stepsizes bounded away from zero (Q1273418)

From MaRDI portal





scientific article; zbMATH DE number 1230440
Language Label Description Also known as
default for all languages
No label defined
    English
    Incremental gradient algorithms with stepsizes bounded away from zero
    scientific article; zbMATH DE number 1230440

      Statements

      Incremental gradient algorithms with stepsizes bounded away from zero (English)
      0 references
      15 June 1999
      0 references
      For the problem \[ \text{minimize}\quad \sum^k_{j= 1} f_j(x)\quad\text{subject to}\quad x\in\mathbb{R}^n, \] where \(f_1,\dots, f_k: \mathbb{R}^n\to \mathbb{R}\) are continuously differentiable functions, the author proposes the following incremental gradient algorithm: Choose any \(x^0\in \mathbb{R}^n\). Having \(x^i\), check a stopping criterion. If not satisfied, compute \(x^{i+ 1}= T(x^i,\eta_i)\), where \(T:\mathbb{R}^n\times \mathbb{R}^+\to \mathbb{R}^n\) is given by \[ T(x,\eta):= x-\eta \sum^k_{j=1}\nabla f_j(z^j) \] and \[ z^1= x,\quad z^{j+1}= z^j- \eta\nabla f_j(z^j),\quad j=1,\dots, k-1. \] The author presents convergence results for a class of incremental gradient algorithms with stepsizes bounded away from zero. Applications on neural network training are discussed.
      0 references
      incremental gradient algorithm
      0 references
      convergence
      0 references
      neural network training
      0 references
      0 references
      0 references

      Identifiers

      0 references
      0 references
      0 references
      0 references
      0 references