Local convergence of the alternating least squares algorithm for canonical tensor approximation (Q2910973)

From MaRDI portal





scientific article; zbMATH DE number 6081317
Language Label Description Also known as
default for all languages
No label defined
    English
    Local convergence of the alternating least squares algorithm for canonical tensor approximation
    scientific article; zbMATH DE number 6081317

      Statements

      0 references
      12 September 2012
      0 references
      low-rank approximation
      0 references
      nonlinear Gauss-Seidel method, PARAFAC
      0 references
      local convergence
      0 references
      alternating least square algorithm
      0 references
      tensor
      0 references
      global minimization
      0 references
      Lagrange multipliers
      0 references
      Local convergence of the alternating least squares algorithm for canonical tensor approximation (English)
      0 references
      The author presents a local convergence result for the alternating least square (ALS) algorithm under the assumption that the Hessian of the loss function at the solution is essentially positive definite, except on a trivial null space caused by the scaling indeterminacy. This assumption necessarily requires the local essential uniqueness of the CP decomposition (that is, uniqueness up to scaling), which is reasonable for tensors of order higher than three. The proof shows that an inbuilt normalization procedure in the ALS algorithm acts (in first order) as a projection onto a subspace complementary to the null space of the Hessian. On this subspace the linear Gauss-Seidel method is known to be contractive.NEWLINENEWLINEThe main ideas are not specifically related to the least squares error as the loss function to be minimized. It turns out that the global minimization in one ALS direction can be replaced by a single Newton step to obtain an approximate scheme which is still convergent (under the same assumptions). One convenient aspect of this approach is that it avoids the explicit use of Lagrange multipliers. It is therefore easily accessible and, in principle, applicable to more delicate types of redundancies as they, for instance, occur in the Tucker format or in the newly developed TT format.
      0 references

      Identifiers

      0 references
      0 references
      0 references
      0 references
      0 references
      0 references
      0 references