Non-negative matrix factorization with fixed row and column sums (Q935373): Difference between revisions

From MaRDI portal
RedirectionBot (talk | contribs)
Removed claim: author (P16): Item:Q208837
ReferenceBot (talk | contribs)
Changed an Item
(3 intermediate revisions by 3 users not shown)
Property / author
 
Property / author: Paul Van Dooren / rank
 
Normal rank
Property / MaRDI profile type
 
Property / MaRDI profile type: MaRDI publication profile / rank
 
Normal rank
Property / full work available at URL
 
Property / full work available at URL: https://doi.org/10.1016/j.laa.2007.02.026 / rank
 
Normal rank
Property / OpenAlex ID
 
Property / OpenAlex ID: W2085937140 / rank
 
Normal rank
Property / cites work
 
Property / cites work: On reduced rank nonnegative matrix factorization for symmetric nonnegative matrices / rank
 
Normal rank
Property / cites work
 
Property / cites work: Nonnegative matrix factorization and I-divergence alternating minimization / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q5689624 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Learning the parts of objects by non-negative matrix factorization / rank
 
Normal rank

Revision as of 13:11, 28 June 2024

scientific article
Language Label Description Also known as
English
Non-negative matrix factorization with fixed row and column sums
scientific article

    Statements

    Non-negative matrix factorization with fixed row and column sums (English)
    0 references
    0 references
    0 references
    6 August 2008
    0 references
    Given a non-negative \(m\times n\) matrix \(A\), find two non-negative matrices \(U\) (\(m\times k\)) and \(V\) (\(n \times k\)) with \(k\ll m, n\) that minimize \(F (A, U V^T)\), where \[ F (A, U V^T)=D(A|| UV^T) :=\sum_{ij}A_{ij} \log\frac{A_{ij}}{[UV^T]_{ij}} -A_{ij}+[UV^T]_{ij}. \] The problem \(\min_{U\geq 0,V\geq 0} ED(A||U V^T)\), where \(A\geq 0\), is called non-negative matrix factorization using the generalized Kullback-Leibler (KL) divergence. The authors show: Let \(A_{m\times n}\) be a non-negative matrix. Then every stationary point \((U, V)\) of the cost function \(D(A||U V^T)\) preserves the column sums of \(A\), the row sums of \(A\), and the matrix sum of \(A\). Every stationary point \((U_{m\times k}, V_{n\times k})\) of the KL minimization problem has the form \(U V^T = P_{m\times k} D_{k\times k} Q_{n\times k}^T\), where \(P, Q\) are column stochastic, \(D\) is diagonal non-negative, and \(\sum_i D_{ii} = \sum_{ij} A_{ij}\). Furthermore, if \(A\) is column stochastic (or row stochastic), then the matrix \(DQ^T\) (or \(P D\)) is also column stochastic (or row stochastic). The authors give an application in stochastic matrix approximation.
    0 references
    non-negative matrix factorization
    0 references
    generalized Kullback-Leibler divergence
    0 references
    stochastic matrix approximation
    0 references
    row sums
    0 references
    column sums
    0 references

    Identifiers