4.5 Article

Algorithms for Nonnegative Matrix Factorization with the β-Divergence

Journal

NEURAL COMPUTATION
Volume 23, Issue 9, Pages 2421-2456

Publisher

MIT PRESS
DOI: 10.1162/NECO_a_00168

Keywords

-

Funding

  1. [ANR-09-JCJC-0073-01 TANGERINE]

Ask authors/readers for more resources

This letter describes algorithms for nonnegative matrix factorization (NMF) with the beta-divergence (beta-NMF). The beta-divergence is a family of cost functions parameterized by a single shape parameter beta that takes the Euclidean distance, the Kullback-Leibler divergence, and the Itakura-Saito divergence as special cases (beta = 2, 1, 0 respectively). The proposed algorithms are based on a surrogate auxiliary function (a local majorization of the criterion function). We first describe a majorization-minimization algorithm that leads to multiplicative updates, which differ from standard heuristic multiplicative updates by a beta-dependent power exponent. The monotonicity of the heuristic algorithm can, however, be proven for beta is an element of (0, 1) using the proposed auxiliary function. Then we introduce the concept of the majorization-equalization (ME) algorithm, which produces updates that move along constant level sets of the auxiliary function and lead to larger steps than MM. Simulations on synthetic and real data illustrate the faster convergence of the ME approach. The letter also describes how the proposed algorithms can be adapted to two common variants of NMF: penalized NMF (when a penalty function of the factors is added to the criterion function) and convex NMF (when the dictionary is assumed to belong to a known subspace).

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available