Automatic Relevance Determination in Nonnegative Matrix Factorization with the /spl beta/-Divergence
Citations Over TimeTop 10% of 2012 papers
Abstract
This paper addresses the estimation of the latent dimensionality in nonnegative matrix factorization (NMF) with the β-divergence. The β-divergence is a family of cost functions that includes the squared euclidean distance, Kullback-Leibler (KL) and Itakura-Saito (IS) divergences as special cases. Learning the model order is important as it is necessary to strike the right balance between data fidelity and overfitting. We propose a Bayesian model based on automatic relevance determination (ARD) in which the columns of the dictionary matrix and the rows of the activation matrix are tied together through a common scale parameter in their prior. A family of majorization-minimization (MM) algorithms is proposed for maximum a posteriori (MAP) estimation. A subset of scale parameters is driven to a small lower bound in the course of inference, with the effect of pruning the corresponding spurious components. We demonstrate the efficacy and robustness of our algorithms by performing extensive experiments on synthetic data, the swimmer dataset, a music decomposition example, and a stock price prediction task.
Related Papers
- → Non-Negative Matrix Factorization with Constraints(2010)60 cited
- → CUR+NMF for learning spectral features from large data matrix(2008)10 cited
- → Sparsity promoted non-negative matrix factorization for source separation and detection(2014)3 cited
- → Detection of Brain Activity in Functional Magnetic Resonance Imaging Data using Matrix Factorization(2013)1 cited
- → PHASL-NMF: Hierarchical ALS Based Power Non-Negative Matrix Factorization(2023)