A fast Multiplicative Updates algorithm for Non-negative Matrix Factorization
Abstract
Nonnegative Matrix Factorization is an important tool in unsupervised machine learning to decompose a data matrix into a sum of parts that are often interpretable. Many dedicated algorithms have been proposed during the last three decades. A well-known method is the Multiplicative Updates algorithm proposed by Lee and Seung in 2002. Multiplicative updates have many interesting features: they are simple to implement, can be adapted to popular variants such as sparse Nonnegative Matrix Factorization, and, according to recent benchmarks, are state-of-the-art for many problems where the loss function is not the squared Frobenius norm. In this manuscript, we propose to improve the Multiplicative Updates algorithm seen as an alternating majorization minimization algorithm by crafting a tighter upper bound of the Hessian matrix for each alternate subproblem. Convergence is still ensured and we observe in practice on both synthetic and real world dataset that the proposed fastMU algorithm is often significantly faster than the regular Multiplicative Updates algorithm, and can even be competitive with state-of-the-art methods for the Frobenius loss.
Keywords— Nonnegative Matrix Factorization, Quadratic majorization, Multiplicative Updates, Alternating Optimization
1 Introduction
Nonnegative Matrix Factorization (NMF) is the (approximate) decomposition of a matrix into a product of nonnegative matrix factors. In many applications, the factors of the decomposition give an interesting interpretation of its content as a sum of parts. While Paatero and Tapper could be attributed with the modern formulation of NMF [23] as known in the data sciences, its roots are much older and stem from many fields, primarily chemometrics, and the Beer Lambert law, see for instance the literature survey in [11]. NMF can be formulated as follows: given a matrix , find two non-negative (entry-wise) matrices and with fixed , typically, one selects , that satisfy
(1) |
Formally, we solve a low-rank approximation problem by minimizing a loss function :
(2) |
Classical instances of are the squared Frobenius norm or the Kullback-Leibler divergence (KL) defined in Equations (5) and (6). The integer is often called the nonnegative rank of the approximation matrix .
There is a large body of literature on how to compute solutions to the NMF problem. Existing algorithms could be classified based on two criteria: the loss they minimize, and whether they alternatively solve for and , that is, for each iteration
(3) | |||
(4) |
or update and simultaneoustly. In many usual loss functions, the subproblems of the alternating algorithms are (strongly) convex. Therefore convergence guarantees can be ensured using the theory of block-coordinate descent. Moreover, these methods generally perform well in practice. For the Frobenius loss, state-of-the-art alternating algorithms include the Hierarchical Alternating Least Squares [7, 12]. For the more general beta-divergence loss which includes the popular KL divergence, the Multiplicative Updates algorithm proposed by Lee and Seung [18, 19], that we will denote by MU in the rest of this manuscript, is still state-of-the-art for many dataset [14].
The MU algorithm is in fact a very popular algorithm for computing NMF even when minimizing the Frobenius loss. It is simple to implement a vanilla MU and easy to modify it to account for regularizations, a popular choice being the sparsity-inducing penalizations [15, 30]. There has been a number of works regarding the convergence and the implementation of MU [25, 20, 33, 29, 26], but these works do not modify the core of the algorithm: the design of an approximate diagonal Hessian matrix used to define matrix-like step sizes in gradient descent. A downside of MU, in particular for the Frobenius loss, is that its convergence can be significantly slower than other methods.
In this work, we aim at speeding up the convergence of the MU algorithm. To this end, we propose to compute different approximate diagonal Hessian matrices which are tighter approximations of the true Hessian matrices. The iterates of the proposed algorithm, which is coined fastMU, are guaranteed to converge to a stationary point of the objective function, but the convergence speed is unknown. The updates are not multiplicative anymore but instead follow the usual structure of forward-backward algorithms. We observe on both synthetic and realistic dataset that fastMU is competitive with state-of-the-art methods for the Frobenius loss, and can speed-up MU by several orders of magnitudes in Frobenius loss and KL divergence. It can however struggle when the data is very sparse or when the residuals are large.
1.1 Structure of the manuscript
Section 1 introduces the motivation and also specifies the position of our work in the topic. Section 2 covers the relevant background material on quadratic majorization technique. Section 3 provides the technical details about the Majorize Minimize framework which underpins the fastMU algorithm. Section 4 details the fastMU algorithm, the main contribution of our work, which allows for improved empirical convergence speed over MU. Section 5 shows several experiments in both simulation and real data cases with many comparisons with different well-known and classical methods in the literature. Finally, Section 6 gives the conclusion of our work and some possible perspectives for future work.
1.2 Notations
To end this section, let us introduce the following notations: and denote the Hadamard (entry-wise) product and division, respectively. denotes the diagonal matrix with entries defined from vector . Uppercase bold letters denote matrices, and lowercase bold letters denote column vectors. Vector with index denotes the th column of the corresponding matrix. For example, will be the -th column of with and . In the same way, denotes entry of matrix . For , we note (resp. ) if (resp. ). (resp. ) means that is positive semi-definite (resp. positive definite), that is, for all , (resp. ). denotes vector of ones with length and .
2 Background
2.1 Loss function
In this paper, we consider two widely used loss functions:
1) the squared Frobenius norm:
(5) |
2) the KL divergence:
(6) |
These two loss functions can be split into a sum of separable functions with respect to each column vector of the matrix . Therefore, as is commonly done in the literature, we propose to solve the subproblem of estimating in parallel for each column . More precisely,
(7) |
where
1) when is the squared Frobenius norm
(8) |
2) when is the KL divergence
(9) |
For simplicity and because of (i) the symmetry between the and updates and (ii) the separability of loss functions, only strategies proposed for updating a column of the matrix are discussed. Thus, we only consider problems of the form
(10) |
The aim of this work is to develop an efficient algorithm to minimize . A well-known technique to tackle this purpose is the Majorize-Minimize (MM) principle [16]. In the following, we will present how to build quadratic majorization functions.
2.2 Quadratic majorization functions
MM algorithms are based on the idea of iteratively constructing convex quadratic majorizing approximation functions (also called an auxiliary function [19], [17] ) of the cost function:
Definition 1
Let be a differentiable function and . Let us define, for every ,
(11) |
where is a positive semi-definite matrix and denotes the weighted Euclidean norm induced by matrix , that is, . Then, satisfies the majoration condition for at if is a quadratic majorization function of at , that is, for every .
In this work, with the purpose of achieving fast convergence, we propose a new approach to design matrix in a way that permits as large moves as possible among successive approximations of the decomposition, while still allowing an easy inversion. To build a family of majorizing functions, we resort to the following result inspired by the convergence proof of MU from Lee and Seung [19]:
Proposition 1
Let be a symmetric matrix, and , a vector with positive entries. Then, is positive semi-definite.
Proof 1
See Appendix Proof of Proposition 1.
This proposition leads to the following practical corollary.
Corollary 1
Let a convex, twice-differentiable function with continuous derivatives in an open ball around the point . Denote the Hessian matrix evaluated at . Then for any , the function
is a quadratic majorization function of at .
3 Construction of quadratic majorant functions for NMF
The choice of matrix is important since a good choice can improve the speed of convergence of the MM algorithm. Indeed, in [6, 24] the authors show that the use of judicious preconditioning matrices can significantly accelerate the convergence of the algorithm. In fact, we may see as an approximation of the Hessian of the cost function, and the choice of constitutes a trade-off between the number of iterations necessary to achieve convergence and update complexity. If is exactly the Hessian matrix, minimizing the majorant amounts to performing a Newton step method which has quadratic convergence. On the other hand, if is diagonal, it can be inverted with negligible time complexity but convergence speed may be sublinear as is the case for gradient descent. Here, we leverage Corollary 1 to build the quadratic majorant function of . We choose
(12) |
with for a well chosen vector .
Here, the Hessian matrix is easily derived:
1) when is the squared Frobenius norm
(13) |
2) when is the KL divergence
(14) |
An approximation of the Hessian is sometimes used instead for KL. In particular, within the MU framework, it is obtained by setting , leading to
(15) |
In their seminal work [18, 19], Lee and Seung proposed to choose to build the majorant of the Hessian matrix. We shell denote it by . This yields
1) when is the squared Frobenius norm:
(16) |
2) when is the KL divergence, under the approximation :
(17) |
It is shown in [19] that these choices not only yield majorizing functions for but also give a minimizer vector of with positive entries.
Although MU is an efficient and widely used algorithm in the literature, its convergence still attracts the attention of many researchers. In particular [10] produced a sequence of iterates with decreasing cost function values . However, the convergence of the cost function to a first-order stationary point is not guaranteed. Zhao and Tan proposed in [33] to add regularizers for and in the objective function to guarantee the convergence of the algorithm. Another family of approaches proposed to enforce the nonnegativity of the entries of and in (2) using and with [28, 27]. These constraints allow the MU algorithm to converge to the stationary points by avoiding various problems occurring at zero. In the same way, in the following, we assume that the factor matrices and have positive entries.
Minimization of the quadratic majorant obtained with the approximate Hessian recovers the well-known MU updates:
1) for the squared Frobenius norm:
(18) |
2) for the KL divergence111For the KL divergence, note that the MU algorithm uses an approximation of the approximate Hessian. This double approximation is justified by the simplicity of the updates obtained this way.
(19) |
At this stage, it is important to discuss the complexity of these updates. A quick analysis easily shows that in a low-rank context, costly operations are products of the form and , both involving multiplications.
The purpose of our work is to find a matrix (which boils down to finding a vector ) that improves the majorant proposed by Lee and Seung without increasing the computational cost of the updates. To this end, we notice from Proposition 1 that the approximate Hessians proposed by Lee and Seung are actually larger than the true Hessian matrices (or their approximation in the KL case). In order to achieve reduced curvature of the function and thus enables more significant moves when minimizing it, we look for tighter upper bounds of the Hessian matrices by choosing another value for . To achieve this goal, we will search for so that the diagonal values of are as small as possible. Finding such a vector can be addressed by minimizing the norm of these diagonal values:
(20) |
The following proposition shows that solutions to these problems are known in closed form.
Proposition 2
Let be a a symmetric matrix with strictly positive entries of size . The optimization problem
(21) |
has solutions in following form:
(22) |
Proof 2
We denote the cost function in (21) as and rewrite it as follows
First we can note that for any and we have . The Problem 21 can be reformulated as follows
(23) |
The Lagrangian of this problem is given by
(24) |
and its partial gradient with respect to is
We can see that for every , otherwise , which implies that for all . Now let and let by using the KKT condition, we have
and
Hence , is the only solutuion of KKT condition.
To show that these solutions are local minima, we compute the Hessian matrix:
where for each and we have
For any non-zero vector , we have
where the equality holds if and only if is a constant vector which is not belong to the tangent space of the problem. Therefore the hessian matrix at the set of vectors is a positive definite over the tangent space as long as is a symmetric and entry-wise positive matrix. Thus the proof of Proposition 2.
Finally, plugging the crafted with Proposition 2 in an approximate Hessian , we are able to provide new simple majorants of the true Hessians, denoted , for each loss function:
1) When is the squared Frobenius norm:
(25) |
2) When is the KL divergence:
(26) |
Importantly, computing either Hessian matrices in (25) or (26) involves similar quantities as in the MU updates, which are also required to compute the gradients. Therefore they are computed with little additional cost compared to classical MU Hessian matrices .
Remark 1
The upper approximation of the Hessian is of the form , with . When we minimize , we choose the element in that class of majorants with minimum mean value. We can think of other choices. For instance, picking which leads to the largest possible step in all directions,
(27) |
We can show that the solution to the this minmax problem is any eigenvector associated with the largest singular value of (see in Appendix Minmax problem for more details), this means that under this criterion, fastMU boils down to gradient descent with stepsize with the local Lipschitz constant.
4 Final proposed algorithm
In the previous section, we showed how to compute diagonal approximate Hessian matrices for updating iterates of NMF subproblems. Let us now detail how to make use of these approximate Hessian matrices. We propose to use an inexact Variable Metric Forward-Backward Algorithm [6] which is a combination between an MM strategy and a Weighted proximity operator. This framework allows to include nonnegativity constraints easily through projection, but other penalizations with tractable proximity operators could easily be considered instead. Within the scope of this work, we only consider nonnegativity constraints.
4.1 The fastMU Algorithm
We represent the following forward-backward algorithm [6, 4, 21, 3, 32] to minimize the function in Algorithm 1. At each iteration, , (resp. ) is updated with one step of gradient descent (forward step) on (resp. ) followed by a projection on the -nonnegative orthant222The set of matrices with entries larger than with . (backward step). For , denotes a “step-size”-like matrix with its -th column given by:
(28) |
With the same principle, denotes the “step-size”-like matrix for at . Moreover, and denote the partial gradients of at with respect to the variables and , respectively. We observed in preliminary experiments that fastMU with KL loss was sensitive to initialization, and therefore we recommend initializing with good estimates, e.g. refine random initialization with one iteration of the classical MU algorithm.
4.2 Convergence guarantee
It may be surprising that Algorithm 1 imposes and for a small given , typically in our implementation. In fact, this means that Algorithm 1 solves a slightly modified NMF problem where the nonnegativity constraint is replaced by the constraint “greater than ”. However, this is both a standard operation in recent versions of MU algorithms that ensures the convergence of MU iterates [29, 11] and a necessary operation for the proposed approximate Hessians to always be invertible. The clipping operator is in fact the proximity operator of the modified constraint so the proposed algorithm is still an instance of a forward-backward algorithm for plain NMF. Convergence guarantees with other regularizations are not provided in this work but should be easily obtained for a few classical regularizations such as the norm.
The convergence of Algorithm 1 when the cost is penalized with the characteristic function of the -nonnegative orthant can be derived from a general result established in [6].
Proposition 3 ([6])
.
Let and be sequences generated by Algorithm 1 with and given by (28). Assume that:
1) There exists such that, for all ,
2) Step-sizes and are chosen in the interval where and are some given positive real constants with .
Then, the sequence converges to a critical point of the problem (2). Moreover, is a nonincreasing sequence converging to .
4.3 Dynamic stopping of inner iterations
Algorithm 1 is a barebone set of instructions for the proposed fastMU algorithm. However, to minimize the time spent when running the algorithm, we suggest to use a dynamic stopping criterion for the inner iterations, inspired from the literature on Hierarchial Alternating Least Squares (HALS) for computing NMF [12], instead of a fixed number of inner iterations and .
The proof of convergence for fastMU allows any number of inner iterations to be run. But in practice, stopping inner iteration early avoids needlessly updating the last digits of a factor. The time saved by stopping the inner iterations can then be used more efficiently to update the other factor, and so on until convergence is observed.
Early stopping strategies have been described in the literature, and we use the same strategy as the accelerated HALS algorithm [12] which we describe next. Suppose factor is being updated. At each inner iteration, the squared norm of the factor update is computed. In principle, this norm should be large for the first inner iterations and should decrease toward zero as convergence occurs in the inner loop. Therefore one may stop the inner iteration if the factors update do not change much relative to the first iteration, i.e. when the following proposition is true:
(29) |
where is a tolerance value defined by the user. The optimal value of for our algorithm will be studied experimentally in Section 5. Note that this acceleration adds to the complexity of the algorithm, which is an order of magnitude smaller than the inner loop complexity.
Moreover, unless specified otherwise, we use an aggressive constant stepsize .
5 Experimental results
All the experiments and figures below can be reproduced by running the code shared on the repository attached to this project333https://github.com/cohenjer/MM-nmf.
Let us summarize the various hyperparameters of the proposed fastMU algorithm:
-
•
The loss function may be the KL divergence or the Frobenius norm, and we respectively denote the proposed algorithm by fastMU-KL and fastMU-Fro. Standard algorithms to compute NMF typically are not the same for these two losses. In particular, for the Frobenius norm, it is reported in the literature [11] that accelerated Hierarchical Alternating Least Squares [12] is state-of-the-art, while for KL divergence the standard MU algorithm proposed by Lee and Seung is still one of the best performing method [14].
-
•
The number of inner iterations needs to be tuned. A small number of inner iterations leads to faster outer loops, but a larger number of inner iterations can help decrease the cost more effectively. A dynamic inner stopping criterion has been successfully used in the literature, parameterized by a scalar defined in equation (29) that needs to be tuned.
5.1 Tuning the fastMU inner stopping hyperparameter
Below, we conduct synthetic tests in order to decide, on a limited set of experiments, what are the best choices for the hyperparameter for both supported losses. We will then fix this parameter for all subsequent comparisons.
5.1.1 Synthetic data setup
To generate synthetic data with approximately low nonnegative rank, factor matrices and are sampled elementwise from i.i.d. uniform distributions on . The chosen dimensions , and the rank vary among the experiments and are reported directly on the figures. The synthetic data is formed as where is a noise matrix also sampled elementwise from a uniform distribution over 444We use uniform noise with both KL and Frobenius loss to avoid negative entries in , even though i) these losses do not correspond to maximum likelihood estimators for this noise model and ii) uniform noise induces bias.. Given a realization of the data and the noise, the noise level is chosen so that the Signal to Noise Ratio (SNR) is fixed to a user-defined value reported on the figures or the captions. Note that no normalization is performed on the factors or the data. Initial values for all algorithms are identical and generated randomly in the same way as the true underlying factors.
The results are collected using a number of realizations of the above setup. If not specified otherwise, we chose . The maximum number of outer iterations is in general set large enough to observe convergence, by default 20000. We report loss function values normalized by the product of dimensions . Experiments are run using toolbox shootout [8] to ensure that all the intermediary results are stored and that the experiments are reproducible.
5.1.2 Early stopping tuning
In this experiment, the dynamic inner stopping tolerance is set in with a maximum number of inner iterations set to 100. When the maximum number of inner iterations is reached (i.e. 100), while for , in general, we observed only one or two inner iterations, see the median number of inner iterations in Figure 1. The total number of outer iterations here is set depending on so that the execution time of all tests are similar. Figure 2 shows the median normalized cost function against time for both the Frobenius norm and the KL divergence.
We may observe that, first, the choice of directly impacts the number of inner iterations on a per-problem basis, which validates that the dynamic inner stopping criterion stops inner iterations adaptively. Second, while there is no clear winner for in the interval and a lot of variability, we observe a significant decrease in performance outside that range. Therefore according to these tests, we may set by default for both losses. Alternatively, we could fix the maximal number of inner iterations to ten iterations since this would lead to a similar performance in this experiment.
5.2 Comparisons with other algorithms
5.2.1 Baseline algorithms
We compare our implementation of the proposed MU algorithm with the following methods:
- •
-
•
Hierarchical Alternating Least Squares (HALS), which only applies to the Frobenius loss. We used a customized version of the implementation provided in nn-fac [22].
-
•
Nesterov NMF [13] (NeNMF), which is an extension of Nesterov Fast Gradient for computing NMF. We observed that NeNMF provided divergent iterates when used to minimize the KL loss, so it is only used with the Frobenius loss in our experiments.
-
•
Vanilla alternating proximal Gradient Descent (GD) with an aggressive stepsize where is the Lipschitz constant, also only for the Frobenius loss. GD was unreasonably slow for KL in our experiments, in particular because the Lipschitz constant is not cheap to compute.
To summarize, for the Frobenius loss we evaluate the performance of MU, HALS, NeNMF, and GD against the proposed fastMU algorithm fastMU_Fro. For KL, we compare the classical MU algorithm with the proposed fastMU (fastMU_KL). Note that all methods are implemented with the same criterion for stopping inner iterations, .
Moreover, in order to further compare the proposed algorithm with the baselines, we also solve the (nonlinear) Nonnegative Least Squares (NLS) problem obtained when factor is known and fixed (to the ground-truth if known). The cost function is then convex and we expect all methods to return the same minimal error. The NLS problem is formally defined as
(30) |
5.2.2 Comparisons on synthetic dataset
Experiments with the Frobenius loss
Synthetic data are generated with the same procedure as in Section 5.1. All methods start from the same random initialization for each run. Two set of dimensions are used: and . We set the SNR to for the NMF problem, and to or for the NLS problem.
Figures 3 and 4 report the results of the Frobenius loss experiments. We may draw several conclusions from these experiments. First, fastMU has competitive performance with the baselines and in particular significantly outperforms MU. In this experiment, fastMU is also faster than HALS for the larger dimensions and rank. Second, the extrapolated algorithms NeNMF is overall the best performing method for both NMF and NLS problems.
The performance of fastMU is very close to vanilla Gradient Descent in the NLS problem. This may happen if the columns of are close to orthogonality (in which case the Hessian is approximately the identity matrix), but note that the two methods perform differently on the NMF problem where the ground truth is not provided. Gradient descent will perform less favorably on a realistic dataset where factor matrices are correlated, see Section 5.2.3.
Experiments with the KL loss
The data generation is more complex for the KL loss. We overall follow the same procedure as described in Section 5.1. However, because we observed that sparsity has an important impact on MU performance when minimizing the KL loss, we design four different dense/sparse setups:
-
•
setup dense: no sparsification555more precisely we used for the factor entries, and for the data with , which are the smallest values the algorithms can reconstruct. is applied, and the factors and the data are dense.
-
•
setup fac sparse: the smallest half of the entries of both and are set to zero.
-
•
setup data sparse: after setting , the fifty percent smallest entries of are set to zero. The NMF model is therefore only approximate and the residuals are in fact much larger than in the fac sparse and dense setups.
-
•
setup fac data sparse: we sparsify both the factors and the data with the above procedure.
In all setups, noise is added to the data, after sparsification if it applies, so that SNR is . The dimensions are for the NMF problem, and for the NLS problem to test the impact of dimensions on this simpler problem.
5.2.3 Comparisons on realistic datasets
We used the following datasets to further the comparisons:
-
•
An amplitude spectrogram of dimensions and computed from an audio excerpt in the MAPS dataset [9]. More precisely, we hand-picked the file MAPS_MUS_bach-847_AkPnBcht.wav which is a recording of Bach’s second Prelude played on a Yamaha Disklavier. Only the first thirty seconds of the recording are kept, and we also discard all frequencies above 5300Hz. This piece is a moderately difficult piano piece, and NMF has been used in this context to perform blind automatic music transcription, see [1] for more details. Taking inspiration from a model called Attack-Decay [5], in which each rank-one component corresponds to either the attack or the decay of a different key on the piano, we suppose that the amplitude spectrogram is well approximated by an NMF of rank , i.e. twice the number of keys on an acoustic piano keyboard.
-
•
An hyperspectral image of dimensions and called “Urban” which has been used extensively in the blind spectral unmixing literature for showcasing the efficiency of various NMF-based methods, see for instance the survey [2]. Blind spectral unmixing consists in recovering the spectra of the materials present in the image as well as their spatial relative concentrations. These quantities are in principle estimated as rank-one components in the NMF. It is reported that Urban has between four and six components. We therefore set .
In what follows we are not interested in the interpretability of the results since the usefulness of NMF for music automatic transcription and spectral unmixing has already been established in previous works and we do not propose any novelty in the modeling. Instead, we investigate if fastMU and its variants are faster than its competitors to compute NMF on these data.
For both dataset, a reasonable ground-truth for the factor is available, which allows us to also compare algorithms for the NLS problem as well. Indeed in the audio transcription problem, the MAPS database contains individual recordings of each key played on the Disklavier, which we can use to infer as detailed in [31]. In the blind spectral unmixing problem, we use a “fake” ground-truth obtained in [34] where six pixels containing the unmixed spectra have been hand-chosen.
Figure 7, 8, 9 and 10 respectively show the KL and Frobenius loss function values over time for the power spectrogram and the hyperspectral image. All experiments use different initializations. Since a reasonable ground truth is available for these problems, we initialized matrix as this ground-truth plus for an i.i.d. uniform noise matrix to stabilize the runs. The NMF HSI (HyperSpectral Image) problem required more outer iterations, the maximum was therefore set to 200.
The results are not entirely consistent with the observations made in the synthetic experiments:
-
•
Frobenius loss: For the audio data, HALS is the best performing method as is generally reported in the literature, followed by MU. When computing NMF, fastMU_Fro is a close competitor but do not outperform MU. Nevertheless, fastMU improves significantly over MU in the HSI experiment and is competitive with HALS. Moreover, interestingly fastMU_Fro performs significantly better than NeNMF on both realistic dataset while the two algorithms were performing similarly on the simulated dataset. On the NLS problem results are more ambiguous, but overall fastMU is not a bad competitor to HALS.
-
•
KL loss: we may observe that fastMU is significantly faster than MU in the HSI experiment, but not in the audio experiment. This may be explained in light of the synthetic experiments: HSI are dense data with sparse underlying factors, while audio data is sparse with sparse factors and large residuals. However further research on fastMU seems required to better understand how to further speed-up the method for sparse datasets when the residuals are large.
6 Conclusions
In this work, we propose a tighter upper bound of the Hessian matrix to improve the convergence of the MU algorithm for Frobenius and KL loss. The proposed algorithm coined fastMU shows promising performance on both synthetic and real-world data. While, in the experiments we conducted, fastMU is not always better than MU proposed by Lee and Seung, in many cases it is significantly faster. Moreover, it is also competitive with HALS for the Frobenius loss, which is one of the state-of-the-art methods for NMF in that setting.
There are many promising research avenues for fastMU. This paper shows how to build a family of majorant functions that contains the Lee and Seung majorant and how to compute a good majorant in that family. While we chose the best majorant in that family according to some criteria, other criterion could be explored which might produce faster algorithms; it is anticipated that an efficient algorithmic procedure to compute a better majorant could be designed. One could also search for other majorant families which are even tighter to the original objective, or families of majorants from which sampling is faster. Finally, the algorithm could be improved to better handle sparse data matrices and large residuals in particular for the KL divergence loss.
Proof of Proposition 1
It is straightforward to prove that a square matrix is positive semi-definite if and only if is a positive semi-definite matrix for any . Therefore the proof of the proposition 1 is equivalent to prove that is positive semi-definite for any . Indeed, for any with we have
This concludes the proof.
Minmax problem
This appendix is to show that the solution of minmax problem 27 is any eigenvector associated with the largest singular value of matrix . Indeed, assume that is a solution of 27, then necessarily we will show that for all ,
If this is not the case, we can show that is not optimal. Indeed, assume that and suppose that there exists an index such that
Since is a strictly decreasing function with respect to and is strictly increasing function with respect to for , we can find a vector with
and with small enough, the following in-equations hold
That means is not optimal points of minmax problem. Therefore it must hold that
which we can write as an eigenvalue problem
We look for the largest eigenvalue of such that an associated eigenvector is nonnegative. By the Perron-Frobenius theorem, the eigenvector associated with the largest eigenvalue is nonnegative (strictly if is strictly positive). Since all other eigenvectors are orthogonal to , they cannot satisfy the nonnegativity constraint. Therefore the only admissible set of solutions is , .
This means that where is the maximum eigenvalue of . FastMU is then a simple scaled gradient descent algorithm.
References
- [1] Emmanouil Benetos, Simon Dixon, Zhiyao Duan, and Sebastian Ewert. Automatic music transcription: An overview. IEEE Signal Processing Magazine, 36(1):20–30, 2018.
- [2] José M Bioucas-Dias, Antonio Plaza, Nicolas Dobigeon, Mario Parente, Qian Du, Paul Gader, and Jocelyn Chanussot. Hyperspectral unmixing overview: Geometrical, statistical, and sparse regression-based approaches. IEEE journal of selected topics in applied earth observations and remote sensing, 5(2):354–379, 2012.
- [3] J. Bolte, P. L. Combettes, and J.-C. Pesquet. Alternating proximal algorithm for blind image recovery. In 2010 IEEE International Conference on Image Processing. IEEE, sep 2010.
- [4] Jérôme Bolte, Shoham Sabach, and Marc Teboulle. Proximal alternating linearized minimization for nonconvex and nonsmooth problems. Mathematical Programming, 146(1-2):459–494, jul 2013.
- [5] Tian Cheng, Matthias Mauch, Emmanouil Benetos, Simon Dixon, et al. An attack/decay model for piano transcription. In ISMIR 2016 proceedings, 2016.
- [6] Emilie Chouzenoux, Jean-Christophe Pesquet, and Audrey Repetti. A block coordinate variable metric forward-backward algorithm. Journal of Global Optimization, pages 1–29, February 2016.
- [7] Andrzej Cichocki, Rafal Zdunek, and Shun-ichi Amari. Hierarchical ALS algorithms for nonnegative matrix and 3D tensor factorization. In International Conference on Independent Component Analysis and Signal Separation, pages 169–176. Springer, 2007.
- [8] Jeremy E. Cohen. Shootout. https://github.com/cohenjer/shootout, 2022.
- [9] Valentin Emiya, Nancy Bertin, Bertrand David, and Roland Badeau. MAPS-A piano database for multipitch estimation and automatic transcription of music. Research Report, pp.11. inria-00544155, 2010.
- [10] Cédric Févotte and Jérôme Idier. Algorithms for nonnegative matrix factorization with the -divergence. Neural Computation, 23(9):2421–2456, sep 2011.
- [11] Nicolas Gillis. Nonnegative Matrix Factorization. Society for Industrial and Applied Mathematics, Philadelphia, PA, 2020.
- [12] Nicolas Gillis and François Glineur. Accelerated multiplicative updates and hierarchical ALS algorithms for nonnegative matrix factorization. Neural computation, 24(4):1085–1105, 2012.
- [13] Naiyang Guan, Dacheng Tao, Zhigang Luo, and Bo Yuan. NeNMF: An optimal gradient method for nonnegative matrix factorization. IEEE Transactions on Signal Processing, 60(6):2882–2898, jun 2012.
- [14] Le Thi Khanh Hien and Nicolas Gillis. Algorithms for nonnegative matrix factorization with the Kullback–Leibler divergence. Journal of Scientific Computing, 87(3):1–32, 2021.
- [15] P.O. Hoyer. Non-negative sparse coding. In Proceedings of the 12th IEEE Workshop on Neural Networks for Signal Processing. IEEE, 2002.
- [16] David R. Hunter and Kenneth Lange. [optimization transfer using surrogate objective functions]: Rejoinder. Journal of Computational and Graphical Statistics, 9(1):52, mar 2000.
- [17] David R. Hunter and Kenneth Lange. A tutorial on MM algorithms. American Statistician, 58(1):30–37, February 2004.
- [18] Daniel D. Lee and H. Sebastian Seung. Learning the parts of objects by non-negative matrix factorization. Nature, 1999.
- [19] Daniel D. Lee and H. Sebastian Seung. Algorithms for non-negative matrix factorization. In Advances in Neural Information Processing Systems. MIT Press, 2001.
- [20] Valentin Leplat, Nicolas Gillis, and Jérôme Idier. Multiplicative updates for NMF with -divergences under disjoint equality constraints. SIAM Journal on Matrix Analysis and Applications, 42(2):730–752, 2021.
- [21] Z. Q. Luo and P. Tseng. On the convergence of the coordinate descent method for convex differentiable minimization. Journal of Optimization Theory and Applications, 72(1):7–35, jan 1992.
- [22] Axel Marmoret and Jérémy Cohen. nn_fac: Nonnegative factorization techniques toolbox, 2020.
- [23] Pentti Paatero and Unto Tapper. Positive matrix factorization: A non-negative factor model with optimal utilization of error estimates of data values. Environmetrics, 5(2):111–126, jun 1994.
- [24] Audrey Repetti, Mai Quyen Pham, Laurent Duval, Emilie Chouzenoux, and Jean-Christophe Pesquet. Euclid in a taxicab: Sparse blind deconvolution with smoothed regularization. IEEE Signal Processing Letters, 22(5):539–543, may 2015.
- [25] Romain Serizel, Slim Essid, and Gaël Richard. Mini-batch stochastic approaches for accelerated multiplicative updates in nonnegative matrix factorisation with beta-divergence. In 2016 IEEE 26th International Workshop on Machine Learning for Signal Processing (MLSP), pages 1–6. IEEE, 2016.
- [26] Yong Sheng Soh and Antonios Varvitsiotis. A non-commutative extension of Lee-Seung's algorithm for positive semidefinite factorizations. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 27491–27502. Curran Associates, Inc., 2021.
- [27] N. Takahashi, J. Katayama, and J. Takeuchi. A generalized sufficient condition for global convergence of modified multiplicative updates for NMF. In Int. Symp. Nonlinear Theory Appl., 2014.
- [28] Norikazu Takahashi and Ryota Hibi. Global convergence of modified multiplicative updates for nonnegative matrix factorization. Computational Optimization and Applications, 57(2):417–440, aug 2013.
- [29] Norikazu Takahashi and Masato Seki. Multiplicative update for a class of constrained optimization problems related to NMF and its global convergence. In 2016 24th European Signal Processing Conference (Eusipco), pages 438–442. IEEE, 2016.
- [30] Leo Taslaman and Björn Nilsson. A framework for regularized non-negative matrix factorization, with application to the analysis of gene expression data. PLoS ONE, 7(11):e46331, nov 2012.
- [31] Haoran Wu, Axel Marmoret, and Jeremy E. Cohen. Semi-supervised convolutive NMF for automatic piano transcription. In Sound and Music Computing 2022. Zenodo, june 2022.
- [32] Yangyang Xu and Wotao Yin. A block coordinate descent method for regularized multiconvex optimization with applications to nonnegative tensor factorization and completion. SIAM Journal on Imaging Sciences, 6(3):1758–1789, jan 2013.
- [33] Renbo Zhao and Vincent YF Tan. A unified convergence analysis of the multiplicative update algorithm for nonnegative matrix factorization. In 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 2562–2566. IEEE, 2017.
- [34] Feiyun Zhu. Hyperspectral unmixing: ground truth labeling, datasets, benchmark performances and survey. arXiv preprint arXiv:1708.05125, 2017.