Abstract
We consider the unconstrained \(L_q\)-\(L_p\) minimization: find a minimizer of \(\Vert Ax-b\Vert ^q_q+\lambda \Vert x\Vert ^p_p\) for given \(A \in R^{m\times n}\), \(b\in R^m\) and parameters \(\lambda >0\), \(p\in [0, 1)\) and \(q\ge 1\). This problem has been studied extensively in many areas. Especially, for the case when \(q=2\), this problem is known as the \(L_2-L_p\) minimization problem and has found its applications in variable selection problems and sparse least squares fitting for high dimensional data. Theoretical results show that the minimizers of the \(L_q\)-\(L_p\) problem have various attractive features due to the concavity and non-Lipschitzian property of the regularization function \(\Vert \cdot \Vert ^p_p\). In this paper, we show that the \(L_q\)-\(L_p\) minimization problem is strongly NP-hard for any \(p\in [0,1)\) and \(q\ge 1\), including its smoothed version. On the other hand, we show that, by choosing parameters \((p,\lambda )\) carefully, a minimizer, global or local, will have certain desired sparsity. We believe that these results provide new theoretical insights to the studies and applications of the concave regularized optimization problems.
References
Chartrand, R.: Exact reconstruction of sparse signals via nonconvex minimization. IEEE Signal Process. Lett. 14, 707–710 (2007)
Chartrand, R., Staneva, V.: Restricted isometry properties and nonconvex compressive sensing. Inverse Probl. 24, 1–14 (2008)
Chen, X., Xu, F., Ye, Y.: Lower bound theory of nonzero entries in solutions of \(\ell _2-\ell _p\) minimization. SIAM J. Sci. Comput. 32, 2832–2852 (2010)
Chen, X., Zhou, W.: Convergence of reweighted \(l_1\) minimization algorithms and unique solution of truncated \(l_p\) minimization. Department of Applied Mathematics, The Hong Kong Polytechnic University, Preprint (2010)
Fan, J., Li, R.: Variable selection via nonconcave penalized likelihood and its oracle properties. J. Am. Stat. Soc. 96, 1348–1360 (2001)
Foucart, S., Lai, M.J.: Sparsest solutions of under-determined linear systems via \(l_q\) minimization for \(0<q\le 1\). Appl. Comput. Harmon. Anal. 26, 395–407 (2009)
Frank, I.E., Freidman, J.H.: A statistical view of some chemometrics regression tools (with discussion). Technometrics 35, 109–148 (1993)
Garey, M.R., Johnson, D.S.: “Strong” NP-Completeness results: motivation, examples, and implications. J. Assoc. Comput. Mach. 25, 499–508 (1978)
Garey, M.R., Johnson, D.S.: Computers and Intractability; A Guide to the Theory of NP-Completeness. W. H. Freeman, New York (1979)
Ge, D., Jiang, X., Ye, Y.: A note on the complexity of \(L_p\) minimization. Math. Program. 129, 285–299 (2011)
Huang, J., Horowitz, J.L., Ma, S.: Asymptotic properties of bridge estimators in sparse high-dimensional regression models. Ann. Stat. 36, 587–613 (2008)
Knight, K., Fu, W.J.: Asymptotics for lasso-type estimators. Ann. Stat. 28, 1356–1378 (2000)
Lai, M., Wang, Y.: An unconstrained \(l_q\) minimization with \(0 < q< 1\) for sparse solution of under-determined linear systems. SIAM J. Optim. 21, 82–101 (2011)
Natarajan, B.K.: Sparse approximate solutions to linear systems. SIAM J. Comput. 24, 227–234 (1995)
Nocedal, J., Wright, S.J.: Numerical Optimization, 2nd edn. Springer, New York (2006)
Tibshirani, R.: Regression shrinkage and selection via the Lasso. J R. Stat. Soc. B 58, 267–288 (1996)
Vazirani, V.: Approximation Algorithms. Springer, Berlin (2003)
Acknowledgments
The authors would like to thank anonymous referees for their helpful comments. The authors acknowledge the support of the Hong Kong Polytechnic University and Hong Kong Research Grant Council. Research by the second author is supported by National Natural Science Foundation of China under Grants 71001062.
Author information
Authors and Affiliations
Corresponding author
Appendix
Appendix
1.1 Proof of Lemma 4
First it is easy to see that when \(p=0\), \(g(z)\) has a unique minimizer at \( z = 1\), and the optimal value is \(\frac{1}{2}\). Now we consider the case when \(p\ne 0\). Note that \(g(z)>g(0)=1\) for all \(z<0\), and \(g(z)>g(1)=\frac{1}{2}\) for all \(z>1\). Therefore the minimum point must lie within \([0,1]\).
To optimize \(g(z)\) on \([0,1]\), we check its first derivative
We have \(g^{\prime }(0^+)=+\infty \) and \(g^{\prime }(1)=\frac{p}{2}>0\). Therefore, if function \(g(z)\) has at most two stationary points in (0,1), the first one must be a local maximum and the second one must be the unique global minimum and the minimum value \(c(p,q)\) must be less than \(\frac{1}{2}\).
Now we check the possible stationary points of \(g(z)\). Consider solving \(g^{\prime }(z)=-q(1-z)^{q-1}+\frac{p z^{p-1}}{2}=0\). We get \(z^{1-p}(1-z)^{q-1}=\frac{p}{2q}\).
Define \(h(z)=z^{1-p}(1-z)^{q-1}\). We have
Note that \(\frac{1-p}{z}-\frac{q-1}{1-z}\) is decreasing in \(z\) and must have a root on \((0,1)\). Therefore, there exists a point \(\bar{z}\in (0,1)\) such that \(h^{\prime }(z)>0\) for \(z<\bar{z}\) and \(h^{\prime }(z)<0\) for \(z>\bar{z}\). This implies that \(h(z)=\frac{p}{2q}\) can have at most two solutions in \((0,1)\), i.e., \(g(z)\) can have at most two stationary points. By the previous discussions, the lemma holds. \(\square \)
1.2 Proof of Theorem 6
We again consider the same \(3\)-partition problem, we claim that it can be reduced to a minimization problem in form (6). Again, it suffices to only consider the case when \(\lambda = \frac{1}{2}\) (Here we consider the hardness result for any given \(\epsilon > 0\). Note that after the scaling, \(\epsilon \) may have changed). Consider:
We have
The last equality comes from the submodularity of the function \((x+\epsilon )^p\) and the fact that one can always choose only one of \(x_{ij}\) to be nonzero in each set such that the equality holds. Consider function \(g_{\epsilon }(z)=|1-z|^q+\frac{1}{2} (|z|+\epsilon )^p\). Similar to Lemma 4, one can prove that \(g_{\epsilon }(z)\) has a unique minimizer in \([0,1]\). Denote this minimum value by \(c(p, q, \epsilon )\), we know that \(P_{\epsilon }(x)\ge n c(p, q, \epsilon )\). Then we can argue that the 3-partition problem has a solution if and only if \(P_{\epsilon }(x)= n c(p, q, \epsilon )\). Therefore Theorem 6 holds.\(\square \)
Rights and permissions
About this article
Cite this article
Chen, X., Ge, D., Wang, Z. et al. Complexity of unconstrained \(L_2-L_p\) minimization. Math. Program. 143, 371–383 (2014). https://doi.org/10.1007/s10107-012-0613-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10107-012-0613-0
Keywords
- Nonsmooth optimization
- Nonconvex optimization
- Variable selection
- Sparse solution reconstruction
- Bridge estimator