Abstract
Low-rank matrix recovery (LRMR) is to recover the underlying low-rank structure from the degraded observations, and has myriad formulations, for example robust principal component analysis (RPCA). As a core component of LRMR, the low-rank approximation model attempts to capture the low-rank structure by approximating the \(\ell _0\)-norm of all the singular values of a low-rank matrix, i.e., the number of the non-zero singular values. Towards this purpose, this paper develops a low-rank approximation model by jointly combining a parameterized hyperbolic tangent (tanh) function with the continuation process. Specificially, the continuation process is exploited to impose the parameterized tanh function to approximate the \(\ell _0\)-norm. We then apply the proposed low-rank model to RPCA and refer to it as tanh-RPCA. Convergence analysis on optimization, and experiments of background subtraction on seven challenging real-world videos show the efficacy of the proposed low-rank model through comparing tanh-RPCA with several state-of-the-art methods.
X. Zhang and Y. Gao—Contributed equally to this work and this work was supported by the National Key Research and Development Program of China [2016YFB0200401], the National Natural Science Foundation of China [U1435222] and the National High-tech R&D Program [2015AA020108].
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
F-measure is the harmonic average of the precision and recall, with the best value at 1 (perfect precision and recall) and the worst at 0.
References
Beck, A., Teboulle, M.: A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2(1), 183–202 (2009)
Candès, E.J., Li, X., Ma, Y., Wright, J.: Robust principal component analysis? J. ACM 58(3), 11 (2011)
Candès, E.J., Recht, B.: Exact matrix completion via convex optimization. Found. Comput. Math. 9(6), 717 (2009)
Cao, Z., Long, M., Wang, J., Yu, P.S.: HashNet: deep learning to hash by continuation. In: IEEE International Conference on Computer Vision, pp. 5609–5618 (2017)
Chen, C.F., Wei, C.P., Wang, Y.C.F.: Low-rank matrix recovery with structural incoherence for robust face recognition. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2618–2625 (2012)
Fan, Q., Jiao, Y., Lu, X.: A primal dual active set algorithm with continuation for compressed sensing. IEEE Trans. Sig. Process. 62(23), 6276–6285 (2014)
Gao, Y., Cai, H., Zhang, X., Lan, L., Luo, Z.: Background subtraction via 3D convolutional neural networks. In: International Conference on Pattern Recognition. IEEE (2018)
Gong, P., Ye, J., Zhang, C.: Multi-stage multi-task feature learning. J. Mach. Learn. Res. 14(1), 2979–3010 (2013)
Gu, S., Xie, Q., Meng, D., Zuo, W., Feng, X., Zhang, L.: Weighted nuclear norm minimization and its applications to low level vision. Int. J. Comput. Vis. 121(2), 183–208 (2017)
Guo, X., Lin, Z., Center, C.M.I.: Route: robust outlier estimation for low rank matrix recovery. In: International Joint Conference on Artificial Intelligence, pp. 1746–1752 (2017)
Hu, Y., Zhang, D., Ye, J., Li, X., He, X.: Fast and accurate matrix completion via truncated nuclear norm regularization. IEEE Trans. Pattern Anal. Mach. Intell. 35(9), 2117–2130 (2013)
Jin, Z.F., Wan, Z., Jiao, Y., Lu, X.: An alternating direction method with continuation for nonconvex low rank minimization. J. Sci. Comput. 66(2), 849–869 (2016)
Kang, Z., Peng, C., Cheng, Q.: Robust PCA via nonconvex rank approximation. In: IEEE International Conference on Data Mining, pp. 211–220 (2015)
Kang, Z., Peng, C., Cheng, Q.: Robust subspace clustering via smoothed rank approximation. IEEE Sig. Process. Lett. 22(11), 2088–2092 (2015)
Kang, Z., Peng, C., Cheng, Q.: Robust subspace clustering via tighter rank approximation. In: ACM International on Conference on Information and Knowledge Management, pp. 393–401 (2015)
LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436 (2015)
Lewis, A.S., Sendov, H.S.: Nonsmooth analysis of singular values. Part i: theory. Set-Valued Anal. 13(3), 213–241 (2005)
Li, L., Huang, W., Gu, I.Y.H., Tian, Q.: Statistical modeling of complex backgrounds for foreground object detection. IEEE Trans. Image Process. 13(11), 1459–1472 (2004)
Lin, Z., Ganesh, A., Wright, J., Wu, L., Chen, M., Ma, Y.: Fast convex optimization algorithms for exact recovery of a corrupted low-rank matrix. J. Marine Biol. Assoc. UK 56(3), 707–722 (2009)
Lin, Z., Liu, R., Su, Z.: Linearized alternating direction method with adaptive penalty for low-rank representation. In: Advances in Neural Information Processing Systems, pp. 612–620 (2011)
Lu, C., Tang, J., Yan, S., Lin, Z.: Generalized nonconvex nonsmooth low-rank minimization. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 4130–4137 (2014)
Mairal, J., Bach, F., Ponce, J., Sapiro, G.: Online learning for matrix factorization and sparse coding. J. Mach. Learn. Res. 11(1), 19–60 (2010)
Mu, Y., Dong, J., Yuan, X., Yan, S.: Accelerated low-rank visual recovery by random projection. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2609–2616 (2011)
Peng, C., Kang, Z., Li, H., Cheng, Q.: Subspace clustering using log-determinant rank approximation. In: ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 925–934 (2015)
Peng, C., Kang, Z., Yang, M., Cheng, Q.: RAP: scalable RPCA for low-rank matrix recovery. In: Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, pp. 2113–2118 (2016)
Recht, B., Fazel, M., Parrilo, P.A.: Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization. SIAM Rev. 52(3), 471–501 (2010)
Shen, X., Wu, Y.: A unified approach to salient object detection via low rank matrix recovery. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 853–860 (2012)
Toyama, K., Krumm, J., Brumitt, B., Meyers, B.: Wallflower: principles and practice of background maintenance. In: IEEE International Conference on Computer Vision, pp. 255–261 (1999)
Wang, N., Yao, T., Wang, J., Yeung, D.-Y.: A probabilistic approach to robust matrix factorization. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7578, pp. 126–139. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-33786-4_10
Wright, J., Ganesh, A., Rao, S., Peng, Y., Ma, Y.: Robust principal component analysis: exact recovery of corrupted low-rank matrices via convex optimization. In: Advances in Neural Information Processing Systems, pp. 2080–2088 (2009)
Xiang, S., Tong, X., Ye, J.: Efficient sparse group feature selection via nonconvex optimization. In: International Conference on Machine Learning, pp. 284–292 (2013)
Xin, B., Tian, Y., Wang, Y., Gao, W.: Background subtraction via generalized fused lasso foreground modeling. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 4676–4684 (2015)
Yuan, X., Yang, J.: Sparse and low rank matrix decomposition via alternating direction method. Pac. J. Optim. 9(1) (2009)
Zhang, C.H., et al.: Nearly unbiased variable selection under minimax concave penalty. Ann. Stat. 38(2), 894–942 (2010)
Zhang, T.: Analysis of multi-stage convex relaxation for sparse regularization. J. Mach. Learn. Res. 11(3), 1081–1107 (2010)
Zhang, Z., Yan, S., Zhao, M.: Similarity preserving low-rank representation for enhanced data representation and effective subspace learning. Neural Netw. 53, 81–94 (2014)
Zhang, Z., Zhao, M., Li, F., Zhang, L., Yan, S.: Robust alternating low-rank representation by joint lp-and l2, p-norm minimization. Neural Netw. 96, 55–70 (2017)
Zheng, Y., Liu, G., Sugimoto, S., Yan, S., Okutomi, M.: Practical low-rank matrix approximation under robust \(\ell_1\)-norm. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1410–1417 (2012)
Zou, W., Kpalma, K., Liu, Z., Ronsin, J.: Segmentation driven low-rank matrix recovery for saliency detection. In: British Machine Vision Conference, pp. 1–13 (2013)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Zhang, X., Gao, Y., Lan, L., Guo, X., Huang, X., Luo, Z. (2018). Low-Rank Matrix Recovery via Continuation-Based Approximate Low-Rank Minimization. In: Geng, X., Kang, BH. (eds) PRICAI 2018: Trends in Artificial Intelligence. PRICAI 2018. Lecture Notes in Computer Science(), vol 11012. Springer, Cham. https://doi.org/10.1007/978-3-319-97304-3_43
Download citation
DOI: https://doi.org/10.1007/978-3-319-97304-3_43
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-97303-6
Online ISBN: 978-3-319-97304-3
eBook Packages: Computer ScienceComputer Science (R0)