skip to main content
note

Active Two Phase Collaborative Representation Classifier

Published:02 July 2019Publication History
Skip Abstract Section

Abstract

The Sparse Representation Classifier, the Collaborative Representation Classifier (CRC), and the Two Phase Test Sample Sparse Representation (TPTSSR) classifier were introduced in recent times. All these frameworks are supervised and passive in the sense that they cannot benefit from unlabeled data samples. In this paper, inspired by active learning paradigms, we introduce an active CRC that can be used by these frameworks. More precisely, we are interested in the TPTSSR framework due to its good performance and its reasonable computational cost. Our proposed Active Two Phase Collaborative Representation Classifier (ATPCRC) starts by predicting the label of the available unlabeled samples. At testing stage, two coding processes are carried out separately on the set of originally labeled samples and the whole set (original and predicted label). The two types of class-wise reconstruction errors are blended in order to decide the class of any test image. Experiments conducted on four public image datasets show that the proposed ATPCRC can outperform the classic TPTSSR as well as many state-of-the-art methods that exploit label and unlabeled data samples.

References

  1. Mohamad M. Al Rahhal, Yakoub Bazi, Haikel AlHichri, Naif Alajlan, Farid Melgani, and Ronald R. Yager. 2016. Deep learning approach for active classification of electrocardiogram signals. Information Sciences 345 (2016), 340--354. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Rana Aamir Raza Ashfaq, Xi-Zhao Wang, Joshua Zhexue Huang, Haider Abbas, and Yu-Lin He. 2017. Fuzziness based semi-supervised learning approach for intrusion detection system. Information Sciences 378 (2017), 484--497. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Mikhail Belkin, Partha Niyogi, and Vikas Sindhwani. 2006. Manifold regularization: A geometric framework for learning from labeled and unlabeled examples. Journal of Machine Learning Research 7, November (2006), 2399--2434. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Deng Cai, Xiaofei He, and Jiawei Han. 2007. Semi-supervised discriminant analysis. In Proceedings of the IEEE International Conference on Conputer Vision. 1--7.Google ScholarGoogle ScholarCross RefCross Ref
  5. Xiaojun Chang, Zhigang Ma, Ming Lin, Yi Yang, and Alexander Hauptmann. 2017. Feature interaction augmented sparse learning for fast kinect motion detection. IEEE Transactions on Image Processing 26, 8 (2017), 3911--3920.Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Sanjoy Dasgupta and Daniel Hsu. 2008. Hierarchical sampling for active learning. In Proceedings of the 25th International Conference on Machine Learning. ACM, 208--215. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. J. Dems̆ar. 2006. Statistical comparisons of classifiers over multiple data sets. Journal of Machine Learning Research 7 (2006), 1--30. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Fadi Dornaika, Radouan Dahbi, Alireza Bosaghzadeh, and Yassine Ruichek. 2017. Efficient dynamic graph construction for inductive semi-supervised learning. Neural Networks 94 (2017), 192--203.Google ScholarGoogle ScholarCross RefCross Ref
  9. Fadi Dornaika and Youssof El Traboulsi. 2016. Learning flexible graph-based semi-supervised embedding. IEEE Transactions on Cybernetics 46, 1 (2016), 206--218.Google ScholarGoogle ScholarCross RefCross Ref
  10. Fadi Dornaika, Youssof El Traboulsi, and Ammar Assoum. 2013. Adaptive two phase sparse representation classifier for face recognition. In Proceedings of the International Conference on Advanced Concepts for Intelligent Vision Systems. Springer, 182--191.Google ScholarGoogle ScholarCross RefCross Ref
  11. Fadi Dornaika and Youssof El Traboulsi. 2017. Matrix exponential based semi-supervised discriminant embedding. Pattern Recognition 61 (2017), 92--103. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Thomas Drugman, Janne Pylkkönen, and Reinhard Kneser. 2016. Active and semi-supervised learning in ASR: Benefits on the acoustic and language models. Interspeech (2016), 2318--2322.Google ScholarGoogle Scholar
  13. Youssof El Traboulsi, Fadi Dornaika, and Ammar Assoum. 2015. Kernel flexible manifold embedding for pattern classification. Neurocomputing 167 (2015), 517--527. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Retrieved from http://www.statisticssolutions.com/manova-analysis-paired-sample-ttest/. {n.d.}.Google ScholarGoogle Scholar
  15. Hong Huang, Jiamin Liu, and Yinsong Pan. 2012. Semi-supervised marginal Fisher analysis for hyperspectral image classification. ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences 3 (2012), 377--382.Google ScholarGoogle Scholar
  16. Sheng-Jun Huang, Rong Jin, and Zhi-Hua Zhou. 2010. Active learning by querying informative and representative examples. In Proceedings of the 23rd International Conference on Neural Information Processing Systems. 892--900. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Abiodun Iwayemi and Chi Zhou. 2017. Saraa: Semi-supervised learning for automated residential appliance annotation. IEEE Transactions on Smart Grid 8, 2 (2017), 779--786.Google ScholarGoogle Scholar
  18. Sina Jafarpour, Weiyu Xu, Babak Hassibi, and Robert Calderbank. 2009. Efficient and robust compressed sensing using optimized expander graphs. IEEE Transactions on Information Theory 55, 9 (2009), 4299--4308. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Prateek Jain and Ashish Kapoor. 2009. Active learning for large multi-class problems. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR’09). IEEE, 762--769.Google ScholarGoogle ScholarCross RefCross Ref
  20. Ajay J. Joshi, Fatih Porikli, and Nikolaos Papanikolopoulos. 2009. Multi-class active learning for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR’09). IEEE, 2372--2379.Google ScholarGoogle ScholarCross RefCross Ref
  21. Chun-Guang Li, Jun Guo, and Hong-Gang Zhang. 2010. Local sparse representation based classification. In Proceedings of the 20th International Conference on Pattern Recognition (ICPR’10). IEEE, 649--652.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Wei Liu, Dacheng Tao, and Jianzhuang Liu. 2008. Transductive component analysis. In Proceedings of the 8th IEEE International Conference on Data Mining (ICDM’08). IEEE, 433--442. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Hieu T. Nguyen and Arnold Smeulders. 2004. Active learning using pre-clustering. In Proceedings of the 21st International Conference on Machine Learning. ACM, 79. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Feiping Nie, Dong Xu, Ivor Wai-Hung Tsang, and Changshui Zhang. 2010. Flexible manifold embedding: A framework for semi-supervised and unsupervised dimension reduction. IEEE Transactions on Image Processing 19, 7 (2010), 1921--1932. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Lishan Qiao, Songcan Chen, and Xiaoyang Tan. 2010. Sparsity preserving discriminant analysis for single training image face recognition. Pattern Recognition Letters 31, 5 (2010), 422--429. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Jing Wang, Canyi Lu, Meng Wang, Peipei Li, Shuicheng Yan, and Xuegang Hu. 2014. Robust face recognition via adaptive sparse representation. IEEE Transactions on Cybernetics 44, 12 (2014), 2368--2378.Google ScholarGoogle ScholarCross RefCross Ref
  27. John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma. 2009. Robust face recognition via sparse representation. IEEE Transactions on Pattern Analysis and Machine Intelligence 31, 2 (2009), 210--227. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Yong Xu, David Zhang, Jian Yang, and Jing-Yu Yang. 2011. A two-phase test sample sparse representation method for use with face recognition. IEEE Transactions on Circuits and Systems for Video Technology 21, 9 (2011), 1255--1262.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Active Two Phase Collaborative Representation Classifier

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image ACM Transactions on Knowledge Discovery from Data
        ACM Transactions on Knowledge Discovery from Data  Volume 13, Issue 4
        August 2019
        235 pages
        ISSN:1556-4681
        EISSN:1556-472X
        DOI:10.1145/3343141
        Issue’s Table of Contents

        Copyright © 2019 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 2 July 2019
        • Accepted: 1 April 2019
        • Received: 1 December 2018
        Published in tkdd Volume 13, Issue 4

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • note
        • Research
        • Refereed

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format .

      View HTML Format