Skip to main content
Log in

Manifold contraction for semi-supervised classification

  • Research Papers
  • Published:
Science China Information Sciences Aims and scope Submit manuscript

Abstract

The generalization ability of classification is often closely related to both the intra-class compactness and the inter-class separability. Owing to the fact that many current dimensionality reduction methods, regarded as a pre-processor, often lead to the poor classification performance on real-life data, in this paper, a new data preprocessing technique called manifold contraction (MC) is proposed for the classification-oriented learning task. The main motivation behind MC lies in seeking a proper mapping of contracting the given multiple-manifold data such that the ratio of the intra-class to the inter-class scatters is minimized. Moreover, in order to properly control the contraction level in MC, an adaptive MC (AMC) criterion is developed in the semi-supervised setting. Due to its generality, MC can be not only applied in original space and reproducing kernel Hilbert space (RKHS), but also easily incorporated with dimensionality reduction method for further improvement of classification performance. The final experimental results show that MC, as a data preprocessor, is effective and promising in the subsequent classification learning, especially in small-size labeled sample case.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Cherkassky V. Model complexity control and statistical learning theory. Nat Comput, 2006, 1: 109–133

    Article  MathSciNet  Google Scholar 

  2. Duin R P W, Pekalska E. Object representation, sample size and data complexity. In: Basu M, Ho T K, eds. Data Complexity in Pattern Recognition. London: Springer, 2006. 25–47

    Chapter  Google Scholar 

  3. Bishop M. Neural Networks for Pattern Recognition. Oxford: Oxford University Press, 1995

    Google Scholar 

  4. Yan S, Xu D, Zhang B, et al. Graph embedding and extension: a general framework for dimensionality reduction. IEEE Trans Patt Anal Mach Intell, 2007, 29: 40–51

    Article  Google Scholar 

  5. Hotelling H. Analysis of a complex of statistical variables into principle components. J Educat Psych, 1933, 24: 417–441

    Article  Google Scholar 

  6. Fisher R A. The use of multiple measurements in taxonomic problem. Annal Eugen, 1936, 7: 179–188

    Google Scholar 

  7. Cai D, He X, Han J. Semi-supervised discriminant analysis. In: Proceedings of IEEE International Conference on Computer Vision (ICCV′07), Rio de Janeiro, Brazil, 2007. 1–7

  8. Zhang D, Zhou Z H, Chen S C. Semi-supervised dimensionality reduction. In: Proceedings of the 7th SIAM International Conference on Data Mining (ICDM′07), Minneapolis, MN, 2007. 629–634

  9. Schölkopf B, Smola A J, Müller K R. Nonlinear component analysis as a kernel eigenvalue problem. Neur Comput, 1998, 10: 1299–1319

    Article  Google Scholar 

  10. Roweis S T, Saul L K. Nonlinear dimensionality reduction by locally linear embedding. Science, 2000, 290: 2323–2326

    Article  Google Scholar 

  11. Tenenbaum J B, de Silva V, Langford J C. A global geometric framework for nonlinear dimensionality reduction. Science, 2000, 290: 2319–2323

    Article  Google Scholar 

  12. Lafon S, Lee A B. Diffusion maps and coarse-graining: a unified framework for dimensionality reduction, graph partitioning, and data set parameterization. IEEE Trans Patt Anal Mach Intell, 2006, 28: 1393–1403

    Article  Google Scholar 

  13. Belkin M, Niyogi P. Laplacian eigenmaps for dimensionality reduction and data representation. Neur Comput, 2003, 15: 1373–1396

    Article  MATH  Google Scholar 

  14. Chapelle O, Schölkopf B, Zien A. Analysis of Benchmarks. Semi-Supervised Learning. Chapelle O, Schölkopf B, Zien A, eds. New York: MIT Press, 2006. 377–390

    Google Scholar 

  15. Van der Maaten L J P, Postma E O, van den Herik H J. Dimensionality reduction: A comparative review. Neurocomputing, 2008, published online

  16. Zhu X J. Semi-supervised learning literature survey. Computer Sciences TR 1530, University of Wisconsin-Madison, Last modified on July 19, 2008

  17. Belkin M, Niyogi P, Sindhwani V. Manifold regularization: a geometric framework for learning from labeled and unlabeled examples. J Mach Learn Res, 2006, 7: 2399–2434

    MathSciNet  Google Scholar 

  18. Szummer M, Jaakkola T. Partially labeled classification with markov random walks. Adv Neur Inf Process Syst, 2002, 15: 945–952

    Google Scholar 

  19. He J R, Li M J, Zhang H J, et al. Manifold-ranking based image retrieval. In: Proceedings of 12th ACM Multimedia, New York: ACM, 2004. 9–16

    Chapter  Google Scholar 

  20. Zhou D, Bousquet O, Lal T N, et al. Learning with local and global consistency. Adv Neur Inf Process Syst, 2004, 16: 321–328

    Google Scholar 

  21. Zhou D, Weston J, Gretton A, et al. Ranking on data manifold. Adv Neur Inf Process Syst, 2004, 17: 169–176

    Google Scholar 

  22. Zhu X J, Ghahramani Z, Lafferty J. Semi-supervised learning using gaussian fields and harmonic functions. In: Proceedings of the 20th International Conferences on Machine Learning (ICML′03), Washington, DC, 2003. 912–919

  23. Bengio Y, Delalleanu O, Roux N L. Label Propagation and Quadratic Criterion. Semi-Supervised Learning. Chapelle O, Schölkopf B, Zien A, eds. New York: MIT Press, 2006. 193–215

    Google Scholar 

  24. Wang F, Zhang C S. Label propagation through linear neighborhoods. IEEE Trans Knowl Data Eng, 2008, 20: 55–67

    Article  Google Scholar 

  25. Keerthi S S, Shevade S K, Bhattacharyya C, et al. A fast iterative nearest point algorithm for support vector machine classifier design. IEEE Trans Neur Netw, 2000, 11: 124–136

    Article  Google Scholar 

  26. Vapnik V N. The Nature of Statistical Learning Theory. 2nd ed. New York: Springer-Verlag, 1995

    MATH  Google Scholar 

  27. Tsang W I, Kocsor A, Kwok J T. Efficient kernel feature extraction for massive data sets. In: Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD′06), Philadelphia, PA, USA, 2006. 724–729

  28. Breitenbach M, Grudic G. Clustering through ranking on manifolds. In: Proceedings of the 22th International Conferences on Machine Learning (ICML’05), New York: ACM, 2005. 73–80

    Chapter  Google Scholar 

  29. Porkaew K, Chakrabarti K, Mehrotra S. Query refinement for multimedia retrieval and its evaluation techniques in MARS. In: ACM International Multimedia Conference, Orlando, Florida, 1999. 235–238

  30. Zhou Z H, Chen K J, Dai H B. Enhancing relevance feedback in image retrieval using unlabeled data. ACM Trans Inf Syst, 2006, 24: 219–244

    Article  Google Scholar 

  31. Van der Maaten L J P. An Introduction to Dimensionality Reduction Using Matlab. Technical Report MICC 07-07. Maastricht University, Maastricht, The Netherlands, 2007

    Google Scholar 

  32. Levina E, Bickel P J. Maximum likelihood estimation of intrinsic dimension. Adv Neur Inf Process Syst, 2004, 17: 777–784

    Google Scholar 

  33. Fukunaga K. Olsen D R. An algorithm for finding intrinsic dimensionality of data. IEEE Trans Comput, 1971, C-20: 176–183

    Article  Google Scholar 

  34. Bengio Y, Larochelle H, Vincent V. Non-local manifold parzen windows. Adv Neur Inf Process Syst, 2005, 18: 115–122

    Google Scholar 

  35. Vincent P, Bengio Y. Manifold parzen windows. Adv Neur Inf Process Syst, 2002, 15: 825–832

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to EnLiang Hu or SongCan Chen.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Hu, E., Chen, S. & Yin, X. Manifold contraction for semi-supervised classification. Sci. China Inf. Sci. 53, 1170–1187 (2010). https://doi.org/10.1007/s11432-010-0066-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11432-010-0066-0

Keywords

Navigation