Skip to main content

Hierarchical Multi-view Fisher Discriminant Analysis

  • Conference paper
Neural Information Processing (ICONIP 2009)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 5864))

Included in the following conference series:

Abstract

Fisher discriminant analysis (FDA) has exhibited its great power to improve the performance in classification and dimensionality reduction tasks. Objects in the real world often have more than one natural feature set and therefore they often can be described by more than one views. However, traditional FDA addresses all problems with a single view. In this paper we propose multi-view FDA (MFDA) which combines traditional FDA with multi-view learning. In order to improve the performance of MFDA for multi-class case, we further propose hierarchical MFDA which combines MFDA with hierarchical metric learning. Experiments are performed on many artificial and real-world data sets. Comparisons with the single-view FDA show the effectiveness of the proposed method.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Duda, R.O., Hart, P.E., Stork, D.: Pattern Classification, 2nd edn. John Wiley and Sons, New York (2001)

    MATH  Google Scholar 

  2. He, Q.P., Qin, S.J., Wang, J.: A New Fault Diagnosis Method Using Fault Directions in Fisher Discriminant Analysis. AIChE Journal 51(2), 555–571 (2009)

    Article  Google Scholar 

  3. Xu, Y., Yang, J., Jin, Z.: A Novel Method for Fisher Discriminant Analysis. Pattern Recognition 37(2), 381–384 (2004)

    Article  MATH  Google Scholar 

  4. Duchene, J., Leclercq, S.: An Optimal Transformation for Discriminant and Principal Component Analysis. IEEE Transaction on Pattern Analysis and Machine Intelligence 10(6), 978–983 (1988)

    Article  MATH  Google Scholar 

  5. Bo, L.F., Wang, L., Jiao, L.C.: Feature Scaling for Kernel Fisher Discriminant Analysis Using Leave-One-Out Cross Validation. Neural Computation 18(4), 961–978 (2006)

    Article  MATH  MathSciNet  Google Scholar 

  6. Kim, S.J., Magnani, A., Boyd, S.: Optimal Kernel Selection in Kernel Fisher Discriminant Analysis. ACM International Conference Proceeding Series 148, 465–472 (2006)

    Google Scholar 

  7. Sugiyama, M.: Dimensionality Reduction of Multimodal Labeled Data by Local Fisher Discriminant Analysis. The Journal of Machine Learning Research 8, 1027–1061 (2007)

    Google Scholar 

  8. Sun, S.: Semantic features for multi-view semi-supervised and active learning of text classification. In: Proceedings of the IEEE International Conference on Data Mining Workshops, pp. 731–735 (2008)

    Google Scholar 

  9. Sun, S.: High reliable multi-view semi-supervised learning with extremely sparse labeled data. In: Proceedings of 8th International Conference on Hybrid Intelligent Systems (HIS), pp. 935–938 (2008)

    Google Scholar 

  10. Diethe, T., Hardoon, D.R., Shawe-Taylor, J.: Multiview Fisher Discriminant Analysis. In: NIPS Workshop Learning from Multiple Sources (2008)

    Google Scholar 

  11. Hotelling, H.: Relations Between Two Sets of Variates. Biometrika 28, 312–377 (1936)

    Google Scholar 

  12. Sugiyama, M.: Local Fisher Discriminant Analysis for Supervised Dimensionality Reduction. In: Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, Pennsylvania, pp. 905–912 (2006)

    Google Scholar 

  13. Domeniconi, C., Gunopulos, D., Peng, J.: Large Margin Nearest Neighbor Classifiers. IEEE Transactions on Neural Networks 16(4), 899–900 (2005)

    Article  Google Scholar 

  14. Domeniconi, C., Peng, J., Gunopulos, D.: Locally Adaptive Metric Nearest-Neighbor Classificaton. IEEE Pattern Analysis and Machine Intelligence 24(9), 1281–1285 (2002)

    Article  Google Scholar 

  15. Friedman, J.H.: Flexible Metric Nearest Neighbor Classification. Technical Report, Stanford University (1994)

    Google Scholar 

  16. Hastie, T., Tibshirani, R.: Discriminant Adaptive Nearest Neighbor Classification. IEEE Pattern Analysis and Machine Intelligence 18(6), 607–616 (1996)

    Article  Google Scholar 

  17. Peng, J., Heisterkamp, D.R., Dai, H.K.: Adaptive Kernel Metric Nearest Neighbor Classificaton. In: Proceedings of 16th International Conference on Pattern Recognition, Quebec, Canada, pp. 33–36 (2002)

    Google Scholar 

  18. Yang, L., Jin, R., Sukthankar, R., Liu, Y.: An Efficient Algorithm for Local Distance Metric Learning. In: Proceedings of the National Conference on Artificial Intelligence (2006)

    Google Scholar 

  19. Chen, Q., Sun, S.: Hierarchical Distance Metric Learning for Large Margin Nearest Neighbor Classification. Submitted to International Journal of Pattern Recognition and Artificial Intelligence

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2009 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Chen, Q., Sun, S. (2009). Hierarchical Multi-view Fisher Discriminant Analysis. In: Leung, C.S., Lee, M., Chan, J.H. (eds) Neural Information Processing. ICONIP 2009. Lecture Notes in Computer Science, vol 5864. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-10684-2_32

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-10684-2_32

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-10682-8

  • Online ISBN: 978-3-642-10684-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics