skip to main content
10.1145/3443467.3443842acmotherconferencesArticle/Chapter ViewAbstractPublication PageseitceConference Proceedingsconference-collections
Article

Data-driven Key Features Selection for Fault Detection in a Complex System

Authors Info & Claims
Published:01 February 2021Publication History

Editorial Notes

The editors have requested minor, non-substantive changes to the VoR and, in accordance with ACM policies, a Corrected VoR was published on March 5, 2021. For reference purposes the VoR may still be accessed via the Supplemental Material section on this page.

ABSTRACT

Fault detection is a difficult but important problem for a complex system. This paper presents a fault detection method based on data-driven key feature selection for the complex system abbreviated as FD-DKFS. By regarding the observable parameters as original features, FD-DKFS first finds the missing correlations among original features and constructs potentially useful features for fault detection. Next, FD-DKFS provides a filter feature selection method to find the best feature subset for fault detection. Then, these selected features are used to detect the fault in a certain complex system with conventional classifiers. Compared with the other methods, the results of the experiment show that the proposed method is more accurate for fault detection in the complex system.

Skip Supplemental Material Section

Supplemental Material

References

  1. Mei, S., Zarrabi, N., Lees, M., Sloot, P. M. A. (2015). Complex agent networks: an emerging approach for modeling complex systems. Applied Soft Computing, 37, 311--321.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Liu, Z., Jia, Z., Vong, C. M., Bu, S., Han, J., & Tang, X.. (2017). Capturing high-discriminative fault features for electronics-rich analog system via deep learning. IEEE Transactions on Industrial Informatics, 1213--1226.Google ScholarGoogle ScholarCross RefCross Ref
  3. Betelin, V. B., Eskov, V. M., Galkin, V. A., & Gavrilenko, T. V.. (2017). Stochastic volatility in the dynamics of complex homeostatic systems. Doklady Mathematics, 95(1), 92--94.Google ScholarGoogle ScholarCross RefCross Ref
  4. Li, D., Zhou, Y., Hu, G., & Spanos, C. J.. (2016). Optimal sensor configuration and feature selection for ahu fault detection and diagnosis. IEEE Transactions on Industrial Informatics, 13(3), 1369--1380.Google ScholarGoogle ScholarCross RefCross Ref
  5. Wang, G., & Yin, S.. (2017). Quality-related fault detection approach based on orthogonal signal correction and modified pls. IEEE Transactions on Industrial Informatics, 11(2), 398--405.Google ScholarGoogle Scholar
  6. Wang, G., Yin, S., & Kaynak, O.. (2014). An lwpr-based data-driven fault detection approach for nonlinear process monitoring. Industrial Informatics IEEE Transactions on, 10(4), 2016--2023.Google ScholarGoogle ScholarCross RefCross Ref
  7. Guyon, I., & Elisseeff, André. (2003). An introduction to variable and feature selection. Journal of Machine Learning Research, 3(6), 1157--1182.Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Chen, Q., Zhang, M., & Xue, B.. (2017). Genetic Programming with Embedded Feature Construction for High-Dimensional Symbolic Regression. Intelligent and Evolutionary Systems. Springer International Publishing.Google ScholarGoogle Scholar
  9. Liu, H., & Motoda, H.. (1999). Feature extraction construction and selection: a data mining perspective. Journal of the American Statistical Association, 94(448), 014004.Google ScholarGoogle Scholar
  10. Piramuthu, S., and Sikora, R. T. (2009). Iterative feature construction for improving inductive learning algorithms. Expert Systems with Applications, 36(2), 3401--3406.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Smith, M. G., & Bull, L.. (2005). Genetic programming with a genetic algorithm for feature construction and selection. Genetic Programming & Evolvable Machines, 6(3), 265--281.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Tang, J., Chai, T. Y., Yu, W., Liu, Z., Zhou, X. J. (2016). A comparative study that measures ball mill load parameters through different single-scale and multiscale frequency spectra-based approaches. IEEE Transactions on Industrial Informatics. 12(6), 2008--2019.Google ScholarGoogle ScholarCross RefCross Ref
  13. Georgoulas, G., Climente, V., Antonino-Daviu, J. A., Tsoumas, I., Stylios, C., Arkkio, A. and Nikolakopoulos, G. (2017). The use of a multi-label classification framework for the detection of broken bars and mixed eccentricity faults based on the start-up transient. IEEE Transactions on Industrial Informatics. 13(2), 625--634.Google ScholarGoogle ScholarCross RefCross Ref
  14. Rajput, D. S., Singh, P. K. and Bhattacharya, M. (2015). PROFIT: A Projected Clustering Technique. Real World Data Mining Applications, Springer International Publishing, 51--70.Google ScholarGoogle Scholar
  15. Liu, Z., Zhao, X., Zuo, M. J., & Xu, H.. (2014). Feature selection for fault level diagnosis of planetary gearboxes. Advances in Data Analysis & Classification, 8(4), 377--401.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Al-Otaibi, R., Jin, N., Wilcox, T., & Flach, P.. (2017). Feature construction and calibration for clustering daily load curves from smart-meter data. IEEE Transactions on Industrial Informatics, 12(2), 645--654.Google ScholarGoogle ScholarCross RefCross Ref
  17. Lillywhite, K., Lee, D. J., Tippetts, B., & Archibald, J.. (2013). A feature construction method for general object recognition. Pattern Recognition, 46(12), 3300--3314.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Saeys, Y., Inza, I. and Larraaga, P. (2007). A review of feature selection techniques in bioinformatics. Bioinformatics, 23(19), 2507--2517.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Chen, G. and Chen, J. (2015). A novel wrapper method for feature selection and its applications. Neurocomputing, 159(1), 219--226.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Khammassi, C., and Krichen, S. (2017). A GA-LR Wrapper Approach for Feature Selection in Network Intrusion Detection. Computers & Security, 70, 255--277.Google ScholarGoogle ScholarCross RefCross Ref
  21. Ferreira, A. J. and Figueiredo, M. A. T. (2014). Incremental filter and wrapper approaches for feature discretization. Neurocomputing, 123, 60--74.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Wang, A., An, N., Chen, G., Li, L., & Alterovitz, G.. (2015). Accelerating wrapper-based feature selection with k-nearest-neighbor. Knowledge-Based Systems, 83(1), 81--91.Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Zhang, K., Li, Y., Scarf, P. and Ball A. (2011). Feature selection for high-dimensional machinery fault diagnosis data using multiple models and Radial Basis Function networks. Neurocomputing, 74(17), 2941--2952.Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Yang, Y., Liao, Y., Meng, G. and Lee, J. (2011). A hybrid feature selection scheme for unsupervised learning and its application in bearing fault diagnosis. Expert Systems with Applications. 38(9), 11311--11320.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Yang, P., Zhou, B. B., Zhang, Z. and Zomaya, A. Y. (2010). A multi-filter enhanced genetic ensemble system for gene selection and sample classification of microarray data, BMC Bioinformatics, 11, 1--12.Google ScholarGoogle ScholarCross RefCross Ref
  26. Zhao, X., Zuo, M. J. and Patel, T. (2010). EMD, ranking mutual information and PCA based condition monitoring. In: Proceedings of ASME 2010 international design engineering technical conferences. Montreal, Canada, 777--782.Google ScholarGoogle ScholarCross RefCross Ref
  27. Wu, B., Abbott, T., Fishman, D., Mcmurray, W., Mor, G., & Stone, K., et al. (2003). Comparison of statistical methods for classification of ovarian cancer using mass spectrometry data. Bioinformatics, 19 (13), 1636--1643.Google ScholarGoogle ScholarCross RefCross Ref
  28. Freeman, C., Dana Kulić, & Basir, O.. (2015). An evaluation of classifier-specific filter measure performance for feature selection. Pattern Recognition. 48(5), 1812--1826.Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. García-Torres, M., Gómez-Vela, F., Melián-Batista, B. and Moreno-Vega, J. M. (2016). High-dimensional feature selection via feature grouping: A Variable Neighborhood Search approach. Information Sciences, 326, 102--118.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Yu, L., Liu, H. (2004). Efficient feature selection via analysis of relevance and redundancy, The Journal of Machine Learning Research, 5 (12): 1205--1224.Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Hall, M. (2000) Correlation-based feature selection for discrete and numeric class machine learning, in: Proceedings of the 17th International Conference on Machine Learning, Morgan Kaufmann, San Francisco, CA, 359--366.Google ScholarGoogle Scholar
  32. Duval, M., & Depabla, A.. (2002). Interpretation of gas-in-oil analysis using new iec publication 60599 and iec tc 10 databases. Electrical Insulation Magazine IEEE, 17(2), 31--41.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Data-driven Key Features Selection for Fault Detection in a Complex System

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      EITCE '20: Proceedings of the 2020 4th International Conference on Electronic Information Technology and Computer Engineering
      November 2020
      1202 pages
      ISBN:9781450387811
      DOI:10.1145/3443467

      Copyright © 2020 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 1 February 2021

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • Article
      • Research
      • Refereed limited

      Acceptance Rates

      EITCE '20 Paper Acceptance Rate214of441submissions,49%Overall Acceptance Rate508of972submissions,52%
    • Article Metrics

      • Downloads (Last 12 months)2
      • Downloads (Last 6 weeks)0

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader