Skip to main content

Neural Network for the Statistical Process Control of HVAC Systems in Passenger Rail Vehicles

  • Conference paper
  • First Online:
Studies in Theoretical and Applied Statistics (SIS 2021)

Part of the book series: Springer Proceedings in Mathematics & Statistics ((PROMS,volume 406))

Included in the following conference series:

  • 375 Accesses

Abstract

In the rail industry, coach temperature regulation has become a crucial task to improve passenger thermal comfort. Over the past few years, European standards have required rail operators to implement monitoring systems for the control of heating, ventilation and air conditioning (HVAC) of passenger rail vehicles. These systems, based on modern automated sensing technologies, have created new data-rich scenarios and call for new methods to deal with high-dimensional, high-correlated and heterogeneous data. In this article, an autoencoder, which is a particular type of neural network developed to model unlabelled data and automatically extract significant features, is utilised to develop a nonparametric process monitoring approach. Two control charts based on statistics \(H^2\) and SPE are built in the feature space and the residual space, respectively. Through operational HVAC data collected on board passenger vehicles, the proposed approach is shown to be capable of simultaneously monitoring and detecting anomalies that may have occurred in the data streams acquired from each train coach, even though it is not limited to the application hereby investigated. Additionally, via a numerical investigation, the Phase II fault detection performance is compared with that of a simpler linear dimension reduction method and two more complex NN architectures.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 189.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 249.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 249.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. UNI-EN 14750-1: Railway Applications—Air Conditioning for Urban and Suburban Rolling Stock. Part 1: Comfort Parameters. British Standard. British Standards Institution, London (2006)

    Google Scholar 

  2. Montgomery, D.C.: Statistical Quality Control. Wiley Global Education (2012)

    Google Scholar 

  3. James, G., Witten, D., Hastie, T., Tibshirani, R.: An Introduction to Statistical Learning, vol. 112. Springer (2013)

    Google Scholar 

  4. Venkatasubramanian, V., Rengaswamy, R., Yin, K., Kavuri, S.N.: A review of process fault detection and diagnosis: Part I: Quantitative model-based methods. Comput. Chem. Eng. 27(3), 293–311 (2003)

    Google Scholar 

  5. Bersimis, S., Psarakis, S., Panaretos, J.: Multivariate statistical process control charts: an overview. Qual. Reliab. Eng. Int. 23(5), 517–543 (2007)

    Article  Google Scholar 

  6. Zhang, Z., Jiang, T., Zhan, C., Yang, Y.: Gaussian feature learning based on variational autoencoder for improving nonlinear process monitoring. J. Process Control 75, 136–155 (2019)

    Article  Google Scholar 

  7. Lee, S., Kwak, M., Tsui, K.-L., Kim, S.B.: Process monitoring using variational autoencoder for high-dimensional nonlinear processes. Eng. Appl. Artif. Intell. 83, 13–27 (2019)

    Google Scholar 

  8. Chen, L., Wang, Z.-Y., Qin, W.-L., Ma, J.: Fault diagnosis of rotary machinery components using a stacked denoising autoencoder-based health state identification. Signal Process. 130, 377–388 (2017)

    Article  Google Scholar 

  9. Zorriassatine, F., Tannock, J.D.T.: A review of neural networks for statistical process control. J. Intell. Manuf. 9(3), 209–224 (1998)

    Article  Google Scholar 

  10. Psarakis, S.: The use of neural networks in statistical process control charts. Qual. Reliab. Eng. Int. 27(5), 641–650 (2011)

    Article  Google Scholar 

  11. Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT Press (2016)

    Google Scholar 

  12. Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning representations by back-propagating errors. Nature 323(6088), 533–536 (1986)

    Google Scholar 

  13. Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pp. 249–256. JMLR Workshop and Conference Proceedings (2010)

    Google Scholar 

  14. Prechelt, L.: Early stopping-but when? In: Neural Networks: Tricks of the Trade, pp. 55–69. Springer (1998)

    Google Scholar 

  15. Baldi, P., Hornik, K.: Neural networks and principal component analysis: learning from examples without local minima. Neural Netw. 2(1), 53–58 (1989)

    Article  Google Scholar 

  16. Jolliffe, I.: Principal Component Analysis. Encyclopedia of Statistics in Behavioral Science (2005)

    Google Scholar 

  17. Japkowicz, N., Hanson, S.J., Gluck, M.A.: Nonlinear autoassociation is not equivalent to PCA. Neural Comput. 12(3), 531–545 (2000)

    Google Scholar 

  18. Jackson, J.E., Mudholkar, G.S.: Control procedures for residuals associated with principal component analysis. Technometrics 21(3), 341–349 (1979)

    Google Scholar 

  19. Yan, W., Guo, P., Li, Z., et al.: Nonlinear and robust statistical process monitoring based on variant autoencoders. Chemom. Intell. Lab. Syst. 158, 31–40 (2016)

    Article  Google Scholar 

  20. MacGregor, J.F., Kourti, T.: Statistical process control of multivariate processes. Control Eng. Pract. 3(3), 403–414 (1995)

    Google Scholar 

  21. Lahiri, S.N.: Theoretical comparisons of block bootstrap methods. Ann. Stat. 386–404 (1999)

    Google Scholar 

  22. Bühlmann, P., Künsch, H.R.: Block length selection in the bootstrap for time series. Comput. Stat. Data Anal. 31(3), 295–310 (1999)

    Google Scholar 

  23. Härdle, W., Horowitz, J., Kreiss, J.-P.: Bootstrap methods for time series. Int. Stat. Rev. 71(2), 435–459 (2003)

    Article  MathSciNet  MATH  Google Scholar 

  24. Phaladiganon, P., Kim, S.B., Chen, V.C.P., Baek, J.-G., Park, S.-K.: Bootstrap-based t2 multivariate control charts. Commun. Stat. Simul. Comput. ® 40(5), 645–662 (2011)

    Google Scholar 

  25. Silverman, B.W.: Density Estimation for Statistics and Data Analysis. Routledge (2018)

    Google Scholar 

  26. Lehmann, E.L., Romano, J.P., Casella, G.: Testing Statistical Hypotheses, vol. 3. Springer (2005)

    Google Scholar 

  27. Iannone, F., Ambrosino, F., Bracco, G., De Rosa, M., Funel, A., Guarnieri, G., Migliori, S., Palombi, F., Ponti, G., Santomauro, G., Procacci, P.: CRESCO ENEA HPC clusters: a working example of a multifabric GPFS spectrum scale layout. In: 2019 International Conference on High Performance Computing Simulation (HPCS), pp. 1051–1052 (2019)

    Google Scholar 

  28. Van Rossum, G., Drake Jr., F.L.: Python reference manual. Centrum voor Wiskunde en Informatica Amsterdam (1995)

    Google Scholar 

  29. Chollet, F.: Keras. https://keras.io (2015)

  30. Abadi, M., et al.: TensorFlow: Large-scale machine learning on heterogeneous systems. Software available from www.tensorflow.org (2015)

Download references

Acknowledgements

This work has been developed in the framework of the R &D project of the multiregional investment programme “REINForce: REsearch to INspire the Future” (CDS000609) with Hitachi Rail Italy, supported by the Italian Ministry for Economic Development (MISE) through the Invitalia agency. The authors are extremely grateful to engineers Vincenzo Criscuolo and Guido Cesaro, from the Operation Service and Maintenance Product Evolution Department of Hitachi Rail STS, for their technological insights in the interpretation of results. The computing resources and the related technical support used for this work have been provided by CRESCO/ENEAGRID High Performance Computing infrastructure and its staff [27]. CRESCO/ENEAGRID High Performance Computing infrastructure is funded by ENEA, the Italian National Agency for New Technologies, Energy and Sustainable Economic Development and by Italian and European research programmes, see http://www.cresco.enea.it/english for information.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Gianluca Sposito .

Editor information

Editors and Affiliations

Appendix

Appendix

In this Appendix, the Phase II fault detection performance of the proposed method is compared with that of PCA, AE with more than one hidden layer, referred to as DeepAE, and Variational AE, referred to as VAE [6, 11]. The PCA is regarded as a simpler linear dimension reduction method and is used as the benchmark to prove the effectiveness of non-linear approaches. Whereas DeepAE and VAE are regarded as alternative to the proposed NN with more a complex architecture.

Note that, to allow for a fair comparison, the proposed AE, the DeepAE and the VAE must be chosen with a number of code layer nodes equal to the number of components retained in the PCA. In this regard, we compare the performance of the proposed AE with that of the competing methods by setting the code layer dimension hyperparameter equal to 2 and 12. The former turned out indeed to be the optimal number of extracted features by PCA, as they explained the \(82\%\) of the total variance [16]. Whereas, the latter is the optimal choice for the number of code layer nodes for the proposed AE, as already reported in Table 2. Also in this case, as stated in Sect. 2.3 for the proposed AE, the \(H^2\) and SPE statistics and the relative UCLs for all competing methods are estimated by means of the block bootstrap with \(\alpha = 0.05\). The Phase II fault detection performance is explored by means of a numerical investigation based on four different Phase II scenarios. The latter is based on data generated by resampling and transforming \(N_s = 10000\) observations from the training set in order to simulate Phase II non-normal operating conditions with the following severity levels:

  1. 1.

    the means of the variables \(T_\text {Return}\) and \(T_\text {Supply}\) of one coach, say coach # 1, are shifted by an amount \(\Delta \) equal to two units of the relative standard deviation, i.e., \(\Delta = 2\sigma \) (Scenario 1);

  2. 2.

    the means of the variables \(T_\text {Return}\) and \(T_\text {Supply}\) of one coach are shifted by \(\Delta = 3\sigma \) (Scenario 2);

  3. 3.

    the means of the variables \(T_\text {Return}\) and \(T_\text {Supply}\) of two coaches, say coach # 1 and # 2, are shifted by \(\Delta = 2\sigma \) (Scenario 3);

  4. 4.

    the means of the variables \(T_\text {Return}\) and \(T_\text {Supply}\) of two coaches are shifted by \(\Delta = 3\sigma \) (Scenario 4).

Trivially, if either the \(H^2\) or SPE statistics exceeds the value of the corresponding UCL, the fault is counted as successfully detected, otherwise, no alarm is signalled. Then, the Phase II performance is measured through the FDR index, calculated as \(\text {FDR = TN}/N_s \times 100 \%\), where TN is the number of fault samples correctly identified. FDRs of the four competing methods are reported in Tables 3 and 4 for a number of code layer nodes equal to 2 and 12, respectively.

Table 3 FDR (%) for a code layer dimension hyperparameter equal to 2 in the four scenarios described in Sect. 2.3. The best performance is highlighted in bold
Table 4 FDR (%) for a code layer dimension hyperparameter equal to 12 in the four scenarios described in Sect. 2.3. The best performance is highlighted in bold

In particular, Table 3 demonstrates that, even when we choose the best latent representation for the PCA, the latter is outperformed by the proposed method, which in fact performs better also than DeepAE and VAE, in all considered scenarios.

From Table 4, when the code layer dimension is set equal to 12, the Phase II performance of the proposed AE, although outperforming the PCA in all scenarios, is instead outperformed by VAE in Scenario 2 and DeepAE in Scenario 3 and 4. This indicates that the proposed AE is still the best in detecting small deviations from normal behaviours with respect to more complex NN architectures. It is however clear that in all cases PCA underperforms all competing non-linear approaches.

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ambrosino, F., Giannini, G., Lepore, A., Palumbo, B., Sposito, G. (2022). Neural Network for the Statistical Process Control of HVAC Systems in Passenger Rail Vehicles. In: Salvati, N., Perna, C., Marchetti, S., Chambers, R. (eds) Studies in Theoretical and Applied Statistics . SIS 2021. Springer Proceedings in Mathematics & Statistics, vol 406. Springer, Cham. https://doi.org/10.1007/978-3-031-16609-9_23

Download citation

Publish with us

Policies and ethics