Skip to main content
Log in

CO emission predictions in municipal solid waste incineration based on reduced depth features and long short-term memory optimization

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

Carbon monoxide (CO) is a toxic gas emitted during municipal solid waste incineration (MSWI). Its emission prediction is conducive to pollutant reduction and optimized control of MSWI. The variables of MSWI exhibit redundant and interdependent correlations with CO emissions. Furthermore, the mapping relationship is difficult to characterize. Therefore, the work proposed a CO emission prediction method based on reduced depth features and long short-term memory (LSTM) optimization. The particle design for reduced depth feature and LSTM optimization was initially developed—incorporating an adaptive threshold range for feature selection based on the inherent characteristics of modeling data. Secondly, the nonlinear depth features were extracted using ultra-one-dimensional convolution and subsequently fed into an LSTM model for prediction construction. The hyperparameters of the convolutional layer and LSTM were updated based on the loss function. The generalization performance of the model was used as the fitness function of the optimization. Finally, the particle swarm optimization (PSO) was used to adaptively reduce depth features and model’s hyperparameters. The rationality and effectiveness of the proposed method were validated using the benchmark dataset and CO dataset of MSWI. R2 of the testing datasets for RB and CO were 0.9097 ± 3.64E-04 and 0.7636 ± 3.19E-03, respectively, by repeating 30 times.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Algorithm 1
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Data availability

The data that support the findings of the work are available upon request from the corresponding author, [Jian Tang, freeflytang@bjut.edu.cn], upon reasonable request.

References

  1. Gómez-Sanabria A, Kiesewetter G, Klimont Z, Schoepp W, Haberl H (2022) Potential for future reductions of global GHG and air pollutants from circular waste management systems. Nat Commun 13(1):106

    ADS  PubMed  PubMed Central  Google Scholar 

  2. Chen A, Chen JR, Cui J, Fan C, Han W (2019) Research on risks and countermeasures of “cities besieged by waste” in China-an empirical analysis based on DIIS. Bull Chin Acad Sci 34(7):797–806

    Google Scholar 

  3. Tang J, Xia H, Yu W, Qiao JF (2023) Research status and prospects of intelligent optimization control for municipal solid waste incineration process. Acta Autom Sin. https://doi.org/10.16383/j.aas.c220810

    Article  Google Scholar 

  4. Vilardi G, Verdone N (2022) Exergy analysis of municipal solid waste incineration processes: the use of O2-enriched air and the oxy-combustion process. Energy 239:122147

    CAS  Google Scholar 

  5. Xia H, Tang J, Aljerf L, Wang TZ, Qiao JF, Xu Q, Ukaogo P (2023) Investigation on dioxins emission characteristic during complete maintenance operating period of municipal solid waste incineration. Environ Pollut 318:120949

    CAS  PubMed  Google Scholar 

  6. Li WS, Yan DH, Li L, Wen ZY, Liu MJ, Lu SX, Huang QF (2023) Review of thermal treatments for the degradation of dioxins in municipal solid waste incineration fly ash: proposing a suitable method for large-scale processing. Sci Total Environ 875:162565

    ADS  CAS  PubMed  Google Scholar 

  7. Wang B, Wang P, Xie LH, Lin RB, Lv J, Li JR, Chen B (2019) A stable zirconium based metal-organic framework for specific recognition of representative polychlorinated dibenzo-p-dioxin molecules. Nat Commun 10(1):3861

    ADS  PubMed  PubMed Central  Google Scholar 

  8. Qiao JF, Guo ZH, Tang J (2020) Dioxin emission concentration measurement approaches for municipal solid wastes in cineration process: a survey. Acta Autom Sin 46(6):1063–1089

    Google Scholar 

  9. Tang J, Qiao JF (2019) Soft sensor of dioxin emission concentration in solid waste incineration process based on selective ensemble kernel learning algorithm. J Chem Eng Technol 70(2):696–706

    CAS  Google Scholar 

  10. Hu HL, Wen XF, Luo QM (2009) Waste incineration: Best available techniques for integrated pollution prevention and control. Chemical Industry Press, Benijing

    Google Scholar 

  11. Chai TY (2016) Industrial process control systems: research status and development direction. Sci Sin Inf 46(8):1003–1015

    Google Scholar 

  12. Martínez JH, Romero S, Ramasco JJ, Estrada E (2022) The world-wide waste web. Nat Commun 13(1):1–13

    ADS  Google Scholar 

  13. Wang T, Leung H, Zhao J, Wang W (2020) Multiseries featural LSTM for partial periodic time-series prediction: a case study for steel industry. IEEE Trans Instrum Meas 69(9):5994–6003

    ADS  CAS  Google Scholar 

  14. Huda RK, Banka H (2019) Efficient feature selection and classification algorithm based on PSO and rough sets. Neural Comput Appl 31:4287–4303

    Google Scholar 

  15. Tang J, Qiao JF, Xu Z, Guo ZH (2021) Soft measuring approach of dioxin emission concentration in municipal solid waste incineration process based on feature reduction and selective ensemble algorithm. Control Theory Appl 38(1):110–120

    Google Scholar 

  16. Akinola OO, Ezugwu AE, Agushaka JO, Zitar RA, Abualigah L (2022) Multiclass feature selection with metaheuristic optimization algorithms: a review. Neural Comput Appl 34(22):19751–19790

    PubMed  PubMed Central  Google Scholar 

  17. Coelho F, Braga AP, Verleysen M (2010) Multi-objective semi-supervised feature selection and model selection based on pearson’s correlation coefficient. Springer, Berlin, pp 509–516

    Google Scholar 

  18. Battiti R (1994) Using mutual information for selecting features in supervised neural net learning. IEEE Trans Neural Netw 5(4):537–550

    CAS  PubMed  Google Scholar 

  19. Vergara JR, Estévez PA (2014) A review of feature selection methods based on mutual information. Neural Comput Appl 24(1):175–186

    Google Scholar 

  20. Jain AK, Duin RPW, Mao J (2000) Statistical pattern recognition: A review. IEEE Trans Pattern Anal Mach Intell 22(1):4–37

    Google Scholar 

  21. Qiao JF, Guo ZH, Tang J (2021) Soft sensing of dioxin emission concentration in solid waste incineration process based on multi-layer feature selection. Inf Control 50(1):75–87

    Google Scholar 

  22. Fleuret F (2004) Fast binary feature selection with conditional mutual information. J Mach Learn Res 5:1531–1555

    MathSciNet  Google Scholar 

  23. Xie S, Hua Y, Lu S, Li X (2023) A novel spatio-temporal adaptive prediction modelling strategy for industrial production process. IEEE Trans Instrum Meas 72:1–11

    Google Scholar 

  24. Lin JD, Wu XY, Chai Y, Yin HP (2020) Structure optimization of convolutional neural networks: a survey. Acta Autom Sin 46(1):24–37

    Google Scholar 

  25. Kiranyaz S, Ince T, Hamila R, Gabbouj M (2015) Convolutional neural networks for patient-specific ECG classification. In: 2015 37th Annual international conference of the IEEE engineering in medicine and biology society (EMBC), pp 2608–2611

  26. Wan B, Zhou X, Zheng B, Yin H, Zhu Z, Wang H, Yan C (2023) LFRNet: localizing, focus, and refinement network for salient object detection of surface defects. IEEE Trans Instrum Meas 72:1–12

    Google Scholar 

  27. Bejani MM, Ghatee M (2021) A systematic review on overfitting control in shallow and deep neural networks. Artif Intell Rev 54:1–48

    Google Scholar 

  28. He Y, Song K, Meng Q, Yan Y (2019) An end-to-end steel surface defect detection approach via fusing multiple hierarchical features. IEEE Trans Instrum Meas 69(4):1493–1504

    ADS  Google Scholar 

  29. Vinayakumar R, Soman KP, Poornachandran P (2017) Secure shell (ssh) traffic analysis with flow based features using shallow and deep networks. In: 2017 International conference on advances in computing, communications and informatics (ICACCI), pp 2026–2032

  30. Zhang Y, Qin N, Huang D, Du J (2022) Precise diagnosis of unknown fault of high-speed train bogie using novel FBM-net. IEEE Trans Instrum Meas 71:1–11

    Google Scholar 

  31. Abdeljaber O, Sassi S, Avci O, Kiranyaz S, Ibrahim AA, Gabbouj M (2018) Fault detection and severity identification of ball bearings by online condition monitoring. IEEE Trans Industr Electron 66(10):8136–8147

    Google Scholar 

  32. Yang D, Lu J, Zhou Y, Dong H (2022) Establishment of leakage detection model for oil and gas pipeline based on VMD-MD-1DCNN. Eng Res Express 4(2):025051

    ADS  Google Scholar 

  33. Cao J, He Z, Wang J, Yu P (2020) An anti-noise fault diagnosis method of bearing based on multi-scale 1DCNN. Preprints.org

  34. Mtibaa F, Nguyen KK, Azam M, Papachristou A, Venne JS, Cheriet M (2020) LSTM-based indoor air temperature prediction framework for HVAC systems in smart buildings. Neural Comput Appl 32:17569–17585

    Google Scholar 

  35. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780

    CAS  PubMed  Google Scholar 

  36. Zhang J, Zhu Y, Zhang X, Ye M, Yang J (2018) Developing a long short-term memory (LSTM) based model for predicting water table depth in agricultural areas. J Hydrol 561:918–929

    Google Scholar 

  37. Kong W, Dong ZY, Jia Y, Hill DJ, Xu Y, Zhang Y (2017) Short-term residential load forecasting based on LSTM recurrent neural network. IEEE Trans Smart Grid 10(1):841–851

    Google Scholar 

  38. Chondrodima E, Pelekis N, Pikrakis A, Theodoridis Y (2023) An efficient LSTM neural network-based framework for vessel location forecasting. IEEE Trans Intell Trans Syst. https://doi.org/10.1109/TITS.2023.3247993

    Article  Google Scholar 

  39. Pisa I, Morell A, Vicario JL, Vilanova R (2020) Denoising autoencoders and LSTM-based artificial neural networks data processing for its application to internal model control in industrial environments—the wastewater treatment plant control case. Sensors 20(13):3743

    ADS  CAS  PubMed  PubMed Central  Google Scholar 

  40. Zha WS, Liu YP, Wan YJ, Luo RL, Li DL, Yang S, Xu Y (2022) Forecasting monthly gas field production based on the CNN-LSTM model. Energy 260:124889

    CAS  Google Scholar 

  41. Cui CL, Tang J, Xia H, Qiao JF, Yu W (2023) Virtual sample generation method based on generative adversarial fuzzy neural network. Neural Comput Appl 35(9):6979–7001

    Google Scholar 

  42. Tang J, Xia H, Zhang J, Qiao JF, Yu W (2021) Deep forest regression based on cross-layer full connection. Neural Neural Comput Appl 33:9307–9328

    Google Scholar 

  43. Xia H, Tang J, Qiao JF, Zhang J, Yu W (2022) DF classification algorithm for constructing a small sample size of data-oriented DF regression model. Neural Comput Appl 34:2785–2810

    Google Scholar 

  44. Xia H, Tang J, Cui CL, Qiao JF (2023) Soft sensing method of dioxin emission in municipal solid waste incineration process based on broad hybrid forest regression. Acta Autom Sin 49(2):343–365

    Google Scholar 

  45. Qiao JF, Sun J, Meng X (2022) Event-triggered adaptive model predictive control of oxygen content for municipal solid waste incineration process. IEEE Trans Autom Sci Eng. https://doi.org/10.1109/TASE.2022.3227918

    Article  Google Scholar 

  46. Ruma JF, Adnan MSG, Dewan A, Rahman RM (2023) Particle swarm optimization based LSTM networks for water level forecasting: a case study on Bangladesh river network. Res Eng 17:100951

    Google Scholar 

  47. Yang X, Maihemuti B, Simayi Z, Saydi M, Na L (2022) Prediction of glacially derived runoff in the muzati river watershed based on the PSO-LSTM model. Water. https://doi.org/10.3390/w14132018

    Article  Google Scholar 

  48. Zhou H, Zuo Y, Li T, Li S (2021) Application of PSO-LSTM for forecasting of ship traffic flow. In: 2021 International conference on security, pattern analysis, and cybernetics (SPAC), pp 298–302

  49. Liang JJ, Qin AK, Suganthan PN, Baskar S (2006) Comprehensive learning particle swarm optimizer for global optimization of multimodal functions. IEEE Trans Evol Comput 10(3):281–295

    Google Scholar 

  50. Wang DD, Tang J, Xia H, Qiao JF (2022) Virtual sample generation method based on hybrid optimization with multi-objective PSO. Acta Autom Sin 45(x):1–22

    CAS  Google Scholar 

  51. Zhuang JB, Tang J, Aljerf L (2022) Comprehensive review on mechanism analysis and numerical simulation of municipal solid waste incineration process based on mechanical grate. Fuel 320:123826

    CAS  Google Scholar 

  52. Arthur JR (1951) Reactions between carbon and oxygen. Trans Faraday Soc 47:164–178

    CAS  Google Scholar 

  53. Hasberg W, May H, Dorn I (1989) Description of the residence-time behaviour and burnout of PCDD, PCDF and other higher chlorinated aromatic hydrocarbons in industrial waste incineration plants. Chemosphere 19:1–6

    Google Scholar 

  54. Blake CL, Merz CJ (1998) UCI repository of machine learning databases, Dept. Inf. Comput. Sci., Univ. California, Irvine, Irvine, CA, USA, vol. 55. [Online], available: http://archive.ics.uci.edu/ml/datasets.html, January 1, 2022

  55. Xu W, Tang J, Xia H, Qiao JF (2022) Soft sensor of dioxin emission concentration based on Bagging semi-supervised deep forest regression. Chin J Sci Instrum 43(6):251–259

    Google Scholar 

Download references

Acknowledgements

The work was financially supported by the National Key Research and Development Program of China (Grant Nos. 2021ZD0112301 and 2021ZD0112302).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jian Tang.

Ethics declarations

Conflict of interest

The authors declare no conflict of interest. The work is considered for publication on the understanding that the article has neither been published nor will be published anywhere else before being published in Neural Computing and Applications.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendices

Appendix A

Calculation process for the LSTM model

\(N = 1\) and \({\mathbf{h}}_{n - 1} = 0\) at the initial sample calculation. The specific process is shown as follows:

(1) The empty set \({\mathbf{C}}_{n}\) is the memory unit. Forget gate \({\mathbf{f}}_{n}\) determines the output ratio of the sample on \({\mathbf{C}}_{n}\) to the input.

$$ {\mathbf{F}}_{n} = {\mathbf{f}}_{n} \odot {\mathbf{C}}_{n} $$
(35)
$$ {\mathbf{f}}_{n} = \sigma \left( {{\mathbf{U}}_{{\text{f}}}^{{}} \cdot {\mathbf{h}}_{n - 1} + {\mathbf{W}}_{{\text{f}}}^{{}} \cdot {\mathbf{X}}_{n}^{{{\text{exa}}}} + {\mathbf{b}}_{{\text{f}}}^{{}} } \right) $$
(36)

where \({\mathbf{U}}_{{\text{f}}}\) and \({\mathbf{W}}_{{\text{f}}}\) are the weight of \({\mathbf{h}}_{n - 1}\) and \({\mathbf{X}}_{n}^{{{\text{exa}}}}\), respectively; \({\mathbf{b}}_{{\text{f}}}\) is the bias; \(\sigma \left( \cdot \right)\) is the sigmoid activation function; \(\odot\) is the Hadamard product.

(2) \({\tilde{\mathbf{C}}}_{n}\) is the candidate value for calculating the current input state.

$$ {\tilde{\mathbf{C}}}_{n} = {\text{tanh}}\left( {{\mathbf{U}}_{{\text{c}}}^{{}} \cdot {\mathbf{h}}_{n - 1} + {\mathbf{W}}_{{\text{c}}}^{{}} \cdot {\mathbf{X}}_{n}^{{{\text{exa}}}} + {\mathbf{b}}_{{\text{c}}}^{{}} } \right) $$
(37)

where \({\mathbf{U}}_{{\text{c}}}\) and \({\mathbf{W}}_{{\mathbf{c}}}\) are the weight of \({\mathbf{h}}_{n - 1}\) and \({\mathbf{X}}_{n}^{{{\text{exa}}}}\), respectively; \({\mathbf{b}}_{{\mathbf{c}}}\) is the bias; \(\tanh \left( \cdot \right)\) is the tanh activation function.

(3) Input gate \({\mathbf{i}}_{n}\) controls the proportion of the nth sample input stored to \({\mathbf{C}}_{n}\).

$$ {\mathbf{i}}_{n} = \sigma \left( {{\mathbf{U}}_{{\text{i}}}^{{}} \cdot {\mathbf{h}}_{n - 1} + {\mathbf{W}}_{{\text{i}}}^{{}} \cdot {\mathbf{X}}_{n}^{{{\text{exa}}}} + {\mathbf{b}}_{{\text{i}}}^{{}} } \right) $$
(38)
$$ {\mathbf{I}}_{n} = {\mathbf{i}}_{n} \odot {\tilde{\mathbf{C}}}_{n} $$
(39)

where \({\mathbf{U}}_{{\text{i}}}\) and \({\mathbf{W}}_{{\text{i}}}\) are the weight of \({\mathbf{h}}_{n - 1}\) and \({\mathbf{X}}_{n}^{{{\text{exa}}}}\), respectively; \({\mathbf{b}}_{{\text{i}}}\) is bias.

(4) The resulting value is stored in \({\mathbf{C}}_{n}\).

$$ {\mathbf{C}}_{n} = {\mathbf{F}}_{n} + {\mathbf{I}}_{n} $$
(40)

The first sample output \({\mathbf{C}}_{n}\) is \({\mathbf{I}}_{n}\), i.e., \({\mathbf{C}}_{n} = {\mathbf{I}}_{n} ,{\mathbf{F}}_{n} = 0\).

(5) The nth sample \({\mathbf{h}}_{n}\) and the proportion of \({\mathbf{o}}_{n}\) stored in output gate control \({\mathbf{C}}_{n}\) are calculated.

$$ {\mathbf{h}}_{n} = {\mathbf{o}}_{n} \odot \tanh \left( {{\mathbf{C}}_{n} } \right) $$
(41)
$$ {\mathbf{o}}_{n} = \sigma \left( {{\mathbf{U}}_{{\text{o}}}^{{}} \cdot {\mathbf{h}}_{n - 1} + {\mathbf{W}}_{{\text{o}}}^{{}} \cdot {\mathbf{X}}_{n}^{{{\text{exa}}}} + {\mathbf{b}}_{{\text{o}}}^{{}} } \right) $$
(42)

where \({\mathbf{U}}_{{\text{o}}}\) and \({\mathbf{W}}_{{\text{o}}}\) are the weight of \({\mathbf{h}}_{n - 1}\) and \({\mathbf{X}}_{n}^{{{\text{exa}}}}\), respectively; \({\mathbf{b}}_{{\text{o}}}\) is the bias.

(6) The output of the nth sample is calculated as \(\hat{y}^{\prime}_{n}\).

$$ \hat{y}^{\prime}_{n} = \sigma \left( {{\mathbf{W}}_{{{\text{out}}}}^{{}} \cdot {\mathbf{h}}_{n} } \right) = \sigma \left( {\hat{y}^{\prime\prime}_{n} } \right) $$
(43)

where \({\mathbf{W}}_{{{\text{out}}}}\) is the weight corresponding to the implied layer.

The process is the same for the remaining sample input. Output \({\mathbf{\hat{y}^{\prime}}}\) is denoted by

$$ {\mathbf{\hat{y}^{\prime}}} = \left\{ {\hat{y}^{\prime}_{n} ,\hat{y}^{\prime}_{n + 1} , \cdots ,\hat{y}^{\prime}_{n + N - 1} } \right\} $$
(44)

Appendix B

Analysis of model hyperparameters

Univariate sensitivity analysis is performed for hyperparameters such as \(\theta_{{{\text{MI}}}}\), \(\theta_{{{\text{convlayers}}}}\), \(\theta_{{{\text{Epochs}}}}\), \(\theta_{{{\text{learningrate}}}}\), and \(\theta_{{{\text{dropout}}}}\). That is, each time a single hyperparameter is changed, other values maintain the optimal values in the above model. Table 7 shows the intervals in the hyperparameters.

Table 7 Intervals for the hyperparameters of the MI-1DCNN-LSTM model

(1) MI threshold: The higher the \(\theta_{{{\text{MI}}}}\), the fewer the selected feature variables. Figure 

Fig. 11
figure 11

Relationship between \(\theta_{{{\text{MI}}}}\) and R2 for the MI-1DCNN-LSTM model

11 shows its relationship to R2.

(1) RB dataset: The model performance is improved with gradually increased \(\theta_{{{\text{MI}}}}\), and the number of corresponding input features decreases. Stability tends to transpire after reaching a certain value. Consequently, higher \(\theta_{{{\text{MI}}}}\) is selected; (2) CO dataset: R2 increases first and then decreases with increased \(\theta_{{{\text{MI}}}}\). Adaptive \(\theta_{{{\text{MI}}}}\) should be selected. There are differences in MI thresholds for different datasets.

(2) Convolutional layer number: The network will converge slowly or will not converge if \(\theta_{{{\text{convlayers}}}}\) is small. The network will be stronger if \(\theta_{{{\text{convlayers}}}}\) is large. Slow training speed and overfitting of the network model are easy to happen and impair the prediction effect. The number of convolutional layers should be selected according to the data characteristics. The upper limit is ten layers based on experience. Figure 

Fig. 12
figure 12

Relationship between \(\theta_{{{\text{convlayers}}}}\) and R2 for MI-1DCNN-LSTM model

12 shows its relationship to R2.

(1) RB dataset: The model performance increases with increased \(\theta_{{{\text{convlayers}}}}\). However, a downward trend is found after reaching a certain value. (2) CO dataset: R2 first decreases and then increases. Finally, it decreases with increased \(\theta_{{{\text{convlayers}}}}\). Therefore, appropriate \(\theta_{{{\text{convlayers}}}}\) should be chosen. Different datasets should choose the appropriate number of convolutional layers.

(3) Iteration number: Increasing \(\theta_{{{\text{Epochs}}}}\) increases model’s weight updates and prediction accuracy. Figure 

Fig. 13
figure 13

Relationship between \(\theta_{{{\text{Epochs}}}}\) and R.2 for MI-1DCNN-LSTM model

13a, b shows the results of setting iteration number in (50, 3000) and successive changes in units of 50. Figure 13c, d presents the results of setting iteration number in (50, 500) and successive changes in units of 5.

(1) RB dataset: Model performance merely improves with increased \(\theta_{{{\text{Epochs}}}}\) within (50, 3000). R2 increases within (50, 500)with performance degradation. Larger \(\theta_{{{\text{Epochs}}}}\) should be selected to stabilize model performance. (2) CO dataset: Model performance is better with increased \(\theta_{{{\text{Epochs}}}}\) within (50, 3000), and R2 changes little. R2 shows an upward trend within (50, 500), but there are some performance degradation points. Thus, appropriate \(\theta_{{{\text{Epochs}}}}\) should be selected to stabilize model performance. There are differences in the iteration number requirements of different datasets.

(4) Learning rate: Suitable \(\theta_{{{\text{learningrate}}}}\) can converge the objective function to a local minimum at a suitable time. The initial value of the learning rate of most networks should be 0.01 and 0.001. The range of the learning rate is (0.01, 0.1), and the step size is changed successively in 0.01 units. (Fig. 

Fig. 14
figure 14

Relationship between \(\theta_{{{\text{learningrate}}}}\) and R2 for the MI-1DCNN-LSTM model

14a, b). The range is (0.001, 0.03), and the step size is changed successively in units of 0.001 (Fig. 14c, d).

(1) RB dataset: Model performance first increases with increased \(\theta_{{{\text{learningrate}}}}\) and then decreases sharply within (0.01, 0.1). Stable \(\theta_{{{\text{learningrate}}}}\) should be selected. Model performance increases and then decreases with increased \(\theta_{{{\text{learningrate}}}}\) within (0.001, 0.03). Therefore, appropriate \(\theta_{{{\text{learningrate}}}}\) should be selected. (2) CO dataset: Model performance decreases significantly with increased \(\theta_{{{\text{learningrate}}}}\) within (0.01, 0.1), indicating that smaller \(\theta_{{{\text{learningrate}}}}\) is suitable. R2 fluctuates largely at the early and later stages within (0.001, 0.03); thus, smaller \(\theta_{{{\text{learningrate}}}}\) should be selected. Different dataset’s \(\theta_{{{\text{learningrate}}}}\) has different influence ranges on model performance. Consequently, suitable \(\theta_{{{\text{learningrate}}}}\) should be selected.

(5) Dropout rate: It probably stop the activation values of certain neurons. The model is more generalized. \(\theta_{{{\text{dropout}}}}\) is within (0.1,0.9) in the work, and Fig. 

Fig. 15
figure 15

Relationship between \(\theta_{{{\text{dropout}}}}\) and R2 for MI-1DCNN-LSTM model

15 shows its relationship with R2.

(1) RB dataset: Model performance first improves and then decreases with increased \(\theta_{{{\text{dropout}}}}\). It fluctuates greatly after reaching a certain value, indicating that moderate \(\theta_{{{\text{dropout}}}}\) should be selected;

(2) CO dataset: R2 decreases with increased \(\theta_{{{\text{dropout}}}}\); thus, smaller \(\theta_{{{\text{dropout}}}}\) should be selected. Different datasets have different requirements for dropout rates.

In summary, distinct hyperparameters exhibit different performance for varying datasets. It is reasonable to use PSO for the global optimized selection of these hyperparameters.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, R., Tang, J., Xia, H. et al. CO emission predictions in municipal solid waste incineration based on reduced depth features and long short-term memory optimization. Neural Comput & Applic 36, 5473–5498 (2024). https://doi.org/10.1007/s00521-023-09329-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-023-09329-8

Keywords

Navigation