Skip to main content

Advertisement

Log in

Neural networks and genetic algorithms can support human supervisory control to reduce fossil fuel power plant emissions

  • Original Article
  • Published:
Cognition, Technology & Work Aims and scope Submit manuscript

Abstract

Artificial neural networks and genetic algorithms are two intelligent approaches initially targeted to model human information processing and natural evolutionary process, with the aim of using the models in problem solving. During the last decade these two intelligent approaches have been widely applied to a variety of social, economic and engineering systems. In this paper, they have been shown as modelling tools to support human supervisory control to reduce fossil fuel power plant emissions, particularly NOx emissions. Human supervisory control of fossil fuel power generation plants has been studied, and the need of an advisory system for operator support is emphasized. Plant modelling is an important block in such an advisory system and is the key issue of this study. In particular, three artificial neural network models and a genetic algorithm-based grey-box model have been built to model and predict the NOx emissions in a coal-fired power plant. In non-linear dynamic system modelling, training data is always limited and cannot cover all system dynamics; therefore the generalization performance of the resultant model over unseen data is the focus of this study. These models will then be used in the advisory system to support human operators on aspects such as task analysis, condition monitoring and operation optimization, with the aim of improving thermal efficiency, reducing pollutant emissions and ensuring that the power system runs safely.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1.
Fig. 2.
Fig. 3.
Fig. 4.
Fig. 5.
Fig. 6.
Fig. 7.
Fig. 8.
Fig. 9.
Fig. 10.
Fig. 11.
Fig. 12.
Fig. 13.
Fig. 14.
Fig. 15.
Fig. 16.
Fig. 17.
Fig. 18.
Fig. 19.

Similar content being viewed by others

References

  • Blanco A, Delgado M, Pegalajar MC (2001) A real-coded genetic algorithm for training recurrent neural networks. Neural Networks 14(1):93–105

    Article  CAS  PubMed  Google Scholar 

  • Bohlin T (1991) Interactive system identification: prospects and pitfalls. Springer, Berlin Heidelberg New York

    Google Scholar 

  • Bye A, Hollnagel E, Brendeford TS (1999) Human–machine function allocation: a functional modelling approach. Reliability Eng Syst Safety 64:291–300

    Article  Google Scholar 

  • Cacciabue PC (1997) A methodology of human factor analysis for systems engineering: theory and applications. IEEE Trans Syst Man Cybernetics 27(3):325–339

    Article  Google Scholar 

  • Cacciabue PC (1998) Modelling and simulation of human behaviour in system control. Springer, Berlin Heidelberg, New York

  • Carpignano A, Piccini M (1999) Cognitive theories and engineering approaches for safety assessment and design of automated systems: a case study of a power plant. Cognition Technol Work 1(1):47–61

    Article  Google Scholar 

  • Coal R&D programme (1997) Technology status report: NOx control for pulverised coal-fired power plant. ETSU, Harwell, UK

    Google Scholar 

  • Copado A et al (2001) Boiler efficiency and NOx optimisation through advanced monitoring and control of local combustion conditions. In: Sixth international conference on technologies and combustion for a clean environment, vol 1, Porto, Portugal, 9–12 July, 2001, pp 903–908

  • De Soete GG (1975) Overall reaction rates of NO and N2 formation from fuel nitrogen. In: 15th symposium (international) on combustion. Combustion Institute, pp 1093–1102

  • Goldberg DE (1989) Genetic algorithms in search, optimization, and machine learning. Addison-Wesley, Reading, MA

  • Hagen MT, Menhaj MB (1994) Training feedforward networks with the marquardt algorithm. IEEE Trans Neural Networks 5(6):989–993

    Article  Google Scholar 

  • Hertz J, Krogh A (1991) Introduction to the theory of neural computation. Addison-Wesley, Reading, MA

  • Hollnagel E, Cacciabue PC (1999) Cognition, technology and work: an introduction. Cognition Technol Work 1(1):1–6

    Article  Google Scholar 

  • Holmes K, Mayes LW (1994) Progress report on the development of a generic NOx control intelligent system (GNOCIS). Coal R & D Program, Project Profile 103, ETSU

  • Irwin GW, Warwick K, Hunt KJ (1995) Neural network applications in control. Institution of Electrical Engineers. Short Run Press, Exeter

  • Johannsen G, Levis AH, Stassen HG (1994) Theoretical problems in man–machine systems and their experimental validation. Automatica 30(2):217–231

    Article  Google Scholar 

  • Kent S, Stewart R (2000) Evolutionary algorithms: a tool for addressing problems which humans cannot solve. Cognition Technol Work 2(1):35–49

    Article  Google Scholar 

  • Li K, Thompson S (2000) Developing NOx emission model for a coal-fired power generation plant using artificial neural networks. In: UKACC international conference on control 2000, Cambridge, UK, 4–7 September 2000

  • Li K, Thompson S (2001a) NOx emission models for operation and control of power generation boilers. In: 6th international conference on technologies and combustion for a clean environment, vol 2, Porto, Portugal, 2001, pp 889–895

  • Li K, Thompson S (2001b) Fundamental grey-box modelling. In: Proceedings of the European control conference 2001, 3–7 September 2001, Oporto, Portugal, pp 3648–3653

  • Li K, Wieringa PA (2000) Understanding perceived complexity in human supervisory control. Cognition Technol Work 2(2): 75–88

    Article  Google Scholar 

  • Li K, Thompson S, Duan GR, Peng J (2002) A case study of fundamental grey-box modeling. In: 15th IFAC world congress on automatic control, Barcelona, July 2002

  • Liu GP (2001) Nonlinear identification and control: a neural network approach. Springer, Berlin Heidelberg New York

    Google Scholar 

  • Liu GP, Daley S (2001) Adaptive predictive control of combustor NOx emissions. Control Eng Practice 9(6):631–638

    Article  Google Scholar 

  • Peng J, Li K, Thompson S (2001) GA based software for power generation plant NOx emission modelling. In: 6th international conference on technologies and combustion for a clean environment, vol 2, Porto, Portugal, 2001, pp 881–887

  • Reinschmidt KF, Ling B (1994) Neural networks for NOx control. In: Proceedings of the American power conference, Vol 56, part 1, pp 354–359

  • Sheridan TB (1992) Telerobotics, automation, and human supervisory control. MIT Press, Cambridge, MA

  • Smink OG, Wieringa PA, van de Wiel JCJ (2001) On genetic programming of fuzzy knowledge-bases. Honeywell Industrial Automation & Control, June 2001

  • Stassen HG, Johanssen G, Moray N (1990) Internal representation, internal model, human performance model and mental load. Automatica 26:811–820

    Article  Google Scholar 

  • Thompson S, Li K (2002) Fundamental grey-box modelling for operation and control of coal-fired power generation plants. QUB Report for BCURA, 2002

  • Visona SP, Stanmore BR (1996) 3-D modelling of NOx formation in a 275 MW utility boiler. J Inst Energy 69:68–79

    CAS  Google Scholar 

  • Wieringa PA, Stassen HG (1993) Assessment of complexity. In: Wise JA et al (eds) Verification and validation of complex systems: human factors issues. Springer, Berlin Heidelberg New York, pp 173–180

    Google Scholar 

  • Woods DD (1988) Coping with complexity: the psychology of human behavior in complex systems. In: Goodstein LP, Anderson HB, Olsen SE (eds) Tasks, errors and mental models. Taylor & Francis, London, pp 128–147

Download references

Acknowledgements

Acknowledgement is made to the British Coal Utilization Research Association and the UK Department of Trade and Industry for a grant-in-aid for this research. The views expressed are those of the authors, and not necessarily those of BCURA or the Department of Trade and Industry.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to K. Li.

Appendix: An ANN configuration selection algorithm

Appendix: An ANN configuration selection algorithm

Suppose we have a network model with p input nodes and q output nodes. A data set Z1 is used for neural network training:

$$ \left\{ {\matrix{ {Z_{\bf 1}^{} = \left\{ {z_{1i} ,{\rm i} = {\rm 1}{\rm ,2},...,{\rm N}_1 } \right\}} \cr {z_{1i} = \{ [u_{11} (i){\rm u}_{12} (i){\rm }...{\rm }u_{1p} (i)]^T ;[t_{11} (i){\rm t}_{12} (i){\rm }...{\rm t}_{1q} (i)]^T \} } \cr } } \right. $$
(22)

where \( z_{1i} ,{\rm i} = {\rm 1}{\rm ,2},...,{\rm N}_{\rm 1} \) are data samples; \( u_{1i} (j),{\rm i} = {\rm 1}{\rm ,2},...,{\rm p}{\rm , j} = {\rm 1}{\rm ,2},...,{\rm N}_{\rm 1} \) are input values; \( t_{1i} (j),{\rm i} = {\rm 1}{\rm ,2},...,{\rm q}{\rm , j} = {\rm 1}{\rm ,2},...,{\rm N}_{\rm 1} \) are the output targets.

Then the cost function is defined as

$$ E_1 ({\bf Z}_1 ;\omega ) = \sum\limits_{j = 1}^{N_1 } {\sum\limits_{i = 1}^q {(y_{1i}^{} (j) - t_{1i}^{} (j))^2 } } = \left\| {\varepsilon _1 } \right\|^2 $$
(23)

where \( y_{1i} (j),{\rm i} = {\rm 1}{\rm ,2},...,{\rm q}{\rm , j} = {\rm 1}{\rm ,2},...,{\rm N}_{\rm 1} \) are the outputs of the network model given inputs from the training data set, q is the number of output nodes, ω is the adjustable weight vector, e1 are error vectors.

A recursive training algorithm to update the weights with respect to the cost function defined in (23) may take the following form:

$$ \left\{ \matrix{ \omega ^{(i + 1)} = \omega ^{(i)} - \mu ^{(i)} {\bf{H}}^{(i)} {\rm{E'}}_{\rm{1}} (\omega ^{(i)} ),{\rm{ j}} = 0,{\rm{1}}{\rm{,2}} \hfill \cr \omega ^{(0)} = \omega _0 \hfill \cr} \right. $$
(24)

where µ is the step size, which is determined by some search along the indicated line. H is some positive definite matrix, \( E'_1 \) is the first derivative of the cost function of (23) with respect to the weights. The initial value for ω could be some prior guess, i.e. ω0 is generated randomly.

Most recursive learning algorithms are based on Newton-type gradient-descent type techniques. This is a simple gradient descent method but it suffers from the problems of slow convergence and is subject to frequent failure. Consequently, many researchers have developed heuristical extensions such as random search methods. Most recent advances in training have used powerful second-order optimization techniques, and typically involve the calculation of at least an approximate Hessian matrix associated with the function to be optimized. One of these powerful and successful algorithms is the LM (Levenberg–Marquart) method, which is able to produce quick convergence to the optimal solution. The LM method will be used in network training for all three types of neural network models, which will be discussed in the following.

For the LM approach (Hagen and Menhaj 1994), (24) will take the form of

$$ \omega ^{(i + 1)} = \omega ^{(i)} - \mu ^{(i)} ({\rm P}_j^T {\bf P}_j^T + \lambda {\bf I})^{ - 1} {\bf P}_j^T \varepsilon _1 (\omega ^{(i)} ) $$
(25)

where \( {\bf P}_j = ({{\partial \varepsilon _j } \over {\partial \omega }}) \), I is an identity matrix, εl is defined in (23) and λ is a small positive real number.

Now consider that we have two ANN architectures denoted as \( ANN( \bullet ;\omega _1 ) \) and \( ANN( \bullet ;\omega _2 ) \), where \( \omega _1 \in {\bf R}^{n_1 } \) and \( \omega _2 \in {\bf R}^{n_2 } \) are adjustable vectors with different dimensions corresponding to different architecture selection of ANN. The performances of these two trained ANNs are denoted as \( E(\omega _1^ * ) \) and \( E(\omega _2^ * ) \) respectively. Then we can define a likelihood-ratio test statistic L:

$$ L = {{\left( {{{E(\omega _2^ * ) - E(\omega _1^ * )} \over {n_1 - n_2 }}} \right)} \mathord{\left/ {\vphantom {{\left( {{{E(\omega _2^ * ) - E(\omega _1^ * )} \over {n_1 - n_2 }}} \right)} {\left( {{{E(\omega _1^ * )} \over {s - n_1 }}} \right)}}} \right. \kern-\nulldelimiterspace} {\left( {{{E(\omega _1^ * )} \over {s - n_1 }}} \right)}} $$
(26)

where s is the number of total samples, and n 1 and n 2 are dimensions of adjustable vectors and ω2 respectively. Although likelihood-ratio test statistic L may not be F-distributed, since neural networks may not have a Gaussian error distribution, it is assumed here that likelihood-ratio test statistic L is approximately F-distributed (with DOF=(n 1 –n 2 )/(s–n 1)). Thus, under the F-distribution assumption, if L exceeds α×100 critical point of the F-distribution, ANN architecture selection \( ANN( \bullet ;\omega _2 ) \) will be rejected. Now likelihood-ratio test statistic L defined in (26) could be used to compare and select between two network architecture selections. The ANN structure selection algorithm is briefed as follows and illustrated in Fig. 20.

Fig. 20.
figure 20

Flow chart for ANN structure selection algorithm

Algorithm: architecture selection based on likelihood test

  1. 1.

    Determine the maximal numbers of hidden nodes that you wish to apply to the ANN architecture.

  2. 2.

    Select an ANN architecture with the maximal number of nodes which is large enough, and denoted as \( ANN( \bullet ;\omega _1 ) \), and let the number of nodes be denoted as N 1.

  3. 3.

    Find the optimal adjustable vector \( \omega _1^ * \) of \( ANN( \bullet ;\omega _1 ) \) obtained by the training algorithm (25) or, in this paper, the particular LM method.

  4. 4.

    Calculate \( E(\omega _1^ * ) \) for \( ANN( \bullet ;\omega _1 ) \).

  5. 5.

    Select another ANN architecture \( ANN( \bullet ;\omega _2 ) \) with fewer number of nodes than \( ANN( \bullet ;\omega _1 ) \), let the number of nodes of \( ANN( \bullet ;\omega _2 ) \) denoted as N 2. If the number of nodes N 2 is less than the minimal number, stop and go to the end.

  6. 6.

    Find the optimal adjustable vector \( \omega _2^ * \) of \( ANN( \bullet ;\omega _2 ) \).

  7. 7.

    Calculate \( E(\omega _2^ * ) \) for \( ANN( \bullet ;\omega _2 ) \).

  8. 8.

    Test the null hypothesis that the reduced model \( ANN( \bullet ;\omega _2 ) \) is equivalent to the \( ANN( \bullet ;\omega _1 ) \) by calculating the likelihood-ratio test statistic L defined in (26).

  9. 9.

    If \( L \le F_\alpha \), accept the reduced model \( ANN( \bullet ;\omega _2 ) \) and replace \( ANN( \bullet ;\omega _1 ) \) with \( ANN( \bullet ;\omega _2 ) \), select a new \( ANN( \bullet ;\omega _2 ) \) whose number of nodes are less than the present N 2, replace the number N 2 with that new number and go to step 5.

  10. 10.

    If \( L > F_\alpha \), keep \( ANN( \bullet ;\omega _1 ) \), and stop. The final network model is described as \( ANN( \bullet ;\omega _1^ * ) \) with the optimal adjustable vector \( \omega _1^ * \) and \( ANN( \bullet ;\omega _1^ * ) \) could now be employed for use.

Remark

To determine the maximal number of hidden nodes is an important issue in the proposed algorithm, since it is the starting point to prune the network nodes. The decision is, however, mostly based on experience or trial and error. Experience shows that for most engineering systems the maximal number can be chosen to be between 20 and 40.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Li, K., Thompson, S., Wieringa, P.A. et al. Neural networks and genetic algorithms can support human supervisory control to reduce fossil fuel power plant emissions. Cogn Tech Work 5, 107–126 (2003). https://doi.org/10.1007/s10111-002-0107-6

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10111-002-0107-6

Keywords

Navigation