Elsevier

Information Sciences

Volumes 367–368, 1 November 2016, Pages 1078-1093
Information Sciences

Random vector functional link network for short-term electricity load demand forecasting

https://doi.org/10.1016/j.ins.2015.11.039Get rights and content

Abstract

Short-term electricity load forecasting plays an important role in the energy market as accurate forecasting is beneficial for power dispatching, unit commitment, fuel allocation and so on. This paper reviews a few single hidden layer network configurations with random weights (RWSLFN). The RWSLFN was extended to eight variants based on the presence or absence of input layer bias, hidden layer bias and direct input–output connections. In order to avoid mapping the weighted inputs into the saturation region of the enhancement nodes’ activation function and to suppress the outliers in the input data, a quantile scaling algorithm to re-distribute the randomly weighted inputs is proposed. The eight variations of RWSLFN are assessed using six generic time series datasets and 12 load demand time series datasets. The result shows that the RWSLFNs with direct input–output connections (known as the random vector functional link network or RVFL network) have statistically significantly better performance than the RWSLFN configurations without direct input–output connections, possibly due to the fact that the direct input–output connections in the RVFL network emulate the time delayed finite impulse response (FIR) filter. However the RVFL network has simpler training and higher accuracy than the FIR based two stage neural network. The RVFL network is also compared with some reported forecasting methods. The RVFL network overall outperforms the non-ensemble methods, namely the persistence method, seasonal autoregressive integrated moving average (sARIMA), artificial neural network (ANN). In addition, the testing time of the RVFL network is the shortest while the training time is comparable to the other reported methods. Finally, possible future research directions are pointed out.

Introduction

Short-term electricity load demand forecasting targets to predict the future load demand ranging from a few minutes to 1 day ahead [2], [19]. In a contemporary competitive energy market, as power systems become deregulated and involve distributed power sources, electricity demand becomes more complicated in both temporally and spatially. Since electricity load demand is closely related to the electricity price, load forecasting plays an important role in the energy market [19]. From the energy generation point of view, load forecasting is beneficial for electric utility operations such as power dispatching, unit commitment, fuel allocation and network diagnosis [2]. The significance of forecasting error with respect to power system operational cost was stated in [2], [19]: 1% increase in error can result in wastage of millions of dollars. Therefore, improving the forecasting accuracy can significantly reduce the power system operational cost.

Fig. 3a in Section 5.1 shows a small fraction of load demand time series (TS). From the plot, we can observe the non-linearity of the load demand TS because it is usually affected by a lot of exogenous factors such as weather, special occasions, economy, and so on [9], thereby making the modeling of load demand TS stochastic. However, it also reveals strong seasonal components such as daily, weekly, monthly or yearly repetitions. These properties, on the other hand, makes the modeling somewhat deterministic [13].

In the literature, there are numerous methods for load forecasting and they can be categorized into (i) statistical methods, (ii) machine learning methods and (iii) hybrid methods. Statistical methods use mathematical equations and statistical theories to model and forecast the TS: it is fast but with limitations (distributions, stationarity, linearity, etc.). The commonly used statistical methods are exponential smoothing [32] and autoregressive moving average (ARMA) and so on [33]. Machine learning methods use supervised learning to model a portion of historical data (training data). Once the model is optimized based on the training data, it can be applied to forecast the unknown future data (testing data). Some well-known machine learning methods are artificial neural networks (ANN) [19], [23], support vector regression (SVR) [12], [18], [22], etc. Combining more than one forecasting method to form a single forecasting method is known as a hybrid method. The combination can be sequential or parallel. For a sequentially combined hybrid method, one forecasting method’s output is another forecasting method’s input, and the final output is the output of the last forecasting method. For a parallel combined hybrid method, the training data is either bootstrapped to multiple training data sets or decomposed to a collection of training data sets first. Subsequently, each data set is trained by an individual method. Finally, the trained forecasting methods are used to predict the future data and the outputs are aggregated to form the final prediction. Negative correlation learning [3], [9] and wavelet [17], [20], [25], [35], [36] and Empirical Mode Decomposition (EMD) [14], [16], [30] based methods are in this category.

ANN is a popular method among the machine learning based load forecasting methods. There are several ANN based forecasting methods reported in the literature [15], [28], [34]. The usual way to train the ANN is through back propagation (BP). But in [31], a single hidden layer neural network with random weights (RWSLFN) was reported with random weight assignment between input and hidden layers and least square estimation on the output weights as the training method. In addition, we will discuss another type of ANN which is called random vector functional link network (RVFL network) [5], [6], [27]. Although it is similar to RWSLFN reported in [31], it has direct input–output connections between input and output neurons. This paper will compare these neural networks and their forecasting capability in short-term load demand forecasting.

The remaining of the paper is organized as follows: Section 2 reviews different variants of RWSLFN/RVFL network; Section 3 states the essentials of the experiment configurations, feature selection methods, parameter optimization procedure and error measures. Section 4 assesses the performance of different variants on six generic TS data; Section 5 discusses the performance of RWSLFN/RVFL network with eight variants on 12 load demand TS, and Section 6 concludes the paper and suggests possible future research directions.

Section snippets

Review of related neural network structures

ANN is one of the most popular algorithms in machine learning. ANN can be categorized according to the connectionism: feedforward and recurrent. The main difference between feedforward neural network (FNN) and recurrent neural network (RNN) is that the FNN does not have connections from the hidden layers or output layer back to the hidden layer. RNN not only has the connections that the FNN has but also has the connections from hidden layers or output layer back to the input layer. The

Variations on RWSLFN/RVFL network

By varying the network components such as input layer bias, hidden layer bias and input– output connections, we can obtain eight different RWSLFN/RVFL network configurations. The eight RWSLFN/RVFL network configurations are shown in Table 1.

Time series cross validation

The time series cross validation (TS-CV) is used to determine the optimal parameters of the methods. The TS-CV process is slightly different from CV for classification and regression because the TS data are sequential and auto-correlated. The conventional k

Assessment on generic time series data

In order to assess the performances of the eight configurations of RWSLFN/RVFL network to obtain generalized conclusions, we chose six generic TS datasets  [26]. The statistics of the six datasets are summarized in Table 2.

There were ten assessments implemented for each dataset and these assessments were on 1–10 steps ahead forecasting. The performances were measured by nRMSE. For each TS, the first 70% was used for training and the remaining 30% was used for testing. The feature set is ‘full’

Results and discussions on load demand forecasting

In this section, eight different variations of RWSLFN/RVFL network were applied to model and forecast the electricity load demand. Then the best of the eight configurations was compared with several reported statistical and machine learning methods for load demand forecasting.

Conclusion

Random weight single hidden layer neural networks (RWSLFN) have been employed for short term time series forecasting. Eight configurations of RWSLFN have been formulated and compared on six generic time series datasets as well as twelve load demand datasets. From the comparison results, the following conclusions are made:

  • The quantile scaling algorithm has improved the performance for 1–4 h and 18–24 h ahead forecasting horizons.

  • Feature selection based on partial auto correlation function or

Acknowledgments

This work was supported by the Singapore National Research Foundation (NRF) under its Campus for Research Excellence And Technological Enterprise (CREATE) program, and Cambridge Advanced Research Centre in Energy Efficiency in Singapore (CARES), C4T project.

Authors thank the managing Guest Editor Associate Professor Dianhui Wang for suggesting us to investigate the scaling of randomization.

The author Ye Ren would also like to thank the Clean Energy Program Office (CEPO) for scholarship.

References (37)

  • L. Breiman

    Random forests

    Mach. Learn.

    (2001)
  • C.L.P. Chen

    A rapid supervised learning neural network for function interpolation and approximation

    IEEE Trans. Neural Netw.

    (1996)
  • C.L.P. Chen et al.

    A rapid learning and dynamic stepwise updating algorithm for flat neural networks and the application to time-series prediction

    IEEE Trans. Syst. Man Cybern. Part B: Cybern.

    (1999)
  • Y.-Y. Cheng et al.

    Random forest based ensemble system for short term load forecasting

    Proceedings of the International Conference on Machine Learning and Cybernetics (ICMLC2012)

    (2012)
  • H. Cruse

    Neural Networks as Cybernetic Systems

    (1996)
  • M. De Felice et al.

    Short-term load forecasting with neural network ensembles: a comparative study [application notes]

    IEEE Comput. Intell. Mag.

    (2011)
  • H. Drucker et al.

    Support vector regression machines

    Adv. Neural Inf. Process. Syst.

    (1997)
  • G. Dudek

    Short-term load forecasting using random forests

  • Cited by (0)

    View full text