Abstract
We investigated the potential of Artificial Neural Networks (ANN), ANN to forecasts in chaotic series of the price of copper; based on different combinations of structure and possibilities of knowledge in big discovery data. Two neural network models were built to predict the price of copper of the London Metal Exchange (LME) with lots of 100 to 1000 data. We used the Feed Forward Neural Network (FFNN) algorithm and Cascade Forward Neural Network (CFNN) combining training, transfer and performance implemented functions in MatLab. The main findings support the use of the ANN in financial forecasts in series of copper prices. The copper price’s forecast using different batches size of data can be improved by changing the number of neurons, functions of transfer, and functions of performance s. In addition, a negative correlation of −0.79 was found in performance indicators using RMS and IA.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Copper is one of the basic metal products listed on major exchanges in the world: the LME, Commodity Exchange of New York (COMEX) and Shanghai Futures Exchange (SHFE). Prices in these exchanges reflect the balance between the supply and demand of copper worldwide, although they may be strongly influenced by the rates of currency exchange and investment flows, factors that may cause fluctuations of volatile prices partially linked to changes in the economic cycle activity [1].
The price of copper is a sensitive issue for major producers such as Codelco, Freeport-McMoRan Copper & Gold, Glencore Xstrata, BHP Billiton, Southern Copper Corporation, American Smelting and Refining Company. Economies such as those of Chile and Zambia rely heavily on copper production and, subsequently, in the evolution of the prices of the same [2], being Chile the largest producer and exporter of the world.
Several studies include copper as one of the products of interest in the evaluations of the forecast to improve the prediction of prices. They employ a variety of different methods and mathematical models: time series [3,4,5], combined with wavelet [6, 7], transformed of Fourier [8], swarm optimisation algorithm [9], and models of multi-products [10].
A fairly accurate time series model could predict several years forward, whose skill is an advantage for the planning of future requirements. Research on nonlinear dynamical systems has allowed in recent years significantly improve the impact of the predictive capacity of the times series. Chaos is a universal complex dynamic phenomenon that exists in different natural and social systems such as communication, economics and biology. The auto-correlation of chaotic behaviour in the economy began in the decade of 1980, applied to macroeconomic variables such as the gross domestic product (GDP) and monetary aggregates [11]. Since then, several studies have been conducted to search chaotic behaviour in economic and financial series [12, 13].
In this context, amongst the most used techniques and tools are graphics analysis of recurrence, Temporal Space Entropy (STE) Hurst coefficient and exponent of the Lyapunov dimension of correlation for the matching of chaotic behaviour in these series [14]. Additionally, the existence of chaotic behaviour in the commodities of copper was corroborated in [15].
The motivations and the need to carry out this research is to evaluate the ANN by different dimensions, type of networks, functions and structures.
2 Methodology
The continuous increase of the power of calculation and availability of data has grown the attention in the use of ANN in many types of problems of prediction. The ANN can model and predict linear and nonlinear time-series with a high degree of accuracy, to capture any relationship between the data without prior knowledge of the problem which is modelled [16].
The methodology used for visual and statistical analysis is summarised according to the Fig. 1 adapted the Loshin [17].
The closing prices of copper traded on the LME were used with the different functions of learning, optimisation and transfer as input. Then, the quality of the forecast of the ANN performed in MatLab R2014a simulation tool was evaluated with Root Mean Square (RMS) and the Adequacy Index (IA). The saved data goes through the process of Extract, Transform and Load (ETL) to be stored in the Data Warehouse (DW) in a multi-dimensional form, using SQL Server 2008. The visual analysis of the results is done by the software BI Tableau Desktop 8.3 as Frontend, and the R version 3.1.2 software to perform the corresponding statistical analysis.
2.1 Segmentation of Data
Data were segmented into two sizes of batches. The former with 100 and the later with thousand data. Periods from 22/01/2015 to 16/06/2015 and the 01/07/2011 to 16/06/2015 respectively.
Segmentation according to the evaluated batch records is according to the Table 1.
2.2 Artificial Neural Networks
Represented FFNN has been taken to this work in the Fig. 2(a) and the represented CFNN in the Fig. 2(b). We have selected some of the features available in the toolbox of Matlab software.
(a) FFNN of 1–20 neurons in the hidden layer. (b) CFNN of 1–20 neurons in the hidden layer [18].
The inputs used in the system correspond to \( t_{ - 1} \), \( t_{ - 2} \), \( t_{ - 3} \), \( t_{ - 4} \), \( t_{ - 1} \), \( t_{ - 6} \) and target \( t_{0} \) the copper price time series.
2.3 Alternatives of Evaluation
The different alternatives of a network are given according to different combinations between their functions as shown in Eq. (1).
where,
-
\( M \), It is an alternative for a network number.
-
\( E \), It is the number of training functions.
-
\( R \), It is the number of performance functions.
-
\( F \), It is the number of transfer functions.
-
n, It is the number of neurons in the hidden layer.
Replacing values in the Eq. (1) gets the number of alternatives for a network.
To calculate the total number of alternatives evaluated in this work is obtained from the Eq. (2).
where,
-
\( T \), It is the number of evaluated alternatives.
-
\( N \), It is the number of networks (structure of similar functions).
-
\( L \), It is the number of batches of data.
Replacing values in the Eq. (2) obtained 4320 alternative evaluated in this work.
Were 27 simulations for each of the 4320 alternatives evaluated in this work with a total of 116640 simulations according to the Eq. (3).
Where,
-
\( T_{s} \), It is the total number of simulations carried out.
-
\( S \), It is the number of simulations for each alternative.
2.4 Multidimensional Model
This research focuses on the evaluation of two networks simultaneously from different dimensions. The FFNN and the CFNN in dimRed, training features (trainbfg, traincgb, traincgf, traincgp, traingd, traingdm, traingda, traingdx, trainlm) in dimFuncTrain, in dimFuncTransfer (logsig, tansig, purelin) transfer functions, functions (mse, sse) performance in dimFuncPerform. Number them 1 to 20 neurons, with batches of 100 and 1,000 data. The function of learning was the learngdm and the metrics of evaluation [RMS, IA, bestEpoch, duration] in fact table.
Figure 3 shows the “star model” [19] used to perform multidimensional analysis according to the methodology proposed by Kimball [20]. The construction of the DW was based on multidimensional modelling, in which the structures of the information on table are facts and dimensions [21].
2.5 Performance Measures
The results were validated using performance measures that allow indicating the degree of generalisation of the model used. The indexes are the average quadratic Error (RMS), and the index of adequacy (IA). Both are shown in the Eqs. (4) and (5), respectively, where \( o_{i} \) and \( p_{i} \) are the values observed and predicted respectively at the time \( i \), and \( N \) is the total number of data. In addition, \( p_{i}^{'} = p_{i} - o_{m} \) and \( o_{i}^{'} = o_{i} - o_{m} \), being \( o_{m} \) value average of the observations [22].
IA indicates the degree of adjustment that has the values estimated with the actual values of a variable. A value close to 1 indicates a good estimate. On the other hand, near-zero RMS indicate a good quality setting.
3 Results
3.1 Knowledge Discovery
The knowledge is in the data, however, must be extracted and built for the consumption of the users. New technological solutions are required for data management. The new software and methodologies are the solution to make sense of the data, allowing to extract useful information for the construction of knowledge, supporting decision making. Data from markets together with the new data generated from different data mining techniques, can help to the analysis and the management of Big Data [23, 24]. In this case, the business intelligence tool enables the discovery of knowledge.
Figure 4 shows the best results obtained in the average assessment (RMS) ̅ in lots of 100 data FFNN. The indexes are shown in a scale of colors, where the blue colour indicates better value performances corresponding to the minimum of the RMS and the red colour lower performance. After several iterations were obtained the best results to correspond to functions purelin, traincgb, traincgf, traincgp, mse and sse between 1 and 5 neurons.
(\( AI \)) in the FFNN average assessment indicates a better fit in the curves, for this case are kept the same functions and expands the range of neurons from 1 to 20.
In the case of the FFNN network for batches of 1000 data, the best results of \( \overline{RMS} \) and \( \overline{AI} \) they correspond to the functions purelin, trainlm, mse and sse with ranges of neurons from 6 to 10 in Table 2.
In the case of CFNN for batches of 100 data, the best results for the averages of \( \overline{RMS} \) and \( \overline{AI} \) they correspond to the functions purelin, traincgb, traincgf, traincgp, mse and sse between 1 and 2 neurons in Table 3.
CFNN batches of 1000 data, for best results unemployment averages \( \overline{RMS} \) and \( \overline{AI} \) they correspond to the functions purelin, trainlm, mse and sse with ranges from 3 to 4 neurons in Table 4.
We found that FFNN and CFNN for batches of 100 data obtained the best results with the training functions named traincgf, traincgb, and traincgp. On the other hand, for batches of 1000 data, the best function of training was the trainlm in FFNN and CFNN.
3.2 Forecast
The charts show the forecast made by the FFNN and chaotic CFNN in series of the copper’s price (Fig. 5a, b, c and d). FFNN and CFNN show similar values in his performance’s ratings.
(a) Forecast of the FFNN for data verification with lot of 100, purelin, traincgp and sse of 5 neurons functions; (b) Forecast of the FFNN for data verification with lot of 1000, purelin, trainlm and sse of 7 neurons functions; (c) Forecast of CFNN for data verification with lot of 100, purelin, traincgb and sse of 1 neurons functions; (d) Forecast of CFNN for data verification with lot of 1000, purelin, trainlm and mse of 3 neurons functions.
3.3 Statistic Analysis
The CFNN (trainFcn = trainlm, performFcn = sse, transferFcn = purelin, learnFcn = learngdm, lot = 1000, numNeuronas = 3) simulo up to 10000 simulations, where obtained good results in the performance of the network, with average measures of RMS of 0.00767957, IA of 0.9725127 and best_epoch of 3.1376; minimum values of 0.007535554; 0.9698627 and 1 respectively, maximum values of 0.007712344; 0.9731063 and 11 respectively, in the Table 5 the statistical summary of the simulation.
To review the distribution of RMS performance measures is observed which is skewed to the left Fig. 6a, Therefore, the bias is negative (less than the median average); Instead the IA is skewed to the right index Fig. 6b, therefore, the bias is positive (greater than the median average) and the best_epoch this biased right in Fig. 6c. On the other hand the correlations (Pearson) graph shows a high negative correlation of −0.79 between RMS and the IA which shows the good performance of the ANN indicated by RMS will also be it according to the IA. On the other hand presents a negative correlation of −0.49 between IA and best_epoch; and a positive correlation of 0.52 between RMS and best_epoch.
4 Conclusions
Forecasts based on neural network, nonlinear models achieved better results compared with linear forecasting models. Combinations of two neural network models, with nine training functions, two functions of performance, lots of 100 to 1000 data and ranges from 1 to 20 neurons and three transfer functions have been studied.
For different batch size of data it is to be more sensitive to the functions of training on the functions of transfer, performance, number of neurons.
The results of the simulations predict in the case of the copper price series; the best combinations of functions are vector (trainFcn = trainlm, performFcn = sse, transferFcn = purelin, learnFcn = learngdm, lot = 1000, number neurons = 3) for batch 1000 network CFNN and vector (trainFcn = traincgp, performFcn = mse, transferFcn = purelin, learnFcn = learngdm, lot = 100, numNeuronas = 1) to the FFNN.
Found a high negative correlation of −0.79 in the performance indicators used RMS e IA.
In future work, it is necessary to review the contribution of other macroeconomic variables in the performance of the model, in particular, market of capitals of other commodities.
Finally, the characteristics of these time series, should be analysed the same set of combinations of functions in other economic series of price, to assess the possible relationships of the systems involved.
References
Roberts, M.C.: Duration and characteristics of metal price cycles. Resour. Policy 34, 87–102 (2009)
Sánchez Lasheras, F., de Cos Juez, F.J., Suárez Sánchez, A., Krzemień, A., Riesgo Fernández, P.: Forecasting the COMEX copper spot price by means of neural networks and ARIMA models. Resour. Policy 45, 37–43 (2015)
Dooley, G., Lenihan, H.: An assessment of time series methods in metal price forecasting. Resour. Policy 30, 208–217 (2005)
Carrasco, R., Astudillo, G., Soto, I., Chacon, M., Fuentealba, D.: Forecast of copper price series using vector support machines. In: 2018 7th International Conference on Industrial Technology and Management (ICITM), pp. 380–384. IEEE (2018)
Seguel, F., Carrasco, R., Adasme, P., Alfaro, M., Soto, I.: A meta-heuristic approach for copper price forecasting. In: Liu, K., Nakata, K., Li, W., Galarreta, D. (eds.) ICISO 2015. IAICT, vol. 449, pp. 156–165. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-16274-4_16
Kriechbaumer, T., Angus, A., Parsons, D., Rivas Casado, M.: An improved wavelet–ARIMA approach for forecasting metal prices. Resour. Policy 39, 32–41 (2014)
Goss, B.A., Avsar, S.G.: Simultaneity, forecasting and profits in London copper futures. Aust. Econ. Pap. 52, 79–96 (2013)
Khalifa, A., Miao, H., Ramchander, S.: Return distributions and volatility forecasting in metal futures markets: evidence from gold, silver, and copper. J. Futures Marks 31, 55–80 (2011)
Ma, W., Zhu, X., Wang, M.: Forecasting iron ore import and consumption of China using grey model optimized by particle swarm optimization algorithm. Resour. Policy 38, 613–620 (2013)
Cortazar, G., Eterovic, F.: Can oil prices help estimate commodity futures prices? The cases of copper and silver. Resour. Policy 35, 283–291 (2010)
Lebaron, B.: Chaos and nonlinear forecastability in economics and finance. Philos. Trans. Phys. Sci. Eng. 348, 397–404 (1994)
Los, C.A.: Visualization of chaos for finance majors. EconWPA (2004)
Los, C.A., Yu, B.: Persistence characteristics of the Chinese stock markets. Int. Rev. Financ. Anal. 17, 64–82 (2008)
Espinosa, C.: Caos en el mercado de commodities. Cuad. Econ. 29, 155–177 (2010)
Carrasco, R., Vargas, M., Alfaro, M., Soto, I., Fuertes, G.: Copper metal price using chaotic time series forecating. IEEE Lat. Am. Trans. 13, 1961–1965 (2015)
Kourentzes, N., Barrow, D.K., Crone, S.F.: Neural network ensemble operators for time series forecasting. Expert Syst. Appl. 41, 4235–4244 (2014)
Loshin, D.: The business intelligence environment, Chap. 5. In: Loshin, D. (ed.) business intelligence, pp. 61–76. Morgan Kaufmann, Boston (2013)
MathWorks: MatLab 2014a (2014)
Howson, C.: Successful Business Intelligence: Secrets to Making BI a Killer App [Hardcover]. McGraw-Hill Osborne Media, New York (2007)
Sherman, R.: Data architecture, Chap. 6. In: Business Intelligence Guidebook, pp. 107–142. Morgan Kaufmann, Boston (2015)
Soler, E., Trujillo, J., Fernández-Medina, E., Piattini, M.: Building a secure star schema in data warehouses by an extension of the relational package from CWM. Comput. Stand. Interfaces 30, 341–350 (2008)
Zambrano Matamala, C., Rojas Díaz, D., Carvajal Cuello, K., Acuña Leiva, G.: Análisis de rendimiento académico estudiantil usando data warehouse y redes neuronales. Ingeniare. Rev. Chil. Ing. 19, 369–381 (2011)
Xhafa, F., Taniar, D.: Big data knowledge discovery. Knowl.-Based Syst. 79, 1–2 (2015)
Lagos, C., Carrasco, R., Fuertes, G., Gutiérrez, S., Soto, I., Vargas, M.: Big data on decision making in energetic management of copper mining. Int. J. Comput. Commun. Control 12, 61–75 (2017)
Acknowledgment
The authors are acknowledgment the financing of the project “Multiuser VLC for underground mining”, code: IT17M10012 and this research has been supported by DICYT (Scientific and Technological Research Bureau) of The University of Santiago of Chile (USACH) and Department of Industrial Engineering.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 IFIP International Federation for Information Processing
About this paper
Cite this paper
Carrasco, R., Vargas, M., Soto, I., Fuentealba, D., Banguera, L., Fuertes, G. (2018). Chaotic Time Series for Copper’s Price Forecast. In: Liu, K., Nakata, K., Li, W., Baranauskas, C. (eds) Digitalisation, Innovation, and Transformation. ICISO 2018. IFIP Advances in Information and Communication Technology, vol 527. Springer, Cham. https://doi.org/10.1007/978-3-319-94541-5_28
Download citation
DOI: https://doi.org/10.1007/978-3-319-94541-5_28
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-94540-8
Online ISBN: 978-3-319-94541-5
eBook Packages: Computer ScienceComputer Science (R0)