1. Introduction
The global energy consumption scenario is dominated by non-renewable sources such as coal, oil and natural gas. In 2022, according to the Energy Information Administration (EIA) [1], the consumption was: oil (29.5%), coal (26.8%), natural gas (23.7%), biomass (9.8%), nuclear energy (5.0%), hydroelectric energy (2.7%) and other sources (2.5%). In the coming years, oil and natural gas are expected to remain prominent, driven by the development of nations such as China, the largest importer and second largest consumer of oil [2].
Oil, a raw material with high industrial value, has its price influenced by global economic and geopolitical aspects [3,4,5,6,7,8,9]. This price is determined by a complex, non-linear system with many uncertainties [10].
Since 2008, the fall in oil prices has been influenced by the global economic slowdown and geopolitical instability, as well as the crisis between China and the US. The COVID-19 pandemic and the war between Russia and Ukraine have added new uncertainties, affecting price formation [6,11,12]. These events have caused fluctuations in prices, challenging market and political decisions, but also offering opportunities to explore forecasting methods.
Forecasting models include linear and non-linear approaches and combination strategies such as hybrids and ensemble. Linear models, such as exponential smoothing, are used to capture patterns in time series by adjusting for trends and seasonality. For example, Simple Exponential Smoothing (SES) is suitable for series with no trend or seasonality, while the Holt-Winters model deals with series that have these characteristics. Box & Jenkins models, such as AR, ARMA and ARIMA, are essential for analyzing time dependencies, where AR captures the linear relationship between an observation and several past lags, MA models the forecast error as a linear combination of past errors and ARIMA handles non-stationary series by incorporating differentiation [13].
Variants of the ARMA model optimized by Genetic Algorithm (GA) and Particle Swarm Optimization (PSO) improve forecast accuracy by automatically adjusting parameters, allowing for more effective modeling of complex dynamics [14]. These optimization techniques provide an enhanced ability to capture subtle patterns and deal with the inherent complexity of time series.
In addition to hybrid models, combination strategies such as ensemble combine outputs from individual predictors [15]. These strategies include averages, medians, weighted averages and other combinations [16,17]. ensemble techniques optimize the accuracy of forecasts by combining results from multiple models, reducing the variance of errors and increasing the consistency of estimates in volatile markets.
The literature has evolved regarding forecasting models for monthly crude oil (WTI) futures prices [18,19,20]. Although new techniques are emerging, linear models are still widely used, from simple comparisons to hybrid models [21].ensemble models have the potential to improve forecast accuracy, but are still little explored [15,16,17].
The aim of this article is to explore linear models, specifically smoothing and Box & Jenkins models, and apply incremental adjustments to the ARMA model using GA and PSO. Combination strategies ensemble considered include mean, median, pseudo-inverse of Moore-Penrose and dynamic adjustments of weights with GA and PSO, significantly improving the performance of the results.
2. Linear Models
2.1. Smoothing Models
This section presents the first set of models used, known as smoothing models. In these models, the main objective is to estimate the smoothing parameters. The models will be divided as follows: Simple Exponential Smoothing SES, Holt Exponential Smoothing (HES), Additive Holt-Winters (A-HW) and Multiplicative Holt-Winters (M-HW).
2.1.1. Simple Exponential Smoothing (SES)
Simple Exponential Smoothing SES is a data smoothing model that applies non-corresponding weights to the fundamental values of the time series [22,23]. The forecast for one period ahead is given by Equation (1):
(1)
where is the forecast, is the actual data in period t, is the forecast in period t and is the smoothing parameter ().Values of close to zero indicate slower forecasts and less reaction to changes, while values close to one result in faster responses to recent changes in the time series.
After defining the SES model equation, the next section will introduce the Holt model.
2.1.2. Holt Exponential Smoothing (HES)
The Holt Exponential Smoothing HES models are widely used in time series with a linear trend [24]. Unlike the SES model, which smooths only the level, the Holt model also models the trend. Represented by Equations (2)–(4).
(2)
(3)
(4)
where is the new smoothed value, is the smoothing coefficient (), is the current value in period t, is the trend smoothing coefficient (), is the predicted trend, and is the predicted value.Low values of indicate a slow adjustment to the trend, while high values indicate a rapid response to changes in the trend.
The next section will introduce the Holt-Winters model, which models seasonality in an additive or multiplicative way.
2.1.3. Holt-Winters Model
The Holt-Winters model, or triple smoothing model, is used for data with trend, level and seasonality [25]. This model has two variations: Additive Method and Multiplicative Method.
Additive Holt-Winters Method (A-HW): Represented by Equation (5):
(5)
where is the level, the trend, the seasonality at time t and the white noise. The estimates of the model components are given by Equations (6)–(8):(6)
(7)
(8)
where , and are the smoothing parameters for level, trend and seasonality ().Multiplicative Holt-Winters Method (M-HW): Represented by Equation (9):
(9)
The estimates of the model components are given by Equations (10)–(12):
(10)
(11)
(12)
Additive models are indicated for seasonal variations of constant amplitude [26], while multiplicative models are suggested for increasing or decreasing seasonal variations [27]. Both models were tested in this study. The next section will present the adjustments of the smoothing models.
After presenting the smoothing models, the next section will introduce the Box & Jenkins models and their variations.
2.2. Box & Jenkins Models
The Box & Jenkins models, such as ARIMA(), are notable for their accuracy in forecasting time series [25,28,29]. In addition to ARIMA, there are AR(p) and ARMA() models, and the challenge is to determine the values of (p), (d) and (q) and their respective coefficients [30].
Next, the AR(p), ARMA() and ARIMA() models are discussed.
2.2.1. Autoregressive Model—AR(p)
The AR(p) model uses p time lags as inputs to forecast future observations, represented by the linear combination of the past terms of the series, multiplied by the coefficients and adding a Gaussian white noise [13,31]. Based on a deterministic approach, AR(p) uses the Yule-Walker Equations to estimate its coefficients, minimizing the error between the observed data and the predictions [28]. Equation (13) represents the model.
(13)
where is the predicted value at time t, is the weighting coefficient for the delay of .Direct application of this model requires stationary data.
2.2.2. Autoregressive Moving Average Model—ARMA(p,q)
The ARMA() model combines autoregression (AR) and moving average (MA) components [13,32]. Equation (14) describes the model:
(14)
2.2.3. Autoregressive Integrated Moving Average Model—ARIMA(p,d,q)
The ARIMA() model extends ARMA with an order of differentiation d to remove trends and make the series stationary [33]. Equation (15) describes the model:
(15)
The direct application of the ARIMA model makes it possible to model random shocks using the forecast error from the previous step , where .
Maximum likelihood estimation can be used to determine the coefficients of the ARMA() and ARIMA() [34] models.
2.3. Bioinspired Optimization Tools
In this section, we present the algorithms used to optimize the ARMA() models, using two different strategies: Genetic Algorithms GA and Particle Swarm Optimization PSO. The details of how these algorithms were applied to the problem in question will be provided in Section 3 along with the application of the Ensemble model.
2.3.1. Genetic Algorithms (GA)
Optimization using Genetic Algorithms GA is widely used among algorithms inspired by biological processes. Based on the principles of the theory of evolution of Darwin [35,36,37], GAs model biological behavior to solve optimization problems. Introduced by [38] and refined by [39,40], GAs are recognized for identifying optimal or suboptimal solutions, and are robust to various problems, as they seek a global optimal solution [41].
In the context of GAs, the problem is modeled by representing individuals associated with the parameters and coefficients of the models to be optimized, evaluated by the degree of adaptability, known as fitness [14]. This establishes an analogy between an individual’s ability to thrive in an environment and the effectiveness of parameters in producing an optimal solution.
2.3.2. Particle Swarm Optimization (PSO)
A major advantage of metaheuristics is that they are derivative-independent, unlike classical optimization techniques such as gradient descent or Newton methods, which require derivatives of the predictor [42]. This makes them especially useful in problems where derivatives are unavailable or difficult to calculate.
Inspired by the social behavior of birds and fish, Particle Swarm Optimization PSO, proposed by [35], uses individual and collective experience to solve problems. PSO limits the distribution of swarm members in the search space by the current position () and velocity () [43]. The search for the best solution is guided by improving the local position () and the best global position ().
Reference The authors [44] proposed adding the inertia coefficient (), according to Equation (16), restricting the area surveyed. Values of vary from 0.9 for broad searches to 0.4 for narrow searches, affecting convergence. Cognitive components and influence the solution using past experiences, initially defined as 2 [44].
(16)
The performance of the PSO is influenced by and , controlling the speed and direction of the search. When the swarm starts, the particles are randomly distributed in the search space. Each particle is evaluated by the fitness function; the best position found is stored in () and (). The speed of each particle is updated in each iteration based on () and (), until the stopping criterion is reached.
After defining the linear models and optimization tools, the next section presents the Ensemble strategies used.
2.4. Tools for Combining Predictors Ensemble
One of the main advantages of ensembles lies in the ability to synergistically combine different individual models, which can result in remarkable improvements in the generalization process and in the accuracy of predictive models, as mentioned by [45,46,47]. Therefore, it can be said that a Ensemble has the ability to reduce error variance.
However, it is important to note that the effectiveness of ensembles is directly linked to the assertiveness of the individual models, which is influenced by the combination method adopted, as pointed out by [48]. There is no definition or consensus on which ensemble strategy should be used [49].
An ensemble can consist of several stages, such as the generation of individual models, the selection of models and, finally, their combination or aggregation [50,51].
The model generation stage is crucial for creating diversity within the Ensemble [51]. It can be classified as heterogeneous, using models with different architectures, or homogeneous, using models with the same architecture [50,52]. The combination of both approaches is common to diversify the Ensembles [53,54], although heterogeneous models can face challenges in maintaining diversity [53]. Homogeneous models, on the other hand, offer greater control over diversity [53].
The selection and combination of models are fundamental steps in the process of forming an ensemble, with the aim of balancing diversity and forecast accuracy. After generating the models, the next stage is selection, which is essential for building an efficient Ensemble.
A fundamental part of the composition of an ensemble is the selection of predictors, which can involve choosing all the available predictors or a specific subgroup, following established criteria. Selection can be static, using one model or subgroup for the entire test set [55,56], or dynamic, choosing models based on the region of competence during the test phase [50,57]. Although selection is not mandatory at this stage, it can influence the results. Considering all predictors for the final stage may be prudent to avoid selecting models that may underperform in the test set [57]. The next step in forming an Ensemble is the final combination of predictors.
This step integrates the results of the forecasts of the individual predictors, forming the Ensemble forecast (). In time series problems, it is common to aggregate the forecasts of k predictors to obtain more accurate results, usually using the mean or median of the forecasts [58,59].
Both the mean and the median are non-trainable ensemble models, reducing computational costs as it is not necessary to retrain the models repeatedly. The mean is represented by Equation (17), where are the predictions of the predictor i and m is the total number of predictors.
(17)
The median, represented by Equation (18), is useful in the presence of outliers, offering a robust estimate of the central tendency of the forecasts.
(18)
Figure 1 illustrates the process of combining previously trained models using a non-trainable approach.
In addition to the non-trainable ensembles, we also investigated the trainable ones, which differ in the assignment of weights to each predictor [60,61]. This stands out as the main contribution of this research.
Weights can be determined in various ways, such as minimum, maximum or product of the predictors’ outputs. A viable strategy is the weighted average, assigning greater weights to the models with the best performance [62,63,64].
The pseudo-inverse of Moore-Penrose can be used to calculate the weights, effectively adapting to the characteristics of the predictors [65]. Equation (19) provides the solution to this problem:
(19)
Figure 2 shows the procedure for combining previously trained models, incorporating an additional re-training step.
This additional step adjusts the weights of the ensemble models according to the evolution of the data or problem conditions, resulting in a trainable approach.
In this paper, we propose an ensemble of predictive models, where the final output is calculated as the weighted average of the models’ individual predictions, represented by Equation (20).
(20)
The weights are optimized using Genetic Algorithm GA and Particle Swarm Optimization PSO, in order to minimize the prediction error of the ensemble [66].
After defining the linear models, optimization and combination tools, the next section will present the evaluation of these steps.
3. Methodology
In this section, the stages for the development of this research will be presented. Figure 3 illustrates the organization of the stages.
All the statistical tests and computational results were developed using the Python 3.11.5 version programming language.
3.1. Database
The data analyzed, from the EIA [1], covers the monthly closing prices of WTI crude oil from January 1987 to February 2023, totaling 434 observations and showing total integrity with no missing or null records. The distribution of this data is illustrated in Figure 4.
To develop the models, 75% of the data was used for training, while the remaining 25% was used for testing, as shown in Figure 4.
3.2. Pre-Processing
After collecting the data, it was analyzed to identify behaviors such as trend, cyclicality, seasonality and a random term. One of the ways to detect these behaviors is to use certain tests, such as the Cox-Stuart test and the Friedman test [67,68].
The tests show that the series has a trend and seasonality. In this case, it is necessary to pre-process the data to ensure stationarity. The so-called stationary series have a constant mean, constant variance and autocovariance that does not depend on time, reflecting a more stable behavior of the data, which for modeling, especially the Box & Jenkins models, is a sine qua non condition [25,28,69].
For this study, we used the logarithmic transformation and the moving average with a 12-month window, along with differentiation, as shown in Equations (21)–( 23).
(21)
in this case will be the value of the logarithm, represents the time series at time t.(22)
and represents the value of the moving average at time t. The moving average is calculated as the average of the values with a window of N-periods. Now represents the iteration over all the points in the N-period window.If we then apply differentiation using Equation (23), you get:
(23)
Combining all the parts, the complete transformation of the time series is represented by .
The next step is to determine the parameters of the smoothing models. In this case, determine the parameters for level, trend and seasonality.
3.3. Estimating the Smoothing Coefficients
Exponential smoothing models require the determination of the , and coefficients. The literature indicates that there is no consensus on an ideal method for this determination [70]. Using the numerical evaluation of the cost function, the L-BFGS-B (The standard algorithm in the statsmodels library minimizes the mean square error (MSE) using the Quasi-Newton method, without the need to provide the Hessian matrix or the structure of the objective function) method was used to minimize the MSE and define the parameters in the training set.
Although Exponential Smoothing models do not require the series to be stationary, it was decided to use stationary data in this study. The adjusted parameters , and are shown in Table 1 and Appendix A.
After determining the coefficients of the smoothing models, we proceeded to apply the Box & Jenkins models.
3.4. Estimating the Coefficients and Orders of the Box & Jenkins
For the Box & Jenkins models, the orders and coefficients were determined in two ways: the classical approach and the optimization of the coefficients of the ARMA() model using GA and PSO.
The candidate orders were evaluated using the Autocorrelation Function (ACF) and Partial Autocorrelation Function (PACF) graphs. The coefficients were estimated by solving the Yule-Walker Equations for AR(p) [71]. For the ARMA() and ARIMA() models, the maximum likelihood estimator [34] was used. The d part of the ARIMA model was determined as 1 by applying a differentiation to the data.
Significant lags were defined by analyzing Figure 5, which makes it easier to understand the autocorrelation patterns in the adjusted time series.
The ACF and PACF help identify significant lag components for the AR and MA models, respectively [72]. Determining these parameters can be challenging due to the complexity and volume of the data. For the AR(p) model, although Figure 5 suggests testing up to lag 2, lags from 1 to 6 were observed. For the ARMA() model, the MA(q) part was tested up to order 6, as indicated by the ACF.
For the ARIMA() model, the same orders were tested, with d set to 1. The coefficients of the Box & Jenkins models are shown in Table 2. After analyzing the orders, it was decided to refine the choice of parameters, especially for the ARMA() model, using GA and PSO optimization for greater precision [73,74].
Siqueira2 Table 3 and Table 4. In GA, the parameters and were optimized with and , using one-point crossover, dynamic mutation and roulette wheel selection. The specific parameters are detailed in
In PSO Table 4, particles represent candidate solutions . The particles adjust their trajectories based on the best individual () and global () experiences. The inertia (w), cognitive () and social () coefficients modulate the search dynamics. 30 simulations were carried out to identify the optimal configuration.
The optimized values of are shown in Table 5. With the application of GA and PSO, all the linear models are ready to be combined. The ensembles can be developed using the techniques presented in Section 2.4.
3.5. Ensemble
To build ensemble 1, the average of the models’ forecasts was used, as described in Equation (17), for the forecast horizons of 1, 3, 6, 9 and 12 steps ahead. The forecasts were collected and the arithmetic mean was calculated, with all model outputs contributing equally to the final forecast, without the need for readjustment or retraining. Individual errors were calculated for each horizon.
Ensemble 2 was developed based on the median of the models’ forecasts, following the same steps as ensemble 1 and using Equation (18). Each model was evaluated individually.
Ensemble 3, unlike ensembles 1 and 2, is trainable. Initially, the model predictions were organized in a Y matrix and the actual values in a y vector. The weights that minimize the quadratic difference between predictions and actual values were calculated by applying the pseudo-inverse of Moore-Penrose to solve the least squares problem Equation (19).
For ensemble 4, the weighted average of the models’ predictions was used, with weights initially optimized using GA. A chromosome was formed representing the weights W, with the restriction that the weights add up to 1 and are non-negative. GA was applied with one-point crossover, dynamic mutation and tournament selection, as detailed in Table 6.
After simulations and tests with GA, the parameters for PSO were defined Table 7. Each particle in the PSO represents a candidate solution and adjusts its trajectory based on the best individual () and global () experiences, following Equation (16).
3.6. Post-Processing
After the initial transformations to the time series data, it was necessary to reverse the modifications to recover the “removed” values and return the forecasts to the original scale. This makes it easier to understand and visualize the results accurately, ensuring that the accuracy metrics are on the same scale as the original data.
The following evaluation metrics were used: MSE, MAE, MAPE and AE.
The Mean Square Error MSE calculates the average of the squared errors with Equation (24).
(24)
The Mean Absolute Percentage Error MAPE avoids different scale penalties, represented with Equation (25).
(25)
And the Mean Absolute Error MAE calculates the average of the absolute errors with Equation (26).
(26)
On the other hand, the Absolute Error AE is the difference between the observed value and the predicted value and can be calculated with Equation (27).
(27)
Section 4 will present the final results with the models adjusted and reversed as described in Section 3.6.
4. Results
This section presents the results of the models evaluated for each forecast horizon, based on the MSE, MAE and MAPE errors, followed by a ranking of the models (Table 8). For each horizon, the best result is illustrated next to the actual data, as well as the Absolute Error AE curves over time for the 14 models evaluated.The graphs are organized as follows:
Figure 6: A corresponds to the prediction of the best model and B o the evaluation of the AE for one-step ahead;
Figure 7: C represents the prediction of the best model and D the AE evaluation for three-steps ahead;
Figure 8: E shows the prediction of the best model and F the AE evaluation corresponding to six-steps ahead;
Figure 9: G shows the best model prediction and H the AE evaluation for nine-steps ahead;
Figure 10: I contains the prediction of the best model and J the AE evaluation of the absolute error considering twelve-steps ahead.
As shown in Table 8, ensemble 5, using the weighted average with PSO, stood out by dynamically adjusting the weights of the models in the Ensemble based on historical performance, maximizing overall accuracy. This flexibility justifies its superior performance compared to ensembles 1 and 2, which assign equal weights to each model, according to Equations (17) and (18). By looking at the scores assigned to each model according to its performance per evaluation metric, it is possible to construct a score with the sum of all the scores. It can be seen that although ensemble 5 had the best overall performance, ensemble 3 stood out with the best position in relation to MAPE error.
Figure 6 illustrates the best model for predicting a step forward on the test set (Observed).
Subfigure A contains the predicted values with the best model. While subfigure B presents the absolute error for each predicted value of all models. Next to it are the values with the MAE errors per model. This analogy is used for the other predicted steps.
After evaluating the one-step ahead forecasts, we moved on to analyze the three-steps ahead horizons. Table 9 shows the results of all the models based on the MSE, MAE and MAPE error metrics.
For this horizon, ensemble 3 stood out, using the pseudo-inverse of Moore-Penrose to combine the models, taking better advantage of their individual characteristics. The ensembles 4 and 5 also outperformed the individual models, indicating the effectiveness of the GA and PSO approaches. Individual models such as AR, ARMA and ARIMA showed relatively high errors, with the multiplicative Holt Winters model obtaining the highest MSE.
In the three-steps horizon, the ARMA-GA model outperformed ARMA-PSO, possibly due to uncertainties in the parameter selection process. The smoothing models behaved similarly to the one-step horizon, with larger errors in multi-step forecasts. Ensemble 3 again stood out in this horizon.
As mentioned above and illustrated in Table 9, ensemble 3 obtained better results than all the predictive models. This is because ensemble 3 is more precise when adjusting the weights, directly minimizing the prediction error. In this case, it provides more sensitive and accurate responses to fluctuations, which are more evident in forecasts with longer horizons. Its performance is also evident when evaluating the final ranking, thus obtaining a better score in all error metrics. It is worth noting, however, that as in the previous step, ensembles 4 and 5 obtained good results compared to the other ensembles, again highlighting the efficiency of using GA and PSO. In this sense, Figure 7 shows the best prediction model, obtained by ensemble 3.
Similarly, we went on to evaluate other forecast horizons, in this case for six-steps ahead, as shown in Table 10.
Ensemble 3 was again superior, reinforcing its ability to determine the best weightings for longer forecasts. The ensemble 5 also stood out, showing the efficiency of the optimization algorithms. Its performance is also evident when evaluating the final ranking, thus obtaining a better score in all the error metrics. Ensemble 2, which uses the median, performed reasonably well, being robust against outliers as the steps increase. Figure 8 shows the best result for this horizon, obtained by ensemble 3.
After these considerations, the forecasts for the models considering nine-steps ahead were evaluated, as illustrated in Table 11.
Ensemble 3 stood out again, as shown in Table 11. As the horizons increase, the errors of the individual models increase significantly, which does not occur in the ensembles. Ensemble 3 was the best for forecasts nine-steps ahead, as illustrated in Figure 9. Its performance is also evident when evaluating the final ranking, thus obtaining a better score in all the error metrics. The Box & Jenkins models maintained their performance, highlighting the efficiency of the GA and PSO algorithms. Although ensemble 2 was not the best, it obtained considerable results, demonstrating its robustness for longer horizon forecasts, due to the reduction in variability when using central values.
Finally, with regard to the last forecast horizon, considering twelve-steps ahead, Table 12 shows the results of all the models.
The Box & Jenkins models maintained the same results as the previous cases. In the Smoothing models, there was a change, with the additive and multiplicative models, previously the worst, becoming the best. The ensemble 3 remains the most effective.
The exponential smoothing models showed variations in results, with the Additive Holt Winters being the most effective, especially in long-term forecasts, due to its stability and predictability. The SES model also benefited from stationarity in shorter horizon forecasts.
Finally, the results reinforce that ensemble 3 significantly outperformed the individual models, and the ensembles in general proved superior at other forecast horizons.
After the aforementioned considerations for a forecast horizon of twelve-steps ahead, Figure 10 shows the best answer, in this case ensemble 3.
Several were analyzed and MSE, MAE and MAPE were used to evaluate them. These metrics illustrate average values (the best overall approximation in the analysis). Abrupt changes in the direction of the time series make it difficult for models to predict, but some models have the ability to adapt better than others. By analyzing AE, it is possible to see which models have the smallest outliers, which is additional behavioral information that the usual averages do not provide.
The results presented in this section confirm the concepts discussed in the Section 2.4, demonstrating the robustness of the ensemble models over different forecast horizons. Specifically, ensemble 5 proved to be superior in the forecast horizon of one-step ahead, while for forecasts of 3, 6, 9 and 12 steps ahead, ensemble 3 was superior to all models.
In general, ensemble models have different advantages and disadvantages. The mean is simple, but can be influenced by outliers. The median is robust against outliers, but can ignore variability. The Moore-Penrose inverse optimizes weights based on historical performance, and is accurate but computationally complex. Weighted averaging with PSO and GA dynamically adjusts the weights, improving accuracy, but requires more computing power. For short-term forecasts, the mean and median are effective; for the long term, the inverse of Moore-Penrose and the weighted mean offer better optimization, provided there is sufficient data.
5. Conclusions
The main contribution of this work is related to the use of the pseudo-inverse of Moore-Penrose to determine the weights of the models to be used in the formation of the Ensemble, in addition to the use of metaheuristics.
It is known that GA and PSO algorithms are widely used in the literature, although not so much for application in ensemble. In this sense, as an initial work, it was decided to use these techniques.
The results show that the ensemble models, especially those that used metaheuristics and the pseudo-inverse of Moore-Penrose, significantly improved the individual results of the predictive models at all forecast horizons.
After pre-processing the data, the model parameters were determined in various ways: for the smoothing models, a numerical model that minimizes the cost function was used; for the Box & Jenkins models, the Yule-Walker equations and maximum likelihood estimators were used, with delays tested exhaustively. Specifically for the ARMA model, two coefficient optimization techniques were used: GA and PSO. For the ensemble, several strategies were tested, including arithmetic mean, median, pseudo-inverse of Moore-Penrose and weighted mean with GA and PSO. The results showed that the ensemble approaches outperformed the individual models, with the weighted average with PSO (ensemble 5) standing out in step 1, and the pseudo-inverse of Moore-Penrose (ensemble 3) in the other steps.
The results indicate the feasibility of using ensembles in time series forecasting, allowing it to be applied to forecasting models other than linear ones.
In this sense, the research can be further developed with the insertion of other approaches aimed at technological development, such as the creation of other Ensembles. Mention could be made of the use of artificial neural networks to form a non-linear ensemble.
Conceptualization, J.L.F.d.S., Y.R.K., S.L.S.J. and H.V.S.; methodology, J.L.F.d.S., Y.R.K., S.L.S.J. and H.V.S.; software, J.L.F.d.S.; validation, J.L.F.d.S., Y.R.K., S.L.S.J., T.A.A. and H.V.S.; formal analysis, A.J.C.V.; investigation, J.L.F.d.S., Y.R.K. and H.V.S.; resources, T.A.A.; data curation, J.L.F.d.S., Y.R.K., S.L.S.J. and H.V.S.; writing—original draft preparation, J.L.F.d.S., Y.R.K., S.L.S.J. and H.V.S.; writing—review and editing, J.L.F.d.S., A.J.C.V., S.L.S.J. and H.V.S.; visualization, J.L.F.d.S.; supervision, H.V.S.; project administration, H.V.S.; funding acquisition, T.A.A. All authors have read and agreed to the published version of the manuscript.
Data are contained within the article.
The authors declare no conflicts of interest.
The following abbreviations are used in this manuscript:
ACF | Autocorrelation function |
AE | Absolute error |
A-HW | Additive Holt-Winters’ models |
AR | Autoregressive |
ARIMA | Autoregressive integrated moving average |
ARMA | Autoregressive-moving-average |
EIA | Energy Information administration |
GA | Genetic algorithm |
HES | Holt simple exponential smoothing |
MAE | Mean absolute error |
MAPE | Mean absolute percentage error |
M-HW | Multiplicative Holt-Winters’ models |
MSE | Mean Squared Error |
PACF | Partial autocorrelation function |
PSO | Particle swarm optimization |
SES | Simple exponential smoothing |
WTI | West Texas Intermediate |
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Smoothing Model Coefficients.
Models | | | |
---|---|---|---|
SES | 1.00 | - | - |
Holt | 1.00 | | - |
A-HW | | | |
M-HW | 1.00 | | |
Source: Own authorship (2024).
Box & Jenkins Model Coefficients.
Models | p | d | q |
---|---|---|---|
AR | 1 | - | - |
ARMA | 1 | - | 3 |
ARIMA | 6 | 1 | 6 |
Source: Own authorship (2024).
Description of GA Parameters—ARMA.
Parameters | Values |
---|---|
Population Size | 50 |
Number of Iterations | 30 |
Offspring Generated by Crossover | 50 |
Mutation Amplitude Regulator | 0.1 |
Percentage Mutation Rate | 10 % |
Frequency of Use of Local Search | 100 |
Local Search Delta | 0.05 |
Maximum Number of Generations | 150 |
Search Interval for | [−1, 1] |
Search Interval for | [−1, 1] |
Source: Own authorship (2024).
Description of the PSO—ARMA Parameters.
Parameters | Values |
---|---|
Number of Particles | 150 |
Number of Iterations | 150 |
Inertia Coefficient w | 0.7 |
Cognitive Term | 2.0 |
Social Term | 2.0 |
Maximum Number of Executions | 30 |
Local Search Delta | 0.05 |
Search Interval for | [−1, 1] |
Search Interval for | [−1, 1] |
Source: Own authorship (2024).
ARMA
Model | | | | |
---|---|---|---|---|
ARMA-GA | 0.1192 | −0.0081 | −0.0586 | −0.1268 |
ARMA-PSO | 0.6036 | −0.5537 | 0.1161 | −0.0438 |
Source: Own authorship (2024).
Description of the GA Parameters for Optimizing the Weights in a Weighted Average Ensemble.
Parameters | Values |
---|---|
Population Size | 150 |
Number of Iterations | 30 |
Offspring Generated by Crossover | By Generation, Half of the Population |
Crossover Rate | 90% |
Mutation Rate | 50% |
Mutation Amplitude (Standard Deviation) | 0.1 |
Tournament Size | 4 |
Number of Individual Models | 9 |
Source: Own authorship (2024).
Description of the PSO Parameters for Optimizing the Weights in a Weighted Average Ensemble.
Parameters | Values |
---|---|
Number of Particles | 150 |
Number of Iterations | 150 |
Inertia Coefficient w | 0.9 |
Cognitive Term | 2.0 |
Social Term | 2.0 |
Maximum Number of Executions | 30 |
Number of Models by Particle | 9 |
Source: Own authorship (2024).
Evaluation and Rankings for One-Step Ahead Forecasts.
Models | MSE | MAE | MAPE | Rank MSE | Rank MAE | Rank MAPE | Total Score | Final Ranking |
---|---|---|---|---|---|---|---|---|
Ensemble 5 | 26.1324 | 3.8052 | 0.0732 | 1 | 1 | 2 | 4 | 1 |
Ensemble 1 | 26.5979 | 3.8855 | 0.0748 | 2 | 3 | 3 | 8 | 2 |
ARMA-PSO | 35.7461 | 4.5461 | 0.0836 | 3 | 4 | 4 | 11 | 3 |
ARMA-GA | 36.0904 | 4.5779 | 0.0845 | 4 | 5 | 5 | 14 | 4 |
SES | 39.3276 | 4.7244 | 0.0878 | 5 | 8 | 8 | 21 | 5 |
Holt | 39.7390 | 4.7383 | 0.0881 | 6 | 9 | 9 | 24 | 6 |
ARMA | 42.1666 | 4.9635 | 0.0917 | 7 | 6 | 7 | 20 | 7 |
ARIMA | 42.2707 | 4.9896 | 0.0915 | 8 | 7 | 6 | 21 | 8 |
AR | 42.5677 | 5.0139 | 0.0924 | 9 | 11 | 10 | 30 | 9 |
M-HW | 43.6446 | 5.0792 | 0.0931 | 10 | 13 | 12 | 35 | 10 |
A-HW | 42.7063 | 5.0471 | 0.0940 | 11 | 12 | 13 | 36 | 11 |
Ensemble 2 | 38.1103 | 4.8385 | 0.0888 | 7 | 10 | 11 | 28 | 12 |
Ensemble 4 | 42.8594 | 5.0280 | 0.9389 | 12 | 14 | 14 | 40 | 13 |
Ensemble 3 | 47.2520 | 3.8363 | 0.0732 | 14 | 2 | 1 | 17 | 14 |
Source: Own authorship (2024).
Evaluation and Rankings for Three-Steps Ahead Forecasts.
Models | MSE | MAE | MAPE | Rank MSE | Rank MAE | Rank MAPE | Total Score | Final Ranking |
---|---|---|---|---|---|---|---|---|
Ensemble 3 | 31.5856 | 4.2631 | 0.0809 | 1 | 1 | 1 | 3 | 1 |
Ensemble 5 | 46.6288 | 5.1197 | 0.1039 | 2 | 6 | 2 | 10 | 2 |
ARMA-GA | 62.9834 | 4.5779 | 0.1230 | 3 | 3 | 5 | 11 | 3 |
ARMA-PSO | 72.1269 | 6.2979 | 0.1148 | 4 | 4 | 3 | 11 | 4 |
Ensemble 4 | 62.1147 | 6.0290 | 0.1206 | 5 | 5 | 4 | 14 | 5 |
SES | 85.8354 | 6.8282 | 0.1371 | 6 | 8 | 6 | 20 | 6 |
Holt | 87.4450 | 6.9225 | 0.1385 | 7 | 9 | 7 | 23 | 7 |
Ensemble 1 | 80.2616 | 6.6111 | 0.1325 | 8 | 7 | 8 | 23 | 8 |
ARMA | 93.7271 | 4.3698 | 0.1500 | 9 | 2 | 9 | 20 | 9 |
ARIMA | 94.1366 | 7.1262 | 0.1451 | 10 | 10 | 11 | 31 | 10 |
AR | 97.7222 | 7.4393 | 0.1572 | 11 | 11 | 12 | 34 | 11 |
A-HW | 100.0355 | 7.1951 | 0.1491 | 12 | 12 | 10 | 34 | 12 |
M-HW | 105.9239 | 7.5685 | 0.1528 | 13 | 13 | 13 | 39 | 13 |
Ensemble 2 | 116.4333 | 8.1610 | 0.1540 | 14 | 14 | 14 | 42 | 14 |
Source: Own authorship (2024).
Evaluation and Rankings for Six-Steps Ahead Forecasts.
Models | MSE | MAE | MAPE | Rank MSE | Rank MAE | Rank MAPE | Total Score | Final Ranking |
---|---|---|---|---|---|---|---|---|
Ensemble 3 | 36.6887 | 4.8088 | 0.0940 | 1 | 1 | 1 | 3 | 1 |
Ensemble 4 | 42.2360 | 5.1421 | 0.0985 | 2 | 2 | 2 | 6 | 2 |
Ensemble 5 | 70.8841 | 6.1835 | 0.1182 | 3 | 3 | 3 | 9 | 3 |
Ensemble 2 | 92.1716 | 6.9610 | 0.1375 | 4 | 4 | 4 | 12 | 4 |
ARMA-PSO | 102.7793 | 7.6440 | 0.1418 | 5 | 5 | 5 | 15 | 5 |
ARMA-GA | 108.8052 | 7.9844 | 0.1469 | 6 | 6 | 6 | 18 | 6 |
Ensemble 1 | 113.8113 | 7.7389 | 0.1515 | 7 | 7 | 7 | 21 | 7 |
SES | 145.3774 | 6.6541 | 0.1718 | 11 | 8 | 10 | 29 | 8 |
ARIMA | 143.9546 | 8.5516 | 0.1633 | 10 | 9 | 7 | 26 | 9 |
AR | 141.2104 | 8.6026 | 0.1633 | 9 | 10 | 8 | 27 | 10 |
ARMA | 136.2695 | 8.8349 | 0.1656 | 8 | 11 | 9 | 28 | 11 |
A-HW | 152.3447 | 8.8144 | 0.1734 | 13 | 12 | 12 | 37 | 12 |
Holt | 151.8369 | 8.8701 | 0.1747 | 12 | 13 | 11 | 36 | 13 |
M-HW | 181.0155 | 9.7933 | 0.1855 | 14 | 14 | 13 | 41 | 14 |
Source: Own authorship (2024).
Evaluation and Rankings for Nine-Steps Ahead Forecasts.
Models | MSE | MAE | MAPE | Rank MSE | Rank MAE | Rank MAPE | Total Score | Final Ranking |
---|---|---|---|---|---|---|---|---|
Ensemble 3 | 40.7999 | 5.1785 | 0.0996 | 1 | 2 | 2 | 5 | 1 |
Ensemble 4 | 42.8763 | 5.0218 | 0.0948 | 2 | 1 | 1 | 4 | 2 |
Ensemble 5 | 76.2204 | 6.6119 | 0.1212 | 3 | 3 | 3 | 9 | 3 |
Ensemble 2 | 107.0554 | 7.7230 | 0.1550 | 4 | 4 | 4 | 12 | 4 |
ARMA-GA | 116.7820 | 8.2592 | 0.1580 | 5 | 5 | 6 | 16 | 5 |
ARMA-PSO | 138.9001 | 8.3752 | 0.1573 | 6 | 6 | 5 | 17 | 6 |
Ensemble 1 | 131.7656 | 8.6014 | 0.1707 | 7 | 7 | 7 | 21 | 7 |
SES | 195.8325 | 10.2751 | 0.2050 | 10 | 8 | 9 | 27 | 8 |
ARMA | 183.8635 | 10.2683 | 0.1816 | 8 | 9 | 8 | 25 | 9 |
AR | 207.4238 | 10.7631 | 0.1998 | 9 | 10 | 10 | 29 | 10 |
Holt | 209.9169 | 10.7295 | 0.2121 | 11 | 11 | 10 | 32 | 11 |
ARIMA | 235.2772 | 11.7777 | 0.2230 | 12 | 12 | 11 | 35 | 12 |
M-HW | 262.9169 | 12.3568 | 0.2253 | 14 | 13 | 12 | 39 | 13 |
A-HW | 254.0866 | 12.4684 | 0.2357 | 13 | 14 | 13 | 40 | 14 |
Source: Own authorship (2024).
Evaluation and Rankings for Twelve-Steps Ahead Forecasts.
Models | MSE | MAE | MAPE | Rank MSE | Rank MAE | Rank MAPE | Total Score | Final Ranking |
---|---|---|---|---|---|---|---|---|
Ensemble 3 | 42.399 | 4.9668 | 0.0980 | 1 | 1 | 1 | 3 | 1 |
Ensemble 4 | 59.7894 | 5.7881 | 0.1103 | 2 | 2 | 2 | 6 | 2 |
Ensemble 5 | 84.6815 | 6.7984 | 0.1286 | 3 | 3 | 3 | 9 | 3 |
Ensemble 2 | 106.8666 | 7.3913 | 0.1541 | 4 | 4 | 4 | 12 | 4 |
ARMA-GA | 120.2188 | 8.4345 | 0.1550 | 5 | 6 | 5 | 16 | 5 |
Ensemble 1 | 122.8396 | 8.0137 | 0.1633 | 6 | 5 | 7 | 18 | 6 |
ARMA-PSO | 147.1678 | 8.9393 | 0.1641 | 7 | 7 | 6 | 20 | 7 |
ARMA | 160.8367 | 9.5246 | 0.1768 | 8 | 8 | 8 | 24 | 8 |
AR | 179.3992 | 9.7180 | 0.1840 | 9 | 9 | 9 | 27 | 9 |
ARIMA | 191.2371 | 9.7794 | 0.1891 | 10 | 10 | 10 | 30 | 10 |
A-HW | 190.4970 | 9.9673 | 0.1967 | 11 | 11 | 11 | 33 | 11 |
M-HW | 238.7421 | 11.3661 | 0.2129 | 12 | 12 | 12 | 36 | 12 |
SES | 346.1113 | 13.1934 | 0.2469 | 13 | 13 | 13 | 39 | 13 |
Holt | 379.3841 | 13.8740 | 0.2570 | 14 | 14 | 14 | 42 | 14 |
Source: Own authorship (2024).
Appendix A
Together with the parameters observed in
For the application of the Autoregressive Moving Average Model-ARMA(
With the parameters observed in
Adjustments of the
Weights Associated with Ensemble 3.
Models | Weights |
---|---|
SES | 5.3110 |
ARMA | 0.3103 |
AR | 0.9854 |
ARMA-PSO | −0.0534 |
A-HW | 0.6266 |
M-HW | −1.0575 |
HOLT | −4.8867 |
ARIMA | −1.1908 |
ARIMA-GA | 0.6770 |
References
1. Administration, U.E.I. Petroleum & Other Liquids. 2023; Available online: https://www.eia.gov/petroleum (accessed on 3 June 2023).
2. Duan, K.; Ren, X.; Wen, F.; Chen, J. Evolution of the information transmission between Chinese and international oil markets: A quantile-based framework. J. Commod. Mark.; 2023; 29, 100304. [DOI: https://dx.doi.org/10.1016/j.jcomm.2022.100304]
3. Balcilar, M.; Gabauer, D.; Umar, Z. Crude Oil futures contracts and commodity markets: New evidence from a TVP-VAR extended joint connectedness approach. Resour. Policy; 2021; 73, 102219. [DOI: https://dx.doi.org/10.1016/j.resourpol.2021.102219]
4. Lyu, Y.; Tuo, S.; Wei, Y.; Yang, M. Time-varying effects of global economic policy uncertainty shocks on crude oil price volatility: New evidence. Resour. Policy; 2021; 70, 101943. [DOI: https://dx.doi.org/10.1016/j.resourpol.2020.101943]
5. Khan, K.; Su, C.W.; Umar, M.; Yue, X.G. Do crude oil price bubbles occur?. Resour. Policy; 2021; 71, 101936. [DOI: https://dx.doi.org/10.1016/j.resourpol.2020.101936]
6. Wang, X.; Li, X.; Li, S. Point and interval forecasting system for crude oil price based on complete ensemble extreme-point symmetric mode decomposition with adaptive noise and intelligent optimization algorithm. Resour. Policy; 2022; 328, 120194. [DOI: https://dx.doi.org/10.1016/j.apenergy.2022.120194]
7. Karasu, S.; Altan, A. Crude oil time series prediction model based on LSTM network with chaotic Henry gas solubility optimization. Energy; 2022; 242, 122964. [DOI: https://dx.doi.org/10.1016/j.energy.2021.122964]
8. Ren, X.; Liu, Z.; Jin, C.; Lin, R. Oil price uncertainty and enterprise total factor productivity: Evidence from China. Int. Rev. Econ. Financ.; 2023; 83, pp. 201-218. [DOI: https://dx.doi.org/10.1016/j.iref.2022.08.024]
9. Yuan, J.; Li, J.; Hao, J. A dynamic clustering ensemble learning approach for crude oil price forecasting. Eng. Appl. Artif. Intell.; 2023; 123, 106408. [DOI: https://dx.doi.org/10.1016/j.engappai.2023.106408]
10. Zhang, T.; Tang, Z. The dependence and risk spillover between economic uncertainties and the crude oil market: New evidence from a Copula-CoVaR approach incorporating the decomposition technique. Environ. Sci. Pollut. Res.; 2023; 83, pp. 104116-104134. [DOI: https://dx.doi.org/10.1007/s11356-023-29624-0]
11. Inacio, C.; Kristoufek, L.; David, S. Assessing the impact of the Russia—Ukraine war on energy prices: A dynamic cross-correlation analysis. Phys. A Stat. Mech. Its Appl.; 2023; 626, 129084. [DOI: https://dx.doi.org/10.1016/j.physa.2023.129084]
12. An, S.; An, F.; Gao, X.; Wang, A. Early warning of critical transitions in crude oil price. Energy; 2023; 280, 128089. [DOI: https://dx.doi.org/10.1016/j.energy.2023.128089]
13. Siqueira, H.; Boccato, L.; Attux, R.; Lyra, C. Unorganized machines for seasonal streamflow series forecasting. Int. J. Neural Syst.; 2014; 24, pp. 1299-1316. [DOI: https://dx.doi.org/10.1142/S0129065714300095] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/24552508]
14. Siqueira, H.; Belotti, J.T.; Boccato, L.; Luna, I.; Attux, R.; Lyra, C. Recursive linear models optimized by bioinspired metaheuristics to streamflow time series prediction. Int. Trans. Oper. Res.; 2023; 30, pp. 742-773. [DOI: https://dx.doi.org/10.1111/itor.12908]
15. Ren, Y.; Zhang, L.; Suganthan, P. Ensemble Classification and Regression-Recent Developments, Applications and Future Directions (Review Article). IEEE Comput. Intell. Mag.; 2016; 11, pp. 41-53. [DOI: https://dx.doi.org/10.1109/MCI.2015.2471235]
16. Yu, L.; Xu, H.; Tang, L. LSSVR ensemble learning with uncertain parameters for crude oil price forecasting. Appl. Soft Comput.; 2017; 56, pp. 692-701. [DOI: https://dx.doi.org/10.1016/j.asoc.2016.09.023]
17. Fathalla, A.; Alameer, Z.; Abbas, M.; Ali, A. A Deep Learning Ensemble Method for Forecasting Daily Crude Oil Price Based on Snapshot Ensemble of Transformer Model. Comput. Syst. Sci. Eng.; 2023; 46, pp. 929-950. [DOI: https://dx.doi.org/10.32604/csse.2023.035255]
18. Cen, Z.; Wang, J. Crude oil price prediction model with long short term memory deep learning based on prior knowledge data transfer. Energy; 2017; 169, pp. 160-171. [DOI: https://dx.doi.org/10.1016/j.energy.2018.12.016]
19. Wang, M.; Tian, L.; Zhou, P. A novel approach for oil price forecasting based on data fluctuation network. Energy Econ.; 2018; 71, pp. 201-212. [DOI: https://dx.doi.org/10.1016/j.eneco.2018.02.021]
20. Bildirici, M.; Bayazit, N.G.; Yasemen, U. Analyzing crude oil prices under the impact of COVID-19 by using lstargarchlstm. Energies; 2020; 13, 2980. [DOI: https://dx.doi.org/10.3390/en13112980]
21. Bekiroglu, K.; Duru, O.; Gulay, E.; Su, R.; Lagoa, C. Predictive analytics of crude oil prices by utilizing the intelligent model search engine. Appl. Energy; 2018; 228, pp. 2387-2397. [DOI: https://dx.doi.org/10.1016/j.apenergy.2018.07.071]
22. Gardner, E.S. Exponential smoothing: The state of the art—Part II. Int. J. Forecast.; 2006; 22, pp. 637-666. [DOI: https://dx.doi.org/10.1016/j.ijforecast.2006.03.005]
23. Saputra, N.D.; Aziz, A.; Harjito, B. Parameter optimization of Brown’s and Holt’s double exponential smoothing using golden section method for predicting Indonesian Crude Oil Price (ICP). Proceedings of the 2016 3rd International Conference on Information Technology, Computer, and Electrical Engineering (ICITACEE); Semarang, Indonesia, 19–20 October 2016; pp. 356-360. [DOI: https://dx.doi.org/10.1109/ICITACEE.2016.7892471]
24. Hyndman, R.J.; Khandakar, Y. Automatic Time Series Forecasting: The forecast Package for R. J. Stat. Softw.; 2008; 27, pp. 1-22. [DOI: https://dx.doi.org/10.18637/jss.v027.i03]
25. Hyndman, R.J.; Athanasopoulos, G. Forecasting Principles and Practice; OTexts: Melbourne, Australia, 2021; Volume 3, pp. 1-442.
26. Awajan, A.M.; Ismail, M.T.; Al Wadi, S. Improving forecasting accuracy for stock market data using EMD-HW bagging. PLoS ONE; 2018; 13, 199582. [DOI: https://dx.doi.org/10.1371/journal.pone.0199582]
27. Papastefanopoulos, V.; Linardatos, P.; Kotsiantis, S. COVID-19: A Comparison of Time Series Methods to Forecast Percentage of Active Cases per Population. Appl. Sci.; 2020; 10, 3880. [DOI: https://dx.doi.org/10.3390/app10113880]
28. Box, G.; Jenkis, G.; Reinsel, C.; Ljung, M. Time series analysis: Forecasting and control. Wiley Ser. Probab. Stat. N. J.; 2015; 301, pp. 1-709.
29. Theerthagiri, P.; Ruby, A.U. Seasonal learning based ARIMA algorithm for prediction of Brent oil Price trends. Multimed. Tools Appl.; 2023; 18, pp. 2485-24504. [DOI: https://dx.doi.org/10.1007/s11042-023-14819-x]
30. Gujarati, D.N.; Porter, D.C. Econometria Básica-5; Amgh Editora: Porto Alegre, Brazil, 2011.
31. Haykin, S. Adaptive Filter Theory; Pearson Education India: Chennai, India, 2002.
32. Shadab, T.; Ahmad, S.; Said, S. Spatial forecasting of solar radiation using ARIMA model. Remote Sens. Appl. Soc. Environ.; 2023; 20, 100427. [DOI: https://dx.doi.org/10.1016/j.rsase.2020.100427]
33. Almasarweh, M.; Wadi, S.A. ARIMA Model in Predicting Banking Stock Market Data. Mod. Appl. Sci.; 2018; 12, pp. 309-312. [DOI: https://dx.doi.org/10.5539/mas.v12n11p309]
34. Zhong, C. Oracle-efficient estimation and trend inference in non-stationary time series with trend and heteroscedastic ARMA error. Comput. Stat. Data Anal.; 2024; 193, e1475. [DOI: https://dx.doi.org/10.1016/j.csda.2024.107917]
35. Kennedy, J.; Eberhart, R. Particle swarm optimization. Proceedings of the Proceedings of ICNN’95—International Conference on Neural Networks; Perth, WA, Australia, 27 November–1 December 1995; Bastos Filho, C.J.A.; Pozo, A.R.; Lopes, H.S. IEEE: New York, NY, USA, 1995; Volume 4, pp. 1942-1948. [DOI: https://dx.doi.org/10.1109/ICNN.1995.488968]
36. Cortez, P.; Rocha, M.; Neves, J. Evolving Time Series Forecasting ARMA Models. J. Heuristics; 2012; 10, pp. 137-151. [DOI: https://dx.doi.org/10.1023/B:HEUR.0000034714.09838.1e]
37. Binitha, S.; Sathya, S.S. A survey of bio inspired optimization algorithms. Int. J. Soft Comput. Eng.; 2012; 2, pp. 137-151.
38. Holland, J.H. Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence; MIT Press: Cambridge, MA, USA, 1975.
39. Goldberg, D. Genetic Algorithms in Search, Optimization, and Machine Learning; Addison-Wesley Publishing Company: London, UK, 1989.
40. Michalewicz, Z.; Schoenauer, M. Evolving Time Series Forecasting ARMA Models. Evol. Comput.; 1996; 4, pp. 1-32. [DOI: https://dx.doi.org/10.1162/evco.1996.4.1.1]
41. Eren, B.; Rezan, U.V.; Ufuk, Y.; Erol, E. A modified genetic algorithm for forecasting fuzzy time series. Appl. Intell.; 2014; 41, pp. 453-463. [DOI: https://dx.doi.org/10.1007/s10489-014-0529-x]
42. Aljamaan, I.; Alenany, A. Identification of Wiener Box-Jenkins Model for Anesthesia Using Particle Swarm Optimization. Appl. Sci.; 2022; 12, 4817. [DOI: https://dx.doi.org/10.3390/app12104817]
43. Edalatpanah, S.A.; Hassani, F.S.; Smarandache, F.; Sorourkhah, A.; Pamucar, D.; Cui, B. A hybrid time series forecasting method based on neutrosophic logic with applications in financial issues. Eng. Appl. Artif. Intell.; 2024; 129, 107531. [DOI: https://dx.doi.org/10.1016/j.engappai.2023.107531]
44. Kennedy, J.; Eberhart, R.; Shi, Y. Swarm Intelligence; Evolutionary Computation Series; Elsevier Science: Amsterdam, The Netherlands, 2001.
45. Donate, J.P.; Cortez, P.; Sánchez, G.G.; de Miguel, A.S. Time series forecasting using a weighted cross-validation evolutionary artificial neural network ensemble. Neurocomputing; 2013; 109, pp. 27-32. [DOI: https://dx.doi.org/10.1016/j.neucom.2012.02.053]
46. Silva, E.G.; de O. Júunior, D.S.; Cavalcanti,, G.D.C.; de Mattos Neto, P.S.G. Improving the accuracy of intelligent forecasting models using the Perturbation Theory. Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN); Rio de Janeiro, Brazil, 8–13 July 2018; IEEE: New York, NY, USA, 2018; pp. 1-7. [DOI: https://dx.doi.org/10.1109/IJCNN.2018.8489697]
47. Wang, L.; Wang, Z.; Qu, H.; Liu, S. Optimal Forecast Combination Based on Neural Networks for Time Series Forecasting. Appl. Soft Comput.; 2018; 66, pp. 1-17. [DOI: https://dx.doi.org/10.1016/j.asoc.2018.02.004]
48. Perrone, M.P.; Cooper, L. When Networks Disagree: Ensemble Methods for Hybrid Neural Networks; World Scientific Publishing: Singapore, 1995; Volume 109, pp. 1-404.
49. Sun, Y.; Tang, K.; Zhu, Z.; Yao, X. Concept Drift Adaptation by Exploiting Historical Knowledge. IEEE Trans. Neural Netw. Learn. Syst.; 2018; 29, pp. 4822-4832. [DOI: https://dx.doi.org/10.1109/TNNLS.2017.2775225] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/29993956]
50. Britto, A.S.; Sabourin, R.; Oliveira, L.E. Dynamic selection of classifiers—A comprehensive review. Pattern Recognit.; 2014; 47, pp. 3665-3680. [DOI: https://dx.doi.org/10.1016/j.patcog.2014.05.003]
51. Nosrati, V.; Rahmani, M. An ensemble framework for microarray data classification based on feature subspace partitioning. Comput. Biol. Med.; 2022; 148, 105820. [DOI: https://dx.doi.org/10.1016/j.compbiomed.2022.105820]
52. Wilson, J.; Chaudhury, S.; Lall, B. Homogeneous—Heterogeneous Hybrid Ensemble for concept-drift adaptation. Neurocomputing; 2023; 557, 126741. [DOI: https://dx.doi.org/10.1016/j.neucom.2023.126741]
53. Mendes-Moreira, J.A.; Soares, C.; Jorge, A.M.; Sousa, J.F.D. Ensemble approaches for regression: A survey. ACM Comput. Surv.; 2012; 45, pp. 1-40. [DOI: https://dx.doi.org/10.1145/2379776.2379786]
54. Heinermann, J.; Kramer, O. Machine learning ensembles for wind power prediction. Renew. Energy; 2016; 89, pp. 671-679. [DOI: https://dx.doi.org/10.1016/j.renene.2015.11.073]
55. Ma, Z.; Dai, Q. Selected an Stacking ELMs for Time Series Prediction. Neural Process. Lett.; 2016; 44, pp. 831-856. [DOI: https://dx.doi.org/10.1007/s11063-016-9499-9]
56. de Oliveira, J.F.L.; Silva, E.G.; de Mattos Neto, P.S.G. A Hybrid System Based on Dynamic Selection for Time Series Forecasting. IEEE Trans. Neural Netw. Learn. Syst.; 2022; 33, pp. 3251-3263. [DOI: https://dx.doi.org/10.1109/TNNLS.2021.3051384] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33513115]
57. Cruz, R.M.; Sabourin, R.; Cavalcanti, G.D. Dynamic classifier selection: Recent advances and perspectives. Inf. Fusion; 2018; 41, pp. 195-216. [DOI: https://dx.doi.org/10.1016/j.inffus.2017.09.010]
58. Barrow, D.K.; Crone, S.F.; Kourentzes, N. An evaluation of neural network ensembles and model selection for time series prediction. Proceedings of the the 2010 International Joint Conference on Neural Networks IJCNN; Barcelona, Spain, 18–23 July 2010; IEEE: New York, NY, USA, 2010; pp. 1-8. [DOI: https://dx.doi.org/10.1109/IJCNN.2010.5596686]
59. Kourentzes, N.; Barrow, D.K.; Crone, S.F. Neural network ensemble operators for time series forecasting. Expert Syst. Appl.; 2014; 41, pp. 4235-4244. [DOI: https://dx.doi.org/10.1016/j.eswa.2013.12.011]
60. Kazmaier, J.; van Vuuren, J.H. The power of ensemble learning in sentiment analysis. Expert Syst. Appl.; 2022; 187, 115819. [DOI: https://dx.doi.org/10.1016/j.eswa.2021.115819]
61. Chung, D.; Yun, J.; Lee, J.; Jeon, Y. Predictive model of employee attrition based on stacking ensemble learning. Expert Syst. Appl.; 2023; 215, 119364. [DOI: https://dx.doi.org/10.1016/j.eswa.2022.119364]
62. Kuncheva, L.I.; Rodríguez, J.J. A weighted voting framework for classifiers ensembles. Knowl. Inf. Syst.; 2014; 38, pp. 259-275. [DOI: https://dx.doi.org/10.1007/s10115-012-0586-6]
63. Large, J.; Lines, J.; Bagnall, A. A probabilistic classifier ensemble weighting scheme based on cross-validated accuracy estimates. Data Min. Knowl. Discov.; 2019; 33, pp. 1674-1709. [DOI: https://dx.doi.org/10.1007/s10618-019-00638-y]
64. Baradaran, R.; Amirkhani, H. Ensemble learning-based approach for improving generalization capability of machine reading comprehension systems. Neurocomputing; 2021; 33, pp. 229-242. [DOI: https://dx.doi.org/10.1016/j.neucom.2021.08.095]
65. Baksalary, O.M.; Trenkler, G. The Moore–Penrose inverse: A hundred years on a frontline of physics research. Eur. Phys. J. H; 2021; 46, 9. [DOI: https://dx.doi.org/10.1140/epjh/s13129-021-00011-y]
66. Safari, A.; Davallou, M. Oil price forecasting using a hybrid model. Energy; 2018; 148, pp. 49-58. [DOI: https://dx.doi.org/10.1016/j.energy.2018.01.007]
67. Cox, D.R.; Stuart, A. Some quick sign tests for trend in location and dispersion. Biometrika; 1955; 42, pp. 80-95. [DOI: https://dx.doi.org/10.1093/biomet/42.1-2.80]
68. Sprent, P.; Smeeton, N.C. Applied Nonparametric Statistical Methods; CRC Press: Boca Raton, FL, USA, 2016.
69. Chyon, F.A.; Suman, M.N.H.; Fahim, M.R.I.; Ahmmed, M.S. Time series analysis and predicting COVID-19 affected patients by ARIMA model using machine learning. J. Virol. Methods; 2022; 301, 114433. [DOI: https://dx.doi.org/10.1016/j.jviromet.2021.114433] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/34919977]
70. Montgomery, D.; Jennings, C.L.; Kulachi, M. Introduction to Time Series Analysis and Forecasting; John Wiley & Sons: Cambridge, MA, USA, 2008.
71. Siqueira, H.V.; Luna, I. Modelos Lineares Realimentados de Previsão: Um Estudo Utilizando Algoritmos Evoluciionários. Proceedings of the Anais do 12 Congresso Brasileiro de Inteligência Computacional; Curitiba, Brazil, 13–16 October 2016; Bastos Filho, C.J.A.; Pozo, A.R.; Lopes, H.S. pp. 1-6.
72. Awan, T.M.; Aslam, F. Prediction of daily COVID-19 cases in European countries using automatic ARIMA model. J. Public Health Res.; 2021; 9, pp. 4101-4111. [DOI: https://dx.doi.org/10.4081/jphr.2020.1765]
73. Alsharef, A.; Aggarwal, K.; Sonia,; Kumar, M.; Mishra, A. Review of ML and AutoML Solutions to Forecast Time-Series Data. Arch. Comput. Methods Eng.; 2022; 46, pp. 5297-5311. [DOI: https://dx.doi.org/10.1007/s11831-022-09765-0] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/35669518]
74. Ubaid, A.; Hussain, F.; Saqib, M. Container Shipment Demand Forecasting in the Australian Shipping Industry: A Case Study of Asia—Oceania Trade Lane. J. Mar. Sci. Eng.; 2021; 9, 968. [DOI: https://dx.doi.org/10.3390/jmse9090968]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
This paper investigated the use of linear models to forecast crude oil futures prices (WTI) on a monthly basis, emphasizing their importance for financial markets and the global economy. The main objective was to develop predictive models using time series analysis techniques, such as autoregressive (AR), autoregressive moving average (ARMA), autoregressive integrated moving average (ARIMA), as well as ARMA variants adjusted by genetic algorithms (ARMA-GA) and particle swarm optimization (ARMA-PSO). Exponential smoothing techniques, including SES, Holt, and Holt-Winters, in additive and multiplicative forms, were also covered. The models were integrated using ensemble techniques, by the mean, median, Moore-Penrose pseudo-inverse, and weighted averages with GA and PSO. The methodology adopted included pre-processing that applied techniques to ensure the stationarity of the data, which is essential for reliable modeling. The results indicated that for one-step-ahead forecasts, the weighted average ensemble with PSO outperformed traditional models in terms of error metrics. For multi-step forecasts (3, 6, 9 and 12), the ensemble with the Moore-Penrose pseudo-inverse showed better results. This study has shown the effectiveness of combining predictive models to forecast future values in WTI oil prices, offering a useful tool for analysis and applications. However, it is possible to expand the idea of applying linear models to non-linear models.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details






1 Graduate Program in Industrial Engineering (PPGEP), Federal University of Technology—Paraná, Ponta Grossa 84017-220, Brazil;
2 Graduate Program in Mechanical Engineering, Federal University of Technology—Paraná, Ponta Grossa 84017-220, Brazil;
3 Department of Industrial Engineering, Federal University of Technology—Paraná, Ponta Grossa 84017-220, Brazil;
4 Graduate Program in Electrical Engineering, Federal University of Technology—Paraná, Ponta Grossa 84017-220, Brazil;
5 Department of Industrial Engineering, Federal University of Technology—Paraná, Ponta Grossa 84017-220, Brazil;