1. Introduction
Ensuring an adequate supply of energy is one of the national priorities for every nation in the world, because all the economic activities of a nation rely on energy in general, electricity in particular. Electricity is also an essential part of people’s daily activities [1]. In addition, electricity consumption affects public/private investment planning of electricity infrastructure, and electricity production/generation and distribution [2]. Given these facts, it is crucial to have an accurate and reliable forecasting model for electricity consumption.
In recent decades, numerous approaches have been proposed, and the most commonly used methods for electricity consumption forecasting mainly consist of grey models [3,4], multiple regression models [5,6], artificial neural network (ANN) models [7,8] and support vector machine (SVM) models [9], exponential smoothing models [10,11], and auto-regressive integral moving average (ARIMA) models [12]. These models can be generally classified into three categories: artificial intelligence models, uncertainty models, and time series models. A trend has emerged in that more and more complex models using artificial intelligence techniques, such as ANN and SVM [13,14], have been proposed to discover the potentially predictive relationships between electricity consumption and input variables. However, there are many factors that affect electricity consumption, such as temperature [15], population [12], economic growth [1], and power facilities [16], and the major disadvantage of SVR and ANN is that they cannot select important input features within a dataset. Furthermore, artificial intelligence models are difficult to implement, which discourages practitioners with poor skills in data analytics. The uncertainty models like grey models can describe the characteristics of an uncertain system; however, the limitation of this type of model is that it is not suitable for forecasting seasonal time series [17]. In this regard, a simple but sufficiently accurate forecasting model is appreciated, especially if the model addresses the specific nature of monthly electricity consumption patterns such as seasonality.
Time series models assume that the historical data have discounted all relevant factors in an internal structure, thus avoiding the need for selection of input variables [18]. Seasonal exponential smoothing (SES) models, as one of the most popular groups of time series models, have motivated a variety of studies in the area of electrical energy, owing to their intuitive function forms, simple implementation, and ability to capture seasonality. In Reference [19], multiplicative Holt–Winters exponential smoothing was used to forecast medium-term electricity consumption, the fruit fly optimization algorithm was applied to smooth parameter selection, and the results indicate that the model allowed the authors to accurately forecast periodic series with relatively few training samples. In Reference [20], the Holt–Winter exponential smoothing method was used to forecast electricity demand, and the experiment results concluded that it outperformed well-fitted ARIMA models. Reference [11] presents the application of a Holt–Winters method to predict the nonresidential consumption of Romania, and the average prediction deviation was less than 5%. Tratar and Strmcnik applied Holt–Winters methods for heat load forecasting, which were compared to a multiple regression method [21]. Due to these successful applications of methods for forecasting electric energy, in this study, SES models including the popular Holt–Winter additive model were employed to forecast medium-term electricity consumption where the horizon ranged from a month to a year.
The SES methods are simple and accurate forecasting models, but their performance depends to a great extent on the values of smoothing parameters. Some studies have attempted to find the values of smoothing parameters, such as maximum likelihood estimates [22,23,24], trial-and-error procedures [25], and grid search [26]. However, these parameter optimization methods have certain limitations; for maximum likelihood estimation, its application premise is that each sample in the sample set is an independent and identically distributed random variable, which is usually difficult to achieve in the real world. The disadvantage of the trial-and-error method is that it cannot ensure that the optimal or near-optimal parameters will be found. Although the grid search approach, as a simple and effective parameter optimization method, is widely used, it should be noted that its computational complexity increases exponentially with the number of parameters and the width of the grid. Therefore, an effective and efficient smoothing parameter optimization approach that is able to find the global optimal or near-optimal solution with the fastest speed is very much needed. Motivated by the fact that the PSO algorithm has been widely used for parameter tuning of artificial intelligence techniques in energy forecasting [27,28,29,30], and the advantages of its simple concept, easy implementation, and quick convergence [31,32], this present study proposes a model integrating the PSO algorithm with SES models for accurate medium-term electricity consumption forecasting.
The contributions of this study include the following three points. Firstly, we propose three PSO-based SES models for electricity consumption forecasting. While there are numerous studies on PSO algorithms and SES models individually, limited work, if any, has investigated the hybridization of both methods in the literature of electricity consumption forecasting. Secondly, to confirm the benefit of PSO algorithm for parameter optimization in SES models, genetic algorithm (GA) and grid search (GS) were used to compare with PSO based on the prediction accuracy and computational cost; the experimental results indicated that whether the accuracy measure or the elapsed time was considered, the PSO performed better than GS and GA. Thirdly, compared with other well-established counterparts, the benefits of the proposed PSO-based SES model with its non-trend component and additive seasonality were verified using two real-world electricity consumption datasets on the basis of prediction accuracy.
This paper is organized as follows. In Section 2, brief introductions to SES models and the PSO algorithm are provided in Section 2.1 and Section 2.2. In Section 3, details of the proposed PSO-based SES approach are presented. Section 4 shows the details of the experimental setup. Following that, the results and a discussion are reported in Section 5. Section 6 finally concludes this work.
2. Methodologies
In this section, SES models and PSO algorithm are briefly introduced in Section 2.1 and Section 2.2 respectively.
2.1. Seasonal Exponential Smoothing Models
The SES models are classified as additive or multiplicative seasonal models. In this section, we introduce the additive SES models, which can be sorted into three major categories according to trend component: non-trend, additive trend, and multiplicative trend, while detailed information about multiplicative seasonal models can be found in [33].
(i) The SES model without a trend term but with the additive seasonal term, which is abbreviated as NA for convenience of description, is presented using the following expressions.
(1)
(ii) The SES model with an additive trend term and additive seasonal term, which is abbreviated as AA for convenience of description, is presented using the following expressions.
(2)
(iii) The SES model with a multiplicative trend term and additive seasonal term, which is abbreviated as MA for convenience of description, is given using the following expressions.
(3)
where are hyperparameters; represent the level term, trend term, and seasonality term at period , respectively; is the forecast value for steps ahead based on all of the data up to time ; and , is the length of a cycle.In this paper, the initial smooth values for the level and trend components were estimated by averaging the first two full cycles of electricity consumption data, and the specific expressions were and . The initial seasonal factors were determined using the expressions .
In the following Section 2.2, the PSO algorithm is described, in order to enable understanding of the PSO-based SES strategy proposed in Section 3.
2.2. PSO Algorithm
In PSO, and denote the position vector and the velocity vector of the i-th particle, where is number of parameter in the SES model. The best positions visited by the i-th particle and the entire particle swarm are denoted as and , respectively. Specially, the velocity and position of the d-th dimension of the i-th particle are updated according to Formulas (4) and (5), respectively:
(4)
(5)
(6)
where and are the acceleration constants; and are two real random numbers uniformly distributed between 0 and 1; is an inertia weight, and are initial and final weight coefficients, respectively; is the maximum number of iterations; and is the current iteration number. Considering that a larger inertia weight is good for global exploration while a smaller inertia weight is conducive to fine-tuning in the current search area, PSO can thus make a good balance between global and local searching in the entire search process by adjusting along with the iteration step [34].Additionally, the velocities and positions of all the particles are constrained to the interval and by the relationships and . The flowchart of the PSO algorithm is given in Figure 1.
3. Proposed PSO-Based SES Modeling Strategy
In this section, the proposed PSO-based SES modeling framework is formulated and corresponding steps involved are presented in detail, and the specific flowchart is shown in Figure 2.
As shown in Figure 2, the detailed PSO-based SES modeling strategy includes four major operations: initialization, evaluation, update, and prediction. The overall learning process in Figure 2 is elaborated step by step below.
-
Initialization
In PSO, the initial positions of all the particles are randomly distributed across a designed D-dimensional space and each dimension corresponds to a specified smoothing parameter. The initial velocity of each particle in the swarm is randomly assigned a real value which is equal to the product of a real random number uniformly distributed between 0 and 1 and the maximum velocity ; here, was usually set at 10–20% of the dynamic range of the variable in each dimension [35].
As mentioned in Section 2.1, the estimation of the trend component at initial time involves the first two full cycles of electricity consumption data; we let , then all of the data up to time were used to calculate the initial level component , the initial trend component , and the initial season factors .
-
2.. Evaluation
In order to elaborate the computational process of the initial fitness value of each particle, we take the i-th particle and steps ahead forecast as an example, and assume that the data on the training set range from to . Firstly, the position of the i-th particle is transformed to the corresponding smoothing parameter of the SES models. Secondly, the corresponding and the prediction value at each period are calculated using the equations of the SES models and the corresponding forecasting formula, respectively. Thirdly, the fitness value is evaluated as defined in Equation (8). Finally, the current positions of all the particles are used to initialize their , and the best position visited by the entire swarm is used to initialize the .
-
3.. Update
The position and velocity of each particle are updated using Equations (4) and (5), respectively, and their fitness values are recalculated as per Step 2. Once the fitness values of all the particles are obtained, and are updated. A judgement is made as to whether the termination condition is satisfied and, if so, the best particle with the near-optimal smoothing parameters is returned; otherwise, the previous step is repeated.
-
4.. Prediction
The forecasting values on the test set are evaluated with the near-optimal smoothing parameters. Assuming that the data in the test set range from to , then the length of the test set is . Consistent with the prediction operation of the training set, the test set still takes periods ahead to forecast. The corresponding and the prediction value at each period are then calculated using the equations of the SES models and the corresponding forecasting formula, respectively.
4. Experimental Setup
4.1. Data Description
To examine the performances of the proposed PSO-based SES methods in terms of their forecast accuracy, two real-world datasets, i.e., monthly electricity consumption data from United States and China, were used in the present study. The period of the first dataset ranged from January 2002 to September 2019 and that of the second dataset was from January 2010 to December 2015.
As there are 12 months in a full period, the estimation of the trend component at initial time involved the first two full cycles of historical electricity consumption data. Therefore, the first dataset was divided into the calculation set (the first two years) used to evaluate the initialized relevant parameters in the SES models, the training set (January 2004 to December 2015), and the testing set (the last 45 months). Analogously, the second dataset was divided into the calculation set (January 2010 to December 2011), the training set (January 2012 to December 2014), and the testing set (the last 12 months in 2015). Here, the testing sets were completely independent of the training sets and were not involved in the learning procedure. Figure 3 depicts the two datasets.
4.2. Experimental Design
There were two goals in the experimental study. The first goal was to verify the benefits of the PSO algorithm for parameter optimization in SES models. To achieve the goal, the Holt–Winters additive model, as a popular seasonal exponential smoothing model, was employed to forecast the electricity consumption in the above two datasets, and the commonly used GS and GA were also used to optimize the parameters of the Holt–Winters additive model for comparison with PSO. The second goal was to examine the performance of the proposed PSO-based SES models and to determine which of the three different additive SES models was the most effective in terms of accuracy. In this regard, seasonal ARIMA (SARIMA) from time series forecasting models as well as SVR and a back propagation neural network model (BPNN) based on machine learning techniques were used for comparative analysis with the PSO-based SES models.
4.3. Performance Measure
To assess the forecasting performance of a model, three alternative forecasting accuracy measures were employed in this study, namely the mean absolute percentage error (MAPE) [19], the root mean square error (RMSE) [36], and the normalized root mean square error (NRMSE) [37]. Definitions of these measures are presented in the following expressions (7)–(9).
(7)
(8)
(9)
where is the sample number in the test set, is the actual value at period , and is the prediction value at period . The smaller the three measure values are, the better the prediction performance of the model.5. Results and Discussion
5.1. Study 1: Examination of the PSO Algorithm for Parameter Optimization in SES Models
To confirm the benefit of the PSO algorithm for parameter selection in SES models, in this paper, the GS method from the classical domain and the genetic algorithm (GA) from the class of evolutionary computation were used for comparative analysis, and the quantitative and comprehensive assessments were performed with two real-world electricity consumption datasets on the basis of the prediction accuracy and computational cost for GS, GA, and PSO. The experiments were implemented in MATLAB 2016a using in-house software executed on a computer with Intel Core i5–3230M, 2.60 GHz CPU, 4 GB RAM. In general, the experimental facility did not require high computer configuration; was easy to operate, simple, and efficient; and produced repeatable experiments.
In these experiments, the MATLAB codes for GS and PSO were written in source code, while the implementation of GA used the GA toolbox that comes with MATLAB 2016a. It should be noted that the selection of parameters in the PSO algorithm (i.e., inertial weight, coefficients, swarm size, and number of iterations) is yet another challenging model selection task. The final parameters, chosen according to relative empirical and theoretical studies [32,38] and a trial-and-error approach based on the prediction performance and computational time, are summarized in Table 1. Aimed at the random of PSO algorithm, the modeling process for each prediction horizon was repeated 10 times and performance of the PSO-based additive Holt–Winters model was judged by the mean of each accuracy measure.
Table 2 shows the prediction performance of the additive Holt–Winters model with GS, PSO, and GA in terms of three accuracy measures (i.e., MAPE, RMSE, NRMSE) for the U.S. and China datasets. The columns labeled “Average 1—h” show the accuracy measures over the prediction horizon 1 to h. The last column shows the average ranking for each method over all forecast horizons of the out-of-sample prediction performance.
As per the results presented in Table 2, it was not difficult to make the following observation: when considering the U.S. dataset, GA performed the worst across three measures and the computation time was the longest. GS(0.05) and PSO had the same ranking in the terms of RMSE, and the average RMSE measures of GS(0.05) and PSO over all forecast horizons were 8.333 and 8.339, respectively. The ranking with respect to MAPE was GS(0.05) then PSO, while the ranking with respect to NRMSE was PSO then GS(0.05). Additionally, the elapsed time of GS(0.05) for a single replicate was more than twice that of PSO. When considering the China dataset, GS(0.05) performed the worst and GA performed the second worst across three measures according to the average rank. The ranking with respect to MAPE was GS(0.01) then PSO, while the ranking with respect to RMSE and NRMSE was PSO then GS(0.01). In terms of the elapsed time, the GA was computationally much more expensive than the other methods. The GS(0.05) was tens of times faster than GS(0.01), and PSO is hundreds of times faster than GS(0.01).
In conclusion, the PSO was significantly superior to GA, GS(0.01), and GS(0.05) in terms of the elapsed time. PSO outperformed GA across three measures for the majority of prediction horizons in the U.S. and China datasets. PSO was consistently better than GS(0.05) across three measures at all forecast horizons for the China dataset. Moreover, there was no significant difference between PSO and GS(0.01) for the China dataset. Thus, it is conceivable that PSO algorithm is suitable for parameter optimization in SES models.
5.2. Study 2: Comparison with Well-Known Forecasting Models
Figure 4 and Figure 5, respectively, show the comparison of actual values and predicted values for all the examined models (i.e., NA-PSO, AA-PSO, MA-PSO, SARIMA, SVR, and BPNN) in the U.S. and China datasets. And the subgraphs (a) and (b) in these two figures show the predicted values of all the examined models at the forecast horizons of 1 and 6, respectively. The prediction performances of the above models in terms of two accuracy measures (i.e., MAPE and NRMSE) are shown in Table 3. Here, “NA” denotes the SES model without a trend term but with an additive seasonal term, “AA” denotes the SES model with an additive trend term and additive seasonal term, and “MA” denotes the SES model with a multiplicative trend term and an additive seasonal term, and PSO was used to optimize the smoothing parameters of these SES models.
The experiments of the SARIMA model were implemented in Eviews 8. The specific process was as follows:
Stationarity test: a seasonal difference and a nonseasonal difference were performed for the original series. The augmented Dickey–Fuller test result showed that the difference series was stationary, so and ;
model identification: the correlation coefficient graph of the difference sequence was analyzed and then the value range of ,,, and was determined;
model selection: the trial-and-error method based on BIC was used to determine the order of the model. In this paper, the models established for the two electricity consumption datasets in U.S. and China were and , respectively, and all the estimated parameters in these two models passed the significance test;
Residual error test: the series correlation LM test was performed on the fitted residual sequences of the above two models. The results showed that the two residual sequences were not correlated, which indicates that the models we established were credible;
Prediction: Eviews 8 provided a dynamic forecast function for SARIMA’s multi-step-ahead prediction. Here, we obtained the predicted values of each sample in the test set from one step ahead to twelve steps ahead by changing the range of prediction sample, and then rearranged these prediction values according to the forecast horizon and further calculated the prediction accuracy of SARIMA at each forecast horizon.
The experiments of SVR and BPNN models were implemented using LibSVM toolbox (version 2.86) and BPNN toolbox in MATLAB 2016a. The parameters of SVR took default values and the embedding dimensions of SVR and BPNN were determined to be 6 and 12 by trial and error.
According to the results presented in Figure 4 and Table 3, we can make several observations about the U.S. dataset:
Among all the examined models, the two models with the worst prediction results were MA-PSO and BPNN, which is obvious in Figure 4b.
According to the average rank in terms of MAPE measure, the top three models turned out to be NA-PSO, then AA-PSO, and then SARIMA. The rankings with respect to NRMSE measure were SARIMA, then NA-PSO, and then AA-PSO. However, the average NRMSE measure of NA-PSO over all forecast horizons was less than that of SARIMA, which indicates that NA-PSO outperformed SARIMA to a certain extent.
The proposed NA-PSO and AA-PSO consistently achieved more accurate forecasts than the MA-PSO regardless of the accuracy measures and forecast horizon considered. A possible reason is that multiplication form was not suitable for fitting the trend term of current electricity consumption data.
As far as the comparison between the NA-PSO, AA-PSO, and SARIMA, we can see that the results were mixed among the horizons examined; there was an interesting phenomenon observed that whatever the accuracy measures considered, SARIMA won for one-step-ahead and two-step-ahead predictions. The main reason could be that the principle of SARIMA for multi-step-ahead prediction is recursive forecasting, which will cause errors to accumulate as the forecast horizon increases.
From Figure 5 and Table 3, several observations about the China dataset can be made.
Among all the examined models, the two models with the worst prediction results were SVR and BPNN, which is obvious in Figure 5a,b. The possible reason is that the methods based on machine learning were suitable for sufficient training samples, while there were only 72 samples in China dataset.
According to the MAPE, the top three models turned out to be SARIMA, then NA-PSO, and then AA-PSO. The rankings with respect to the NRMSE measure were NA-PSO, and then AA-PSO and SARIMA tied for second. It should be noted that the reason why MA-PSO is not presented in Table 3 and Figure 5 for the China dataset is the MA-PSO performed the worst regardless of the accuracy measures and forecast horizon considered.
As far as the comparison between the NA-PSO, AA-PSO, and SARIMA, we saw that the results were mixed among the horizons examined, but SARIMA won for one-step-ahead prediction regardless of the accuracy measures considered, which indicates that SARIMA is suitable for short-term forecasting.
Overall, it is clear that the proposed NA-PSO held one of the top positions. The main reason could be that the electricity consumption associated with people’s daily activities and various economic activities has basically reached a saturated level, which is more intuitively reflected in Figure 3. It is not difficult to see that growth trends of electricity consumption in the U.S. and China were no longer obvious after 2005 and 2012, respectively.
In order to determine whether there existed a statistically significant difference among the five models in the hold-out sample, ANOVA procedures are performed for each performance measure, prediction horizon, and dataset. All ANOVA results were significant at the 0.05 level, which suggests that there were significant differences among the five models. The results are omitted here to save space. Tukey’s HSD test was used to further identify the significant difference between any two models. The results of these multiple comparison tests for the U.S. and China datasets are shown in Table 4. For each accuracy measure, prediction horizon, and dataset, we have ranked the models in order from 1 (the best) to 5 (the worst).
As per the results shown in Table 4, it is not difficult to see that the three models (i.e., NA-PSO, AA-PSO, and SARIMA) significantly outperformed SVR and BPNN in most cases, and SVR significantly outperformed BPNN in almost all cases.
When considering the U.S. dataset, the SARIMA significantly outperformed AA-PSO and NA-PSO in the cases of one-step-ahead prediction across two measures and horizon 2 for the NRMSE measure. The NA-PSO outperformed SARIMA for the overwhelming majority of prediction horizons. As far as the comparison of NA-PSO vs. AA-PSO, the NA-PSO performed significantly better than AA-PSO in most cases.
When considering the China dataset, the SARIMA significantly outperformed AA-PSO and NA-PSO in the cases of one-step-ahead prediction for MAPE measures and nine-step-ahead prediction across two measures. The AA-PSO significantly outperformed SARIMA and NA-PSO in the cases of twelve-step-ahead prediction across two measures. In addition, the NA-PSO significantly outperformed SARIMA and AA-PSO for the majority of prediction horizons.
6. Conclusions
Electricity consumption forecasting is an important issue in investment planning of electricity infrastructure, and in electricity production/generation and distribution. The purpose of this study was to perform accurate monthly electricity consumption forecasting using the proposed PSO-based SES models. The results show that PSO was significantly superior to GA and GS in terms of the elapsed time, PSO outperformed GA for the majority of predictions, and PSO and GA are almost evenly matched across three accuracy measures. Justified with the real-world electricity consumption datasets, the proposed NA-PSO and AA-PSO consistently achieved more accurate forecasts than the MA-PSO regardless of the accuracy measures and forecast horizon considered, and the three models (i.e., NA-PSO, AA-PSO, and SARIMA) significantly outperformed SVR and BPNN in most cases, and the proposed NA-PSO performed best among the three models (i.e., NA-PSO, AA-PSO, and SARIMA), which indicates that the NA-PSO could be a promised alternative for electricity consumption forecasting.
Author Contributions
Conceptualization, Y.B.; methodology, X.Z.; software, C.D. and Y.H.; validation, X.Z. and Y.H.; formal analysis, C.D.; investigation, X.Z. and C.D.; resources, Y.B.; data curation, Y.H.; writing—original draft preparation, Y.B. and X.Z.; writing—review and editing, Y.B. and Y.H.; supervision, Y.B.; funding acquisition, Y.B. All authors have read and agreed to the published version of the manuscript.
Funding
This work was funded by Humanities and Social Sciences Projects of Jiangxi under project no.GL19115, Science and Technology Research Projects of Jiangxi under project no. GJJ191188, Key R&D Projects of Jiangxi under project no.20192BBHL80015, Jiangxi Principal Academic and Technical Leaders Program under project no.20194BCJ22015, Science and Technology Research Projects of Jiangxi under project no. GJJ202905 and Natural Science Foundation of China under Project Nos. 71571080 and 71871101.
Data Availability Statement
The first data can be obtained from the website of the United States Energy Administration (
Conflicts of Interest
The authors declare no conflict of interest.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Figures and Tables
Figure 4. The forecasted values of all the examined models at forecast horizons of 1 and 6 in the U.S. dataset.
Figure 5. The forecasted values of all the examined models at forecast horizons of 1 and 6 in the China dataset.
Parameter selection in PSO.
Parameters | Values | Parameters | Values |
---|---|---|---|
Population size (ps) | 30 | Interaction coefficient (c2) | 2.0 |
Number of iterations (T) | 200 | Initial inertial weight (w1) | 0.9 |
Cognitive coefficient (c1) | 2.0 | Final inertial weight (w2) | 0.4 |
Prediction accuracy measures and elapsed time of GS, PSO, and GA for the U.S. and China datasets.
Dataset | Strategy | Forecast Horizon (h) | Average 1—h | Average Rank | ||||||
---|---|---|---|---|---|---|---|---|---|---|
1 | 2 | 3 | 6 | 9 | 12 | 1–6 | 1–12 | |||
U.S. | MAPE | |||||||||
GS(0.05) | 1.993 | 2.019 | 2.077 | 1.835 | 1.799 | 1.799 | 1.954 | 1.915 | 1.625 | |
PSO | 1.992 | 2.065 | 2.061 | 1.854 | 1.790 | 1.901 | 1.967 | 1.918 | 1.875 | |
GA | 2.001 | 2.075 | 2.078 | 1.769 | 1.787 | 2.008 | 1.973 | 1.962 | 2.500 | |
RMSE | ||||||||||
GS(0.05) | 8.257 | 8.507 | 8.763 | 8.270 | 7.720 | 8.599 | 8.432 | 8.333 | 1.750 | |
PSO | 8.256 | 8.595 | 8.741 | 8.269 | 7.766 | 8.667 | 8.431 | 8.339 | 1.750 | |
GA | 8.307 | 8.726 | 8.724 | 8.203 | 7.813 | 9.157 | 8.536 | 8.577 | 2.500 | |
NRMSE | ||||||||||
GS(0.05) | 0.02530 | 0.02608 | 0.02683 | 0.02511 | 0.02475 | 0.02641 | 0.02592 | 0.02567 | 2.250 | |
PSO | 0.02529 | 0.02561 | 0.02653 | 0.02502 | 0.02462 | 0.02672 | 0.02573 | 0.02552 | 1.375 | |
GA | 0.02545 | 0.02675 | 0.02671 | 0.02491 | 0.02399 | 0.02803 | 0.02607 | 0.02622 | 2.375 | |
Elapsed Time | ||||||||||
GS(0.05) | 0.596 | 0.481 | 0.452 | 0.428 | 0.453 | 0.47 | 0.478 | 0.468 | 2.000 | |
PSO | 0.187 | 0.188 | 0.181 | 0.182 | 0.179 | 0.182 | 0.184 | 0.183 | 1.000 | |
GA | 293.13 | 540.22 | 663.69 | 1041.09 | 361.03 | 723.59 | 707.96 | 591.15 | 3.000 | |
China | MAPE | |||||||||
GS(0.05) | 4.030 | 4.416 | 3.504 | 3.442 | 4.678 | 2.813 | 3.660 | 3.802 | 3.750 | |
GS(0.01) | 3.802 | 3.918 | 3.326 | 2.690 | 3.694 | 2.075 | 3.192 | 3.090 | 1.375 | |
PSO | 3.819 | 3.961 | 3.337 | 2.671 | 3.693 | 2.108 | 3.225 | 3.100 | 1.875 | |
GA | 3.817 | 4.067 | 3.317 | 5.266 | 4.449 | 2.513 | 3.896 | 3.811 | 3.000 | |
RMSE | ||||||||||
GS(0.05) | 215.229 | 215.850 | 184.854 | 185.216 | 237.467 | 140.873 | 191.187 | 195.588 | 3.500 | |
GS(0.01) | 209.073 | 208.990 | 176.490 | 150.861 | 190.631 | 105.548 | 173.594 | 164.099 | 1.875 | |
PSO | 209.082 | 207.852 | 175.738 | 151.670 | 191.163 | 103.904 | 172.216 | 163.795 | 1.625 | |
GA | 208.587 | 215.862 | 175.248 | 262.722 | 224.159 | 125.866 | 203.947 | 199.078 | 3.000 | |
NRMSE | ||||||||||
GS(0.05) | 0.04655 | 0.04690 | 0.03930 | 0.03858 | 0.05074 | 0.02813 | 0.04069 | 0.04124 | 3.500 | |
GS(0.01) | 0.04522 | 0.04541 | 0.03754 | 0.03142 | 0.04076 | 0.02114 | 0.03697 | 0.03578 | 2.000 | |
PSO | 0.04521 | 0.04516 | 0.03736 | 0.03159 | 0.04084 | 0.02075 | 0.03669 | 0.03458 | 1.500 | |
GA | 0.04511 | 0.04690 | 0.03726 | 0.05473 | 0.04790 | 0.02513 | 0.04334 | 0.04150 | 2.875 | |
Elapsed Time | ||||||||||
GS(0.05) | 0.257 | 0.278 | 0.253 | 0.297 | 0.278 | 0.333 | 0.265 | 0.274 | 2.000 | |
GS(0.01) | 12.45 | 12.397 | 12.585 | 12.451 | 12.293 | 12.327 | 12.466 | 12.414 | 3.000 | |
PSO | 0.0938 | 0.1007 | 0.093 | 0.0976 | 0.0951 | 0.1001 | 0.0970 | 0.0980 | 1.000 | |
GA | 446.89 | 677.99 | 894.85 | 561.53 | 410.7 | 645.47 | 690.710 | 577.91 | 4.000 |
Note: For each column of table, the entry with smallest value is set in boldface type. ‘GS(0.05)’ corresponds to the GS with the grid width of 0.05.
Table 3Prediction accuracy measures for the U.S. and China datasets.
Dataset | Strategy | Forecast horizon (h) | Average 1—h | AverageRank | ||||||
---|---|---|---|---|---|---|---|---|---|---|
1 | 2 | 3 | 6 | 9 | 12 | 1–6 | 1–12 | |||
U.S. | MAPE | |||||||||
NA-PSO | 2.005 | 2.047 | 2.057 | 1.921 | 1.679 | 1.712 | 1.988 | 1.866 | 1.625 | |
AA-PSO | 1.992 | 2.065 | 2.061 | 1.854 | 1.790 | 1.901 | 1.973 | 1.922 | 2.000 | |
MA-PSO | 3.226 | 5.692 | 7.362 | 10.17 | 5.235 | 2.309 | 7.585 | 6.200 | 5.750 | |
SARIMA | 1.767 | 2.010 | 2.149 | 2.035 | 1.912 | 1.843 | 2.010 | 1.975 | 2.375 | |
SVR | 2.236 | 2.912 | 2.751 | 2.116 | 1.930 | 1.989 | 2.359 | 2.167 | 4.000 | |
BPNN | 4.116 | 4.184 | 4.213 | 5.205 | 3.901 | 4.568 | 4.487 | 4.530 | 5.250 | |
NRMSE | ||||||||||
NA-PSO | 0.02534 | 0.02639 | 0.02639 | 0.02530 | 0.02260 | 0.02292 | 0.02579 | 0.02459 | 1.875 | |
AA-PSO | 0.02529 | 0.02561 | 0.02653 | 0.02502 | 0.02462 | 0.02672 | 0.02573 | 0.02552 | 2.500 | |
MA-PSO | 0.04189 | 0.06890 | 0.09325 | 0.12613 | 0.06467 | 0.03068 | 0.09471 | 0.08150 | 5.750 | |
SARIMA | 0.02261 | 0.02513 | 0.02645 | 0.02647 | 0.02420 | 0.02414 | 0.02571 | 0.02546 | 1.750 | |
SVR | 0.03055 | 0.03754 | 0.03439 | 0.02845 | 0.02733 | 0.02632 | 0.03117 | 0.02922 | 3.875 | |
BPNN | 0.05385 | 0.05430 | 0.05360 | 0.07244 | 0.06309 | 0.05897 | 0.06060 | 0.05991 | 5.250 | |
China | MAPE | |||||||||
NA-PSO | 3.255 | 2.635 | 2.085 | 2.726 | 3.519 | 3.933 | 2.374 | 2.540 | 1.875 | |
AA-PSO | 3.819 | 3.961 | 3.337 | 2.671 | 3.693 | 2.108 | 3.225 | 3.100 | 2.625 | |
SARIMA | 3.054 | 2.761 | 2.726 | 2.684 | 2.772 | 2.622 | 2.786 | 2.747 | 1.750 | |
SVR | 5.189 | 5.080 | 4.856 | 5.336 | 3.644 | 3.790 | 5.158 | 4.518 | 3.875 | |
BPNN | 5.026 | 7.835 | 5.821 | 6.865 | 8.882 | 7.468 | 5.997 | 6.390 | 4.875 | |
NRMSE | ||||||||||
NA-PSO | 0.04373 | 0.03255 | 0.02527 | 0.03146 | 0.04040 | 0.03933 | 0.03001 | 0.03034 | 1.500 | |
AA-PSO | 0.04521 | 0.04516 | 0.03736 | 0.03159 | 0.04084 | 0.02075 | 0.03669 | 0.03458 | 2.250 | |
SARIMA | 0.04338 | 0.03928 | 0.03785 | 0.04116 | 0.03956 | 0.03781 | 0.03963 | 0.03906 | 2.250 | |
SVR | 0.07696 | 0.07060 | 0.06424 | 0.07119 | 0.05189 | 0.04827 | 0.06891 | 0.06039 | 4.000 | |
BPNN | 0.07712 | 0.10210 | 0.07600 | 0.08085 | 0.09905 | 0.09239 | 0.07996 | 0.08254 | 5.000 |
Note: For each column of table, the entry with smallest value is set in boldface type.
Table 4Multiple comparison results with ranked models for hold-out sample on U.S. and China datasets.
Dataset | Measure | Prediction | Rank of Strategies | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Horizon (h) | 1 | 2 | 3 | 4 | 5 | ||||||
U.S. | MAPE | 1 | SARIMA | <* | AA-PSO | < | NA-PSO | <* | SVR | <* | BPNN |
2 | SARIMA | < | NA-PSO | < | AA-PSO | <* | SVR | <* | BPNN | ||
3 | NA-PSO | < | AA-PSO | <* | SARIMA | <* | SVR | <* | BPNN | ||
6 | AA-PSO | <* | NA-PSO | <* | SARIMA | <* | SVR | <* | BPNN | ||
9 | NA-PSO | <* | AA-PSO | <* | SARIMA | < | SVR | <* | BPNN | ||
12 | NA-PSO | <* | SARIMA | < | AA-PSO | < | SVR | <* | BPNN | ||
NRMSE | 1,2 | SARIMA | <* | AA-PSO | < | NA-PSO | <* | SVR | <* | BPNN | |
3 | NA-PSO | < | SARIMA | < | AA-PSO | <* | SVR | <* | BPNN | ||
6 | AA-PSO | < | NA-PSO | <* | SARIMA | <* | SVR | <* | BPNN | ||
9 | NA-PSO | <* | AA-PSO | < | SARIMA | <* | SVR | <* | BPNN | ||
12 | NA-PSO | <* | SARIMA | <* | SVR | < | AA-PSO | <* | BPNN | ||
China | MAPE | 1 | SARIMA | <* | NA-PSO | <* | AA-PSO | < | BPNN | < | SVR |
2,3 | NA-PSO | <* | SARIMA | <* | AA-PSO | <* | SVR | <* | BPNN | ||
6 | AA-PSO | < | SARIMA | < | NA-PSO | <* | SVR | <* | BPNN | ||
9 | SARIMA | <* | NA-PSO | <* | SVR | < | AA-PSO | <* | BPNN | ||
12 | AA-PSO | <* | SARIMA | <* | SVR | < | NA-PSO | <* | BPNN | ||
NRMSE | 1 | SARIMA | < | NA-PSO | <* | AA-PSO | <* | SVR | < | BPNN | |
2 | NA-PSO | <* | SARIMA | <* | AA-PSO | <* | SVR | <* | BPNN | ||
3 | NA-PSO | <* | SARIMA | < | AA-PSO | <* | SVR | <* | BPNN | ||
6 | AA-PSO | < | NA-PSO | <* | SARIMA | <* | SVR | <* | BPNN | ||
9 | SARIMA | <* | NA-PSO | < | AA-PSO | <* | SVR | <* | BPNN | ||
12 | AA-PSO | <* | SARIMA | <* | NA-PSO | <* | SVR | <* | BPNN |
Note: * denotes that the mean difference between the two adjacent methods is significant at the 0.05 level.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2021 by the authors.
Abstract
Electricity consumption forecasting plays an important role in investment planning of electricity infrastructure, and in electricity production/generation and distribution. Accurate electricity consumption prediction over the mid/long term is of great interest to both practitioners and academics. Considering that monthly electricity consumption series usually show an obvious seasonal variation due to their inherent nature subject to temperature during the year, in this paper, seasonal exponential smoothing (SES) models were employed as the modeling technique, and the particle swarm optimization (PSO) algorithm was applied to find a set of near-optimal smoothing parameters. Quantitative and comprehensive assessments were performed with two real-world electricity consumption datasets on the basis of prediction accuracy and computational cost. The experimental results indicated that (1) whether the accuracy measure or the elapsed time was considered, the PSO performed better than grid search (GS) or genetic algorithm (GA); (2) the proposed PSO-based SES model with a non-trend component and additive seasonality term significantly outperformed other competitors for the majority of prediction horizons, which indicates that the model could be a promising alternative for electricity consumption forecasting.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details

1 Center of Big Data Analytics, Jiangxi University of Engineering, Xinyu 338029, China;
2 Center for Modern Information Management, School of Management, Huazhong University of Science and Technology, Wuhan 430074, China