1. Introduction
In the last decade, wind and solar energy production, and more specifically photovoltaic (PV) energy, has increased significantly and, therefore, a large PV penetration with a rapid growth has taken place in the electricity market [1]. To achieve this high penetration, it is important to have accurate point forecasts and most of the research has focused on this issue. However, due to the high variability of several meteorological factors, solar power prediction is inherently uncertain and, therefore, it is also important to estimate the uncertainty around point forecasts. Accurate knowledge of uncertainty on renewable energy production can provide both economic opportunities [2] and reliability improvement [3]. Recent literature shows how research into new uncertainty estimation methods, as well as their applications to the renewable energy industry, is a field of great activity [4,5,6].
The most commonly used ways of representing uncertainty are quantiles and prediction intervals (PIs). The estimation of quantiles has been widely addressed in the literature, especially through methods that estimate each desired quantile independently. For example, quantile regression (QR ) [7] constructs linear models by minimizing quantile loss (or pinball score) for a particular quantile-probability. Nonlinear models can also be constructed using this cost function, for example quantile regression neural networks (QRNN ) [8] or tree-based ensembles such as Gradient Tree Boosting (GB ) [9,10]. There are also techniques which build a single model from which all quantiles can be extracted. For instance, quantile regression forests (QRF ) [11] is another non-linear ensemble technique and analog ensembles (AnEn) [12,13,14] generate probability distributions using a set of past measurements that correspond to the most similar past numerical weather prediction (NWP) forecasts to the current situation.
Another way of representing forecast uncertainty is by means of PIs, where two bounds, lower and upper, contain the actual value with a given probability (called prediction interval probability coverage (PICP)) [15]. For a given PICP, the narrower is the PI, the less uncertain is the forecast. There are in the literature many classic methods for computing PIs [16], but, more recently, an evolutionary approach, known as LUBE (lower upper bound estimation), has shown better performance in several domains [16], including renewable energy forecasting [17,18]. The approach uses a two-output artificial neural networks, for the lower and upper bounds of the interval, respectively. For point forecasting, the actual output to the networks is available, and, thus, supervised training techniques, such as backpropagation can be used. However, this is not the case for LUBE, because the two outputs of the network (i.e., the lower and upper bounds) are not directly available. For this reason, the network parameters are usually optimized using evolutionary computation techniques such as simulated annealing (SA) [19] or particle swarm optimization (PSO) [20]. Optimizing PIs is clearly a multi-objective problem, because obtaining high-coverage but narrow intervals are conflicting goals. However, most LUBE approaches translate the multi-objective problem into a single-objective one, by aggregating the two goals, but this requires some reasonable weighting between the objectives to be decided in advance. LUBE can also be addressed as an actual multi-objective approach that does not require goal weighting. This was the approach proposed by the authors in a previous work [21], where a multi-objective particle swarm optimization evolutionary algorithm (MOPSO) showed very good performance in a solar energy production forecasting problem on Oklahoma solar sites [22], where only meteorological forecasts were used as input. An important advantage of using a multi-objective algorithm is that in a single run the method is able to return a whole set of solutions (Pareto front), which represent the best trade-offs between coverage and width, out of which the user can select a solution for some particular coverage.
Most works that deal with energy prediction use meteorological forecast variables as inputs for the prediction models. However, some articles indicate that prediction error can be improved by using actual measurements available at prediction time, in addition to meteorological forecasts [23,24]. For instance, a solar plant that, at timet0, needs to issue a power output forecast for the next hour (t0+1h) can apply a model that uses (as inputs) meteorological forecasts fort0+1has well as the current power output (measured att0 ). Martín-Vázquez et al. [25] showed that, in fact, the use of measured output wind electrical power was helpful to improve point forecasts for short horizons in wind energy forecasting. However, not much research has been done to study the influence of using measurements for constructing PIs, although preliminary results point in that direction [26].
The purpose of this article is to study whether the quality of PIs for PV power forecast can be improved by using measured solar output power in addition to meteorological forecast variables. Given that it could be expected that measurements att0may influence short horizons more strongly than longer ones, in this work, several forecast horizons are considered. In principle, using measured information at prediction time should reduce the uncertainty of forecasts (and, therefore, PIs should be narrower), but this advantage is expected to decrease for long prediction horizons. Thus, the main contributions of this article are to study both the degree of improvement of PIs when using measured information and the dependency of that improvement on the prediction horizon. We used the LUBE approach optimized withMOPSO [21]. TheMOPSO approach has already been used for estimating PIs for the aggregated daily energy production forecast for next day [21]. However, in the context of solar forecasting, it is also important to provide uncertainty forecasts for intra-day horizons, which is what we do in the current work. We also used for comparison purposes two benchmark methods recommended in recent literature for probabilistic solar forecasting [4]. The first isQR [7], a commonly used linear method for obtaining quantiles in solar forecasting [27]. The second one isQRF [11], a non-linear method successfully used in recent works on solar energy uncertainty estimation [28,29]. All methods were tested using as inputs only the meteorological variables and using the meteorological variables together with measured solar power att0.
The study was done using data from the Global Energy Forecasting Competition 2014 (GEFCom2014) concerning the probabilistic solar power forecasting problem [30]. To advance beyond the preliminary results [26], in this study, extensive experimentation was carried out. Data from three different solar plants in Australia (Stations 1–3) were used and the PIs were estimated for five intra-day forecasting horizons (1–5 h, UTC time). Given thatMOPSOandQRFare stochastic techniques, they were run 90 times and statistical tests for comparison were done by means of the Wilcoxon test. The extensiveness of the experiments and the statistical test allowed extracting reliable conclusions.
The rest of the article is organized as follows. Section 2 describes the dataset used for experiments. Section 3 summarizes the evolutionary multi-objective approach for interval optimization. Section 4 describes the two baseline methods for comparison. Section 5 describes the experimental setup and the results, including statistical significance tests. Conclusions are finally drawn in Section 6.
2. Data Description
The data used in this work were obtained from Task 15 in the probabilistic solar power forecasting problem [30] one of the four tracks from the Global Energy Forecasting Competition 2014 (GEFCom2014). These data include measured solar power generation and meteorological forecasts for three solar stations (Stations 1–3) in Australia whose exact location was been revealed. Measured solar power generation, expressed in proportions between 0 and 1 relative to the nominal power of the plant, was provided hourly, from 1 April 2012 01:00 to 1 July 2014 00:00 UTC. The meteorological forecasts were obtained from the European Centre for Medium-range Weather Forecasts (ECMWF) and include 12 weather variables. Forecasts for those weather variables are issued by ECMWF once per day, at midnight, and they are forecasted for the each hour for the next 24 h. Thus, ift0is the time at which forecasts are issued (00:00 UTC), each meteorological variable gets a forecast fort0+1,t0+2, …,t0+24.
The 12 weather variables are: total column liquid water, i.e., the vertical integral of cloud liquid water content (kg · m−2); total column ice water, i.e., the vertical integral of cloud ice water content (kg · m−2); surface pressure (Pa); relative humidity with respect to saturation at 1000 mbar; total cloud cover (0–1); 10-m U wind component (m · s−1); 10-m V wind component (m · s−1); 2-m temperature (K); surface solar radiation downward (J · m−2); surface thermal radiation downward (J · m−2); top net solar radiation, i.e., net solar radiation at the top of the atmosphere (J · m−2); and total precipitation (convective precipitation + stratiform precipitation) (m).
Figure 1 displays a box plot of the normalized measured power generation for each station corresponding to the whole period (2012-04-01 01:00 to 2014-07-01 00:00) and broken down by UTC hour. It is a standard box plot where the lower and upper hinges correspond to the first and third quartiles and the center black horizontal bar is the median. Whiskers are used to denote data in the 1.5*inter-quartile range and black points are used for outliers. It is important to remark that the hour on the x-axis corresponds to EMCWF time. Given that GEFCom2014 did not disclose the actual location of the three solar stations, it is not possible to know with certainty the local time for each of the stations. It is estimated that noon must correspond to the maximum of solar radiation (Time 1–2 for Station 1; Time 3 for Station 2; and Time 2 for Station 3). Zero is the hour at which ECMWF forecasts are issued. Based on the box plot in Figure 1, and given that measured power should be expected to have influence mostly for short time horizons, it was decided to study Hours 1–5. This also excludes hours too close to the night period.
Given that, for each horizon, there is a single forecast everyday, the available data are made of 820 instances, corresponding to 820 days. To ensure that both train and test sets are representative, every 30 days, 20 consecutive days were chosen for training and the remaining 10 for testing. Therefore, there were 550 days for training and 270 days for testing. To perform hyper-parameter tuning tasks, such as choosing the best neural network architecture, the best number of optimization iterations or the best configuration forQFR, it is necessary to split the available training data into another two sets: the training set and the validation set (also known as the development (dev) set). In this work, for every 20-day block of the original training partition, the first 14 days were finally used for training, and the remaining six for validation.
3. Multi-Objective Optimization for Prediction Intervals
The following section describes the multi-objective particle swarm optimization algorithm (MOPSO ) applied to PI, originally reported in [21]. This application was inspired by LUBE [16], which used an artificial neural network (ANN) with a single hidden layer to make an estimation of the lower and upper bounds of the PIs. The ANN was optimized using multi-objective evolution techniques.
In this study, the inputs to the ANN can be meteorological forecast variables and measured solar output power. The outputs of the network are the boundaries (lower and upper) forecasted with the network for the given inputs. For each input in the dataset, an observation of power is found. However, the boundaries are never included, therefore a backpropagation optimization is not suitable for this network. As the desired outputs are unknown, this is not a standard supervised regression, and an alternative optimization technique is required.
Therefore, an evolutionary optimization algorithm was chosen to optimize the weights of the network with two primary goals: PI coverage and interval width. Coverage is measured using the prediction interval coverage probability (PICP), which computes the frequency of observations laying within the interval. The PICP can be maximized by also maximizing the width of the PI, producing trivial solutions. To avoid this, a second goal was set to minimize the width of the PI. The following paragraphs formalize the formulation of these goals.
LetM={(Xi,ti)i=0⋯N}be a set of observations, whereXiis a vector with the input variables andtiis the observed output variable. LetPIi=[Lowi,Uppi)be the prediction interval for observationXi([Lowi,Uppi)are the outputs of the ANN). ThePICP goal can be calculated as described in Equation (1) and the average interval width (AIW ) in Equation (2).
PICP=1N∑i=0NχPIi(Xi)
AIW=1N∑i=0N(Uppi−Lowi)
where N is the number of samples,χPIi(Xi)is the indicator function for intervalPIi(it is 1 ifti∈PIi=[Lowi,Uppi)and 0 otherwise), andUppiandLowirepresent the PI boundaries.
OptimizingPICPrequires sacrificing a higherAIW, while a narrowAIW will lose some coverage in return. The goals are opposed and a multi-objective approach was applied [21]. Every particle in theMOPSOrepresents a single solution to the network, i.e., a different configuration of weights. The goals of this algorithm are:1−PICP (Equation (1)) andAIW (Equation (2)). This network receives the meteorological data (from the given dataset in Section 2) as inputs and measured solar power att0.
TheMOPSOalgorithm produces a set of multiple ANNs laid in a Pareto front of non-dominated solutions. Typically, PIs with a desired coverage, called PIs’ nominal coverage (orPINC), are needed by the user. Sometimes, it is more appropriate to refer to1−PINC, which is named asalphafrom now on. In this approach, the solution in the Pareto front withalpha closest to the one requested by the user is selected [21].
4. Benchmark Methods
To have a baseline to compareMOPSOresults, linearQR [7] and non-linearQRF [11] were used.
4.1. Quantile Regression
TheQR [7] algorithm is able to forecast quantiles by using linear models. In comparison, the least squares method makes a prediction of the target conditional average, whileQRpredicts the median or other quantiles. This is done by minimizing the quantile loss instead of the quadratic loss. Quantiles can be used for obtaining PIs. Letq1andq2be the1−PINC2and1+PINC2quantiles, respectively. There are two quantilesq1andq2for the left and right1+PINC2probability tails of the distribution, respectively. The coverage of the PI[q1,q2]is equal toPINC. To summarize,QRbuilds a couple of linear models to estimate theq1andq2quantiles, forming the[q1,q2]PI in turn.
4.2. Quantile Regression Forest
QRF [11] is an adaption of the random forest (RF ) algorithm [31] for estimating quantiles and follows a different strategy to linear quantile regression. RF is an ensemble learning algorithm whose individual models are regression trees. Each of the trees is obtained from different subsets of the training data by means of bootstrap sampling. To obtain an output for an instance, it is dropped down the tree until it reaches a leaf. The average of the response value of the training instances in the leaf is returned. RF can becomeQRFby first adapting the regression trees by changing the information that is stored in the leaf nodes. In the case ofQRF, leaf nodes store not only the average response, but all the response values of the training instances that reached the leaf. When an instance is dropped down each of the trees in theQRFensemble, every regression tree returns a set of response values (the ones in the leaves reached by the instance). From the union of all these response values, quantiles can be estimated. PIs can be computed from the quantiles, similar to what was done with quantile regression.
5. Experimental Validation
MOPSOand the two benchmark approaches (QRandQRF ) were evaluated to construct PIs for the study case described in Section 2. The main comparison refers to using and not using the power measures att0. The latter is identified as+Pt0 . Therefor, six configurations were tested:MOPSO,MOPSO+Pt0,QR,QR+Pt0,QRF, andQRF+Pt0. PIs obtained from the different approaches were evaluated using two metrics: average interval width (AIW) and ratio (obtained asPICP/AIW). The larger is the ratio, the better is the trade-off betweenPICPandAIW. Solutions that reach largePICPwith wide intervals are penalized with small ratio values.
PIs were obtained independently for each of the five forecast horizons (1–5 h). Several target nominal coverage values (alpha) were considered within the range 0.02–0.20.QRandQRFapproaches must be run for each desiredalphavalue. TheMOPSOapproach is able to obtain PIs for allalphas in a single run (see Section 3). Ninety runs were executed forMOPSO(each run was valid for all thealphavalues) andQRFbecause they are stochastic methods. ForQR, only one run for eachalphawas executed, because it is not stochastic. The following three subsections describe the tuning of the hyper-parameters of the methods, the experimental results, and the statistical significance tests.
5.1. Hyper-Parameter Tuning
Two hyper-parameters were tuned forMOPSO : number of hidden neurons and number of iterations of PSO. The methodology used a training and validation set approach, where the validation set was used to compare and select the best hyper-parameter values. The values of hidden neurons explored were: 2, 4, 6, 8, 10, 15, 20, 30, and 50. The different values for the PSO iterations ranged from 1000 to 16,000 in steps of 1000. Ninety runs were carried out for each number of neurons and iterations, starting with different random seeds. Therefore, each of the 90 runs had its own hyper-parameter tuning process, which resulted in an optimal number of neurons and iterations for each run. The optimal hyperparameters were chosen using the validation set hypervolume ([21]). It is important to remark that this parameter optimization process was carried out independently for each of the five forecasting horizons. The 90-run average (and standard deviation) of the best hyper-parameters for each horizon and configuration (MOPSOandMOPSO+Pt0 ) is displayed in Table 1.
It is observed that the number of hidden neurons depends on the horizon, and long horizons seem to require fewer hidden neurons than short horizons. This trend is not systematic, but it is true for all stations and horizons that more neurons are required for the first hour than for the fifth hour. A possible explanation is that, for longer horizons, predictions are more noisy, and simpler models are required to avoid overfitting. In any case, it shows that tuning the number of neurons separately for each horizon is useful. The number of iterations is always between 12,000 and 14,000, below the maximum value of 16,000, which shows that the iteration limit was appropriate. Nevertheless, we tried increasing this value for a few runs, but no significant changes in the Pareto fronts were observed.
ForQR, no parameter tuning is required. However, the performance of RF method depends on two main parameters: nodesize (minimum number of samples in the tree leaves) and mtry (number of input attributes used en the trees). Therefore, forQRF, we also performed hyper-parameter tuning to select the most suitable configuration. The values explored for nodesize were 2, 5, 10, 20, 30, 40, 50, and 100 and for mtry 2, 4, 6, 8, 10, 12, and 14 (the last one only when past value of power (+Pt0 ) was used as input). The measure used for selecting the best combination of parameters forQRFwas the ratio in the validation set. The study of best parameters was also carried out for each prediction horizon, as forMOPSOapproach. In this case, values of the parameters are very similar for the three stations, the five forecasting horizons, and the different targetalphavalues. Nodesize values are around5.6±0.86with no past power and around5.6±0.80when+Pt0 was used. For mtry, the average values are around11.5±0.62when no past power was used and12.5±0.78when+Pt0 was used (bearing in mind that, when past power was used, an additional attribute was added to the set of inputs).
5.2. Experimental Results
To have a first view of the performance of the different methods, Table 2 shows the mean of the ratio values for each method and each horizon, separately for Stations 1–3. Each ratio value is the mean of the ratios for all thealphavalues (0.02, 0.05, 0.08, 0.1, 0.15, and 0.2).
The results show thatMOPSO+Pt0obtains better results thanMOPSOfor the nearest horizons (especially Horizons 1 and 2) and for the three locations. For the other methods (QRandQRF), the+Pt0variant also behaves better than the original method, and this situation occurs for all the prediction horizons and all the stations. ComparingMOPSO+Pt0with the other methods, we can observe that, for all horizons and stations, it obtains higher ratio values. When we compareQR+Pt0andQRF+Pt0methods, we can observe that there is not a clear predominance of any method. In some cases, the ratio values ofQRF+Pt0are better (Station 1), but, in other cases, it is the opposite, as in Station 3. Next, the results are analyzed in more detail.
Figure 2, Figure 3 and Figure 4 display the results (ratio andAIW) for each of the three stations, respectively. Each of the figures contain two plots, top for the ratio (a) and bottom forAIW(b). These figures compareMOPSO,QR,QRF,MOPSO+Pt0,QR+Pt0, andQRF+Pt0. Information forMOPSO,MOPSO+Pt0,QRF, andQRF+Pt0is displayed by means of box plots, because they are a good summary of the 90 runs.QRandQR+Pt0are shown as horizontal lines, becauseQRis not stochastic. Plots can be seen for some representativealpha values (0.02, 0.05, 0.08, 0.1, 0.15, and 0.2). Figure 2a displays the ratio of the six methods tested for Station 1. It is readily apparent that ratio ofMOPSO+Pt0is larger thanMOPSOfor the first horizon, and, to a lesser degree, for the second one. This should be expected, because the influence of measures at timet0should decrease as the horizon is farther away fromt0. It can also be noticed that the difference betweenMOPSO+Pt0andMOPSOdecreases asalphaincreases. For instance, for the second horizon,MOPSO+Pt0is clearly better thanMOPSOforalpha=0.02but quite similar foralpha=0.2 . The second issue to notice in Figure 2 is that, althoughQRandQRFalso benefit from usingt0at short horizons,MOPSOconfigurations typically outperform theirQRandQRFcounterparts (i.e.,MOPSO+Pt0’s ratio is larger than those ofQR+Pt0andQRF+Pt0, and it also is whent0is not used). This is true for all horizons andalphas. TheAIW plot on the bottom of Figure 2b confirms the previous findings:MOPSO+Pt0intervals are narrower thanMOPSOfor the first two horizons (and that is also true forQR,QR+Pt0,QRF, andQRF+Pt0), highlighting the importance of usingt0for obtaining narrow intervals. Again,MOPSOconfigurations always obtain narrower intervals than theQRandQRF baselines. Figure 3 supports the previous results for Station 2. In this case, higher ratios are obtained byMOPSO+Pt0(vs.MOPSO,QR+Pt0andQRF+Pt0 ) for longer horizons (from 1 to 4 h). Finally, those results are mirrored by Station 3, as shown in Figure 4, although in this case there is an anomaly foralpha= 0.1, whereQR+Pt0manages to obtain better ratios (and narrower intervals) thanMOPSO+Pt0for the first forecasting horizon. In this station, it is also observed thatQRF+Pt0obtains similar ratios asMOPSO+Pt0foralpha>0.05and Horizon 1. For the remaining horizons,MOPSO+Pt0outperformsQRF+Pt0.
To show the PIs constructed for some specific days, Figure 5 displays the PIs obtained byMOPSO,MOPSO+Pt0, andQRF+Pt0on 2013-02-25, 2014-03-19, 2014-04-19, 2013-05-18, and 2012-06-26. PIs were obtained for Station 2 andalpha=0.1 . Each row in Figure 5 corresponds to each different day. The left column comparesMOPSO+Pt0(grey) withMOPSO(yellow) and the right column comparesMOPSO+Pt0(grey) withQRF+Pt0(yellow). The x-axis displays the five forecasting horizons and the y-axis shows the PIs (in grey and yellow) and the actual normalized power (a black line). Although the results depend on the particular day, in general, it is observed thatMOPSO+Pt0provides narrower intervals thanMOPSO(except in the last horizon (5 h) for some of the days) andQRF+Pt0.
5.3. Statistical Significant Tests
The above figures offer a qualitative view of the results, but, given thatMOPSOandQRF were run 90 times, statistical tests could also be computed. In this work, we used the Wilcoxon test. The results are summarized in Figure 6, for Stations 1–3, repectively. In this case, only tests for the ratio are shown (butAIW follows a similar pattern and can be found in Appendix A. Part (a) (top) of Figure 6 comparesMOPSO+Pt0andMOPSO. A blue point is used when theMOPSO+Pt0’s ratio is significantly larger thanMOPSO, and red for the other way around. Color grey signals thatMOPSO+Pt0andMOPSO are not statistically different. lThe lower plots (b) and (c) of Figure 6 perform the same comparison, but forMOPSO+Pt0versusQR+Pt0and forMOPSO+Pt0versusQRF+Pt0 . Results for Station 1 (see Figure 6a) show that, indeed, for the first two horizons,MOPSO+Pt0’s ratios are better thanMOPSO’s (blue points for allalphas). However, beyond the second hour,MOPSO+Pt0is not necessarily significantly better (and, in very few cases, it is significantly worse: red points). The comparisons betweenMOPSO+Pt0andQR+Pt0andMOPSO+Pt0andQRF+Pt0offer more systematic results: they are almost all blue points, which means thatMOPSO+Pt0is basically significantly better thanQR+Pt0and better thanQRF+Pt0 . The second station shows even cleaner results: Figure 7a displays blue points for the first four hours, and mostly grey for the fifth horizon. Again, the blue-filled plot of Figure 7b means thatMOPSO+Pt0is significantly better thanQR+Pt0andQRF+Pt0in all cases. In summary, for Station 2,t0has a beneficial effect onMOPSO for even longer horizon spans (up to the fourth hour). Finally, Figure 8 plots results for Station 3. In this case, the top plot shows blue points up to the third hour, all grey for the fourth horizon, but mostly red for the fifth one. With respect toMOPSO+Pt0vs.QR+Pt0 (Figure 8b), it is again mostly blue points, but the anomalies mentioned in the qualitative results pop up again here (red points), for the first horizon andα=0.09, 0.1, and 0.11. In this figure it is also observed than the observed similarity ofQRF+Pt0andMOPSO+Pt0previously observed for Station 3 and for the first horizon manifests as differences not statistically significant (grey points).
In summary,MOPSO+Pt0is significantly better thanMOPSOfor the short horizons for allalphavalues. For long horizons,MOPSO+Pt0is not significantly worse thanMOPSOexcept for somealphavalues (mostly in Station 1).MOPSO+Pt0is significantly better thanQR+Pt0andQRF+Pt0for all horizons, except for a fewalphavalues for the first horizon in Station 3.
6. Conclusions
The LUBE approach is an interesting alternative for estimating PIs in the context of probabilistic forecasting because it is able to estimate directly the lower and upper bounds of PIs. LUBE can be optimized byMOPSO, a multi-objective evolutionary technique, so that neural networks can be trained to simultaneously optimize the two conflicting properties of PIs: coverage and width. The result of this optimization is a Pareto front, from which solutions can be extracted according to the desired coverage value. In this study, this approach was used to address two issues to obtain PIs for solar power forecasting. The first one was to estimate PIs for five intra-day forecasting horizons from 1 to 5 h. The second, and more important one, was to study the influence of using measured solar power at timet0on the quality of PIs. Previous work has shown that this is useful for improving point forecasting, but we studied this issue in the context of probabilistic forecasting.
The approach was applied to data from three solar stations in Australia and experiments were carried out using two main configurations: (1) using meteorological forecasts variables as inputs to the methods; and (2) using additionally the measured solar output power as input. To analyze how far the influence of measured output reaches, hourly forecasts horizons from 1 to 5 h were used. The quality of prediction intervals was estimated using the coverage/width ratio and the width of the intervals. High values for ratio mean that intervals have a good trade-off between coverage and width. This study was done for several desired coverage values (oralpha) from 0.01 to 0.20.
The results show that the ratio is improved by using the measured additional information for the two first horizons on the three locations studied and all the desired coverage values. However, although for one of the station this beneficial influence reaches up to the fourth horizon, in none of the stations the ratio is improved for the farthest horizon tested (5 h). The same trend can be observed for interval width. Experiments were replicated 90 times for different random seeds and statistical significance tests were performed, which show that the mentioned results are statistically significant. Thus, it can be concluded that using measured solar power reduces the uncertainty of the intervals for short forecasting horizons.
TheMOPSOapproach was compared withQRandQRFas baseline methods. Both were tested in the same conditions asMOPSO, with and without the measured power at timet0. For all approaches, the use of measured power helped to obtain better PIs, especially in the first horizons. The comparison also shows thatMOPSO+Pt0is significantly better in all cases except a fewalphavalues in one of the stations.
Figure 1. Box plot of solar normalized power, broken down by hour, for: Station 1 (left); Station 2 (middle); and Station 3 (right). The x-axis displays UTC hour.
Figure 2. Station 1: (a)PICP/AIWratio per prediction horizon; and (b)AIWper prediction horizon.Alphavalues = 0.02 to 0.2.
Figure 3. Station 2: (a)PICP/AIWratio per prediction horizon; and (b)AIWper prediction horizon.Alphavalues = 0.02 to 0.2.
Figure 4. Station 3: (a)PICP/AIWratio per prediction horizon; and (b)AIWper prediction horizon.Alphavalues = 0.02 to 0.2.
Figure 5. PIs for the five forecast horizons for four specific days of the year (one day per row): The left column comparesMOPSO+Pt0(grey) withMOPSO(yellow) and the right column comparesMOPSO+Pt0(grey) withQRF+Pt0(yellow).
Figure 6. Station 1. Statistical tests for ratio: (a)MOPSO+Pt0vs.MOPSO; (b)MOPSO+Pt0vs.QR+Pt0; and (c)MOPSO+Pt0vs.QRF+Pt0.
Figure 7. Station 2. Statistical tests for ratio: (a)MOPSO+Pt0vs.MOPSO; (b)MOPSO+Pt0vs.QR+Pt0; and (c)MOPSO+Pt0vs.QRF+Pt0.
Figure 8. Station 3. Statistical tests for ratio: (a)MOPSO+Pt0vs.MOPSO; (b)MOPSO+Pt0vs.QR+Pt0; and (c)MOPSO+Pt0vs.QRF+Pt0.
Figure 8. Station 3. Statistical tests for ratio: (a)MOPSO+Pt0vs.MOPSO; (b)MOPSO+Pt0vs.QR+Pt0; and (c)MOPSO+Pt0vs.QRF+Pt0.
Station 1 | MOPSO | MOPSO+Pt0 | ||
---|---|---|---|---|
Horizon | Neurons | Iterations | Neurons | Iterations |
1 h | 23.7(±15.0) | 12889(±3030) | 23.2(±13.8) | 13922(±2284) |
2 h | 20.9(±13.1) | 13644(±2849) | 24.2(±16.1) | 13533(±2695) |
3 h | 14.1(±11.1) | 12622(±3337) | 12.4(±9.74) | 12900(±3198) |
4 h | 15.9(±11.8) | 12244(±3709) | 15.6(±12.8) | 12833(±3188) |
5 h | 14.0(±9.65) | 13322(±3072) | 12.1(±7.25) | 13633(±2542) |
Station 2 | MOPSO | MOPSO+Pt0 | ||
Horizon | Neurons | Iterations | Neurons | Iterations |
1 h | 21.2(±13.0) | 12989(±2696) | 19.6(±13.3) | 13544(±2995) |
2 h | 22.2(±14.7) | 13522(±2837) | 22.7(±15.1) | 13822(±2401) |
3 h | 18.7(±13.9) | 13056(±2977) | 15.2(±12.1) | 12744(±3248) |
4 h | 14.7(±10.8) | 12967(±3303) | 14.2(±9.75) | 12433(±3250) |
5 h | 11.8(±9.29) | 12878(±3395) | 13.2(±10.1) | 12656(±3529) |
Station 3 | MOPSO | MOPSO+Pt0 | ||
Horizon | Neurons | Iterations | Neurons | Iterations |
1 h | 21.7(±13.0) | 13544(±2623) | 18.0(±13.5) | 12756(±3520) |
2 h | 17.7(±13.5) | 12689(±3114) | 20.9(±13.7) | 13044(±3263) |
3 h | 18.6(±13.9) | 13056(±3279) | 19.2(±13.5) | 12789(±3224) |
4 h | 16.6(±12.1) | 12422(±3219) | 15.3(±11.4) | 12911(±3100) |
5 h | 13.5(±10.9) | 12744(±3092) | 14.8(±11.4) | 13433(±2825) |
Station 1 | ||||||
---|---|---|---|---|---|---|
Horizon | MOPSO | MOPSO+Pt0 | QR | QR+Pt0 | QRF | QRF+Pt0 |
1 | 2.07 | 2.69 | 1.71 | 2.47 | 1.80 | 2.58 |
2 | 2.23 | 2.30 | 1.80 | 1.98 | 1.88 | 1.96 |
3 | 2.35 | 2.34 | 1.90 | 1.91 | 1.97 | 1.98 |
4 | 2.37 | 2.36 | 1.92 | 1.95 | 2.01 | 2.01 |
5 | 2.33 | 2.33 | 2.04 | 2.08 | 2.16 | 2.18 |
Station 2 | ||||||
Horizon | MOPSO | MOPSO+Pt0 | QR | QR+Pt0 | QRF | QRF+Pt0 |
1 | 2.23 | 2.79 | 1.77 | 2.52 | 1.96 | 2.61 |
2 | 2.14 | 2.34 | 1.77 | 1.95 | 1.83 | 2.00 |
3 | 2.06 | 2.15 | 1.76 | 1.92 | 1.76 | 1.87 |
4 | 2.02 | 2.10 | 1.68 | 1.78 | 1.68 | 1.81 |
5 | 1.89 | 1.89 | 1.58 | 1.60 | 1.61 | 1.63 |
Station 3 | ||||||
Horizon | MOPSO | MOPSO+Pt0 | QR | QR+Pt0 | QRF | QRF+Pt0 |
1 | 2.37 | 2.89 | 2.02 | 2.77 | 2.09 | 2.84 |
2 | 2.30 | 2.41 | 2.02 | 2.20 | 1.97 | 2.14 |
3 | 2.15 | 2.20 | 1.96 | 2.01 | 1.90 | 1.95 |
4 | 2.13 | 2.14 | 1.92 | 1.95 | 1.85 | 1.89 |
5 | 2.11 | 2.09 | 1.90 | 1.91 | 1.86 | 1.87 |
Author Contributions
All authors participated in all tasks related to this article (conceptualization, software, validation, data curation, writing original draft and review/editing).
Funding
This work was funded by the Spanish Ministry of Science under contract ENE2014-56126-C2-2-R (AOPRIN-SOL project).
Conflicts of Interest
The authors declare no conflict of interest.
Appendix A
This appendix displays the statistical results forAIW.
Figure A1.Station 1. Statistical tests for wideness: (a)MOPSO+Pt0vs.MOPSO ; (b)MOPSO+Pt0vs.QR+Pt0 ; and (c)MOPSO+Pt0vs.QRF+Pt0.
Figure A1.Station 1. Statistical tests for wideness: (a)MOPSO+Pt0vs.MOPSO ; (b)MOPSO+Pt0vs.QR+Pt0 ; and (c)MOPSO+Pt0vs.QRF+Pt0.
Figure A2.Station 2. Statistical tests for wideness: (a)MOPSO+Pt0vs.MOPSO ; (b)MOPSO+Pt0vs.QR+Pt0 ; and (c)MOPSO+Pt0vs.QRF+Pt0.
Figure A2.Station 2. Statistical tests for wideness: (a)MOPSO+Pt0vs.MOPSO ; (b)MOPSO+Pt0vs.QR+Pt0 ; and (c)MOPSO+Pt0vs.QRF+Pt0.
Figure A3.Station 3. Statistical tests for wideness: (a)MOPSO+Pt0vs.MOPSO ; (b)MOPSO+Pt0vs.QR+Pt0 ; and (c)MOPSO+Pt0vs.QRF+Pt0.
Figure A3.Station 3. Statistical tests for wideness: (a)MOPSO+Pt0vs.MOPSO ; (b)MOPSO+Pt0vs.QR+Pt0 ; and (c)MOPSO+Pt0vs.QRF+Pt0.
1. Raza, M.Q.; Nadarajah, M.; Ekanayake, C. On recent advances in PV output power forecast. Sol. Energy 2016, 136, 125-144.
2. Pinson, P.; Chevallier, C.; Kariniotakis, G.N. Trading wind generation from short-term probabilistic forecasts of wind power. IEEE Trans. Power Syst. 2007, 22, 1148-1156.
3. Matos, M.A.; Bessa, R.J. Setting the operating reserve using probabilistic wind power forecasts. IEEE Trans. Power Syst. 2010, 26, 594-603.
4. Van der Meer, D.W.; Widén, J.; Munkhammar, J. Review on probabilistic forecasting of photovoltaic power production and electricity consumption. Renew. Sustain. Energy Rev. 2018, 81, 1484-1512.
5. Bessa, R.; Möhrlen, C.; Fundel, V.; Siefert, M.; Browell, J.; Haglund El Gaidi, S.; Hodge, B.M.; Cali, U.; Kariniotakis, G. Towards improved understanding of the applicability of uncertainty forecasts in the electric power industry. Energies 2017, 10, 1402.
6. Dobschinski, J.; Bessa, R.; Du, P.; Geisler, K.; Haupt, S.E.; Lange, M.; Möhrlen, C.; Nakafuji, D.; de la Torre Rodriguez, M. Uncertainty forecasting in a nutshell: Prediction models designed to prevent significant errors. IEEE Power Energy Mag. 2017, 15, 40-49.
7. Koenker, R.; Hallock, K.F. Quantile regression. J. Econ. Perspect. 2001, 15, 143-156.
8. Cannon, A.J. Quantile regression neural networks: Implementation in R and application to precipitation downscaling. Comput. Geosci. 2011, 37, 1277-1284.
9. Friedman, J.H. Greedy Function Approximation: A Gradient Boosting Machine. Ann. Stat. 2000, 29, 1189-1232.
10. Friedman, J.H. Stochastic gradient boosting. Comput. Stat. Data Anal. 2002, 38, 367-378.
11. Meinshausen, N. Quantile regression forests. J. Mach. Learn. Res. 2006, 7, 983-999.
12. Delle Monache, L.; Eckel, F.A.; Rife, D.L.; Nagarajan, B.; Searight, K. Probabilistic weather prediction with an analog ensemble. Mon. Weather Rev. 2013, 141, 3498-3516.
13. Cervone, G.; Haupt, S.E. Probabilistic Forecasting of solar energy generation using numerical modeling and remote sensing. In Proceedings of the AGU Fall Meeting Abstracts, San Francisco, CA, USA, 12-16 December 2018.
14. Clemente-Harding, L.; Young, G.; Cervone, G.; Delle Monache, L.; Haupt, S.E. Toward a Search Space Extension for the Analog Ensemble technique to Improve Probabilistic Predictions. In Proceedings of the AGU Fall Meeting Abstracts, San Francisco, CA, USA, 12-16 December 2018.
15. Pinson, P.; Nielsen, H.A.; Møller, J.K.; Madsen, H.; Kariniotakis, G.N. Non-parametric probabilistic forecasts of wind power: Required properties and evaluation. Wind Energy 2007, 10, 497-516.
16. Khosravi, A.; Nahavandi, S.; Creighton, D.; Atiya, A.F. Lower upper bound estimation method for construction of neural network-based prediction intervals. IEEE Trans. Neural Netw. 2011, 22, 337-346.
17. Wan, C.; Xu, Z.; Pinson, P. Direct interval forecasting of wind power. Power Syst. IEEE Trans. 2013, 28, 4877-4878.
18. Khosravi, A.; Nahavandi, S. Combined nonparametric prediction intervals for wind power generation. Sustain. Energy IEEE Trans. 2013, 4, 849-856.
19. Kirkpatrick, S.; Gelatt, C.D.; Vecchi, M.P. Optimization by simulated annealing. Science 1983, 220, 671-680.
20. Eberhart, R.C.; Shi, Y.; Kennedy, J. Swarm Intelligence; Elsevier: Amsterdam, The Netherlands, 2001.
21. Galván, I.M.; Valls, J.M.; Cervantes, A.; Aler, R. Multi-objective evolutionary optimization of prediction intervals for solar energy forecasting with neural networks. Inf. Sci. 2017, 418, 363-382.
22. McGovern, A.; Gagne, D.J.; Basara, J.; Hamill, T.M.; Margolin, D. Solar energy prediction: An international contest to initiate interdisciplinary research on compelling meteorological problems. Bull. Am. Meteorol. Soc. 2015, 96, 1388-1395.
23. Aguiar, L.M.; Pereira, B.; Lauret, P.; Díaz, F.; David, M. Combining solar irradiance measurements, satellite-derived data and a numerical weather prediction model to improve intra-day solar forecasting. Renew. Energy 2016, 97, 599-610.
24. Wolff, B.; Kühnert, J.; Lorenz, E.; Kramer, O.; Heinemann, D. Comparing support vector regression for PV power forecasting to a physical modeling approach using measurement, numerical weather prediction, and cloud motion data. Sol. Energy 2016, 135, 197-208.
25. Martín-Vázquez, R.; Aler, R.; Galván, I. Wind Energy Forecasting at Different Time Horizons with Individual and Global Models. In Proceedings of the IFIP International Conference on Artificial Intelligence Applications and Innovations, lRhodes, Greece, 25-27 May 2018; Springer: New York, NY, USA, 2018; pp. 240-248.
26. Martín-Vázquez, R.; Huertas-Tato, J.; Aler, R.; Galván, I. Studying the Effect of Measured Solar Power on Evolutionary Multi-objective Prediction Intervals. In Proceedings of the International Conference on Intelligent Data Engineering and Automated Learning, Madrid, Spain, 21-23 November 2018; Springer: New York, NY, USA, 2018; pp. 155-162.
27. Lauret, P.; David, M.; Pedro, H. Probabilistic solar forecasting using quantile regression models. Energies 2017, 10, 1591.
28. Zamo, M.; Mestre, O.; Arbogast, P.; Pannekoucke, O. A benchmark of statistical regression methods for short-term forecasting of photovoltaic electricity production. Part II: Probabilistic forecast of daily production. Sol. Energy 2014, 105, 804-816.
29. Abuella, M.; Chowdhury, B. Hourly probabilistic forecasting of solar power. In Proceedings of the IEEE 2017 North American Power Symposium (NAPS), Morgantown, WV, USA, 17-19 September 2017; pp. 1-5.
30. Hong, T.; Pinson, P.; Fan, S.; Zareipour, H.; Troccoli, A.; Hyndman, R.J. Probabilistic energy forecasting: Global Energy Forecasting Competition 2014 and beyond. Int. J. Forecast. 2016, 32, 896-913.
31. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5-32.
Ricardo Aler*, Javier Huertas-Tato, José M. Valls and Inés M. Galván
Computer Science Department, Universidad Carlos III de Madrid, 30 Avenida Universidad, Leganes, 28911 Madrid, Spain
*Author to whom correspondence should be addressed.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2019. This work is licensed under http://creativecommons.org/licenses/by/3.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Prediction Intervals are pairs of lower and upper bounds on point forecasts and are useful to take into account the uncertainty on predictions. This article studies the influence of using measured solar power, available at prediction time, on the quality of prediction intervals. While previous studies have suggested that using measured variables can improve point forecasts, not much research has been done on the usefulness of that additional information, so that prediction intervals with less uncertainty can be obtained. With this aim, a multi-objective particle swarm optimization method was used to train neural networks whose outputs are the interval bounds. The inputs to the network used measured solar power in addition to hourly meteorological forecasts. This study was carried out on data from three different locations and for five forecast horizons, from 1 to 5 h. The results were compared with two benchmark methods (quantile regression and quantile regression forests). The Wilcoxon test was used to assess statistical significance. The results show that using measured power reduces the uncertainty associated to the prediction intervals, but mainly for the short forecasting horizons.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer