1. Introduction
In the face of deteriorating environment and increasing energy demand, countries around the world are actively developing renewable energy and vigorously building low-carbon economy. With the advantages of pollution-free, noiseless and renewable approaches [1], photovoltaic (PV) solar power generation has huge market and industrial development potential. However, PV solar power generation is easily affected by weather factors, resulting in significant randomness of the output power. With the increasing penetration of solar power, this randomness brings great challenges to power system dispatch and the safe and stable operation of the power system. Accurate prediction of PV solar power generation can effectively address the above challenges and has important practical significance.
1.1. Literature Review
Compared with physical and statistical models, machine learning models have higher prediction accuracy and reliability [2] and have attracted much attention in PV prediction research [3]. For example, artificial neural network (ANN) [4,5,6], support vector machine (SVM) [7,8], and extreme learning machine (ELM) [9,10] were proven to have acceptable performances in the field of PV prediction. Combining optimization algorithms with machine learning models is also promising and widely studied [11,12]. As a temporal recurrent neural network, a long short-term memory (LSTM) network is very suitable for time series prediction tasks. It has been investigated by many researchers for PV prediction. For example, the LSTM network was employed for solar irradiation prediction [13,14]. The LSTM network was used for solar power prediction together with wavelet transform and K-singular value decomposition [15,16]. Moreover, the LSTM model can be combined with other prediction models to achieve better accuracy and reliability, such as convolutional neural network (CNN)-LSTM models [17,18,19] and LSTM-recurrent neural network (RNN) models [20].
Integrated models combine multiple base learners in serial/parallel manners, which helps to achieve more robust and generalized performance than any individual base learner [21,22,23]. In the field of PV prediction, Chaouachi et al. [21] proposed an integrated model containing four neural networks for short-term solar power forecasting. Experimental results proved that the proposed model provided more accurate forecasts than some conventional networks. Liu et al. [22] provided an integrated CNN-LSTM-multi layer perceptron (MLP) model for solar irradiance prediction. This integrated model was superior to the traditional models in terms of prediction accuracy and robustness. Zhou et al. [23] developed an ensemble PV power prediction model. Four base learners were selected and their prediction results were linearly combined. Results showed the developed model had better accuracy than any single base learner.
For complex time series prediction tasks, the choice of prediction model is critical. However, the data processing and analysis are equally important to the prediction performance, yet they are prone to be ignored. For renewable energy prediction, in addition to conventional data processing such as normalization and outlier detection, data clustering can provide a better understanding of observed data’s internal patterns and characteristics and thus help improve the prediction performance. Traditional clustering methods have been widely adopted [24,25,26,27] and proved to be able to improve the final prediction accuracy. Benmouiza and Cheknane [24] combined K-means clustering with autoregressive neural networks for solar radiation prediction and demonstrated the advantages of this method. Azimi et al. [25] developed a hybrid solar radiation prediction model using the transformation-based K-means algorithm and an MLP neural network, demonstrating the effectiveness of data clustering in prediction. Malakar et al. [26] built an adaptive cluster-specific solar prediction model using the K-medoids algorithm and proved the model effectiveness through data of nine solar stations in India. Note that the basic K-means clustering methods have drawbacks such as sensitivity to noise and computational complexity susceptible to the initial clustering centroids [25], which makes the clustering effect limited. The above literature works on PV solar energy prediction are summarized in Table 1.
1.2. Main Contributions of This Paper
From the above literature review, it is known that integrated models have performance advantages for PV solar power prediction. To improve this kind of generalized modeling performance for multiple horizon prediction tasks, this paper focuses on data analysis and model design simultaneously to build a more accurate and robust PV prediction framework. The LSTM network is suitable as an individual prediction model, but combining it with an integrated learning strategy can further improve the prediction accuracy. In addition, introducing data clustering with optimized clustering centroids to PV solar power prediction has the potential to further enhance the prediction performance. The main contributions of this paper are as follows:
I. An optimized data clustering method is proposed, in which the clustering centroids are optimized by the improved particle swarm optimization (iPSO) algorithm. Compared with no clustering and the common K-means clustering, iPSO-based clustering can obtain a better understanding of the PV data’s internal characteristics and thus has the potential to improve the prediction accuracy.
II. A novel framework for PV solar power prediction is proposed using iPSO-based data clustering and the AdaLSTM network. The AdaLSTM network belongs to the integrated model family, in which multiple LSTM networks are serially integrated together through the AdaBoost algorithm. Two real PV case studies demonstrate that the AdaLSTM network outperforms the persistence, BPNN, CNN, and LSTM in terms of the model accuracy. Combining AdaLSTM with iPSO-based clustering can further improve the prediction accuracy.
1.3. Paper Organization
The rest of paper is organized as follows. In Section 2, some methods and principles are provided such as the improvements of PSO, iPSO-based data clustering, the process of AdaLSTM modeling, and the overall PV solar power prediction framework. In Section 3, comprehensive comparisons and analysis based on two real PV case studies are performed to demonstrate the validity of the proposed prediction model. In Section 4, a brief discussion is presented to show the novelty and superiority of this study. Finally, some conclusions are given in Section 5.
2. Methods
2.1. iPSO Based Data Clustering
2.1.1. Improved PSO Algorithm
PSO [28] is a classic swarm intelligence algorithm. Its basic formulas can be expressed as
(1)
(2)
where and are acceleration coefficients. and are random numbers within (0, 1). and are the velocity and position of particle i in the kth iteration, respectively. is the best position of particle i. is the position of the best particle. is the inertia weight.To reduce the possibility of falling into local optimum by a basic PSO algorithm [29,30], two improvement measures are combined and embedded into the algorithm through the ideas of comprehensive learning and adaptive weight adjustment. Details are as follows (called Improvements I and II):
I. Particles learn from only [31,32,33]. It is possible to solve the problem of becoming trapped in local optimum for complex multimodal optimization. The velocity update is modified as follows:
(3)
where is the best position of particle q (q may be the same as i). Refresh pointer m is defined to count the number of iterations during which the best positions of all particles are not updated. If m reaches a specified refresh interval M, it is set to 0 and a random value is given to particle i. If this random value is larger than (expressed as follows), particle i will learn from . Otherwise, particle i will learn from the of other particles (). Note that m is initialized to 0 and set as m = m + 1 if the best positions of all particles are not changed.(4)
where denotes the learning probability of particle i. denotes the population size.II. To ensure the diversity of the inertia weight, the population is divided into three subgroups according to the individual fitness. For particle i with fitness , inertia weight is adjusted as follows:
(5)
where and denote the specified maximum and minimum of . and are the minimum and average fitness of all particles in the kth iteration, respectively. is the average fitness of particles whose fitness is less than . Note that in Equation (5), three conditions for weight segmentation can be considered to correspond to three different types of particles: better, common, and poorer particles. For the better particles which are already closer to the global optimum, their inertia weights should be smaller. For the poorer particles, their inertia weights are referred to the adaptive adjustment of the control parameters in the genetic algorithm.The proposed iPSO algorithm based on the above Improvement I and Improvement II is used for data clustering. We refer to Figure 1 for the flowchart of the improvement process.
2.1.2. Process of iPSO-Based Clustering
The K-means clustering algorithm divides the data into K clusters according to a similarity measure of Euclidean distance. It is known that the accuracy and computational complexity of the algorithm are significantly affected by the clustering centroids. Therefore, in this paper, the iPSO is adopted to determine the clustering centroids such that the fitness function (sum of the Euclidean distances of all data to the nearest centroid) is minimized. The flowchart of iPSO-based data clustering is shown in Figure 1. The clustering process is described as follows:
Step 1: Input PV data, initialize the clustering centroids (particles) and set the refresh pointer to 0.
Step 2: Perform the clustering operations based on the clustering centroids and calculate the fitness.
Step 3: Check if the stopping criteria are met. If yes, stop optimization and obtain the optimized clustering results. Otherwise, implement PSO Improvement II (according to Equation (5)) and PSO Improvement I (according to Equations (2)–(4)) to update the clustering centroids, and then proceed to Step 2.
The determination of the number of clusters is an important task for the data clustering. The clustering validation index—standard deviation-based index (STDI) [34]—is utilized to select the number of clusters. This index integrates the intra-cluster sample similarity as well as the inter-cluster sample variability, described as follows:
(6)
where denotes the centroid of cluster k. denotes the centroid of all samples, denotes the ith sample in cluster k. denotes the number of samples in cluster k. K denotes the number of clusters. denotes the Euclidean norm of the vector. The larger STDI value indicates the better clustering effect. Figure 1Flowchart of iPSO-based data clustering.
[Figure omitted. See PDF]
2.2. AdaLSTM
Different from the typical RNN, the hidden layer of the LSTM is no longer constructed by some nonlinear activation function (such as tanh). A series of complex calculations on the model input are performed through three gating units: forget gate, input gate, and output gate. See literature [35] for more detailed information of LSTM.
The AdaBoost algorithm is an integrated learning algorithm. In this algorithm, multiple base learners are trained to form a weighted strong learner, thereby improving the accuracy and adaptability of the model. In this paper, the AdaLSTM model, in which the AdaBoost algorithm and three basic LSTM networks (base learners) are combined, is proposed for PV solar power prediction. The structure of AdaLSTM is shown in Figure 2. It is seen that the training samples are assigned weights according to the training results of the previous base learner. The process of AdaLSTM modeling is shown in Algorithm 1.
Algorithm 1 AdaLSTM modeling |
Input Import training dataset (X, Y). X denotes feature input and Y denotes target output. Initialization Initialize the weight vector . N denotes the total number of samples. for i = 1 to P (number of base learners) Train the regression model . Calculate the error using testing set: e=. denotes the model output. for j = 1 to N Assign the weight vector according to the error. (7) end Set parameters of the ith base learner as: = 0.5 [36]. denotes the mean absolute error of the ith base learner. end Output Integrate the base learners into a strong learner: Y = . |
2.3. Overall Prediction Framework
The overall prediction framework adopted in this paper is shown in Figure 3. Detailed prediction steps are as follows:
Step 1: Collect the PV data and divide them into training and testing sets.
Step 2: For the training set, implement the iPSO-based data clustering as described in Section 2.1 (STDI is used to determine the number of clusters).
Step 3: For each data cluster, train the AdaLSTM model as introduced in Section 2.2.
Step 4: For the testing set, adopt the K-nearest neighbor (KNN) method [37] to achieve data classification. Then, obtain the final prediction results using the corresponding trained AdaLSTM model and carry out performance evaluation of the prediction model.
In this paper, three performance metrics are chosen to evaluate the prediction accuracy: mean absolute error (MAE), root mean square error (RMSE), and coefficient of determination ().
(8)
(9)
(10)
where denotes the ith measured value. denotes the ith predicted value. denotes the average of measured values. n denotes the number of data points.3. Case Studies
3.1. Data Description
Two cases from the DKA Solar Energy Center in Australia are adopted in this paper. Alice Springs PV station [38] is a solar technology demonstration facility which has been operating in arid conditions in central Australia for over 15 years. The Yulara Solar System [39] is a 1.8 MW solar PV plant with sites distributed across the vicinity of Australia’s landmark Uluru (Ayers Rock). The data records at Alice Springs PV station from 6 January to 16 February in 2020 (6 January 2020–16 February 2020) are used as Case A. The meteorological variables (wind direction, humidity, ambient temperature, diffuse irradiance, and global irradiance) and solar power are chosen as the input features of the prediction model. The data records from Yulara Solar System are used as Case B (13 January 2020–23 February 2020). Wind speed, air pressure, humidity, ambient temperature, global irradiance, and solar power are chosen as the input features.
Datasets of these two cases are both sampled every 5 min. To reduce zero power data at night, only the data from 5:30 to 20:00 are retained. For the collected data records, the former 80% are used as the training set, and the latter 20% are used as the testing set. The input features at time t are utilized to predict PV solar power generation at time , where is called the prediction horizon. Since the time resolution of the raw data is 5 min, the prediction resolution and prediction update rate are set as 5 min as well. The prediction horizon is set as 10 min, 30 min, and 60 min in this study.
3.2. Validation of the AdaLSTM Model
In this section, without the data clustering, the AdaLSTM prediction model is compared with some other models such as persistence [40], BPNN [41], CNN [42], and LSTM [43]. Among them, the CNN model basically comprises three convolutional layers and a maximum pooling layer. The pooling layer reduces the dimensions of the data by combining the outputs of neurons at one layer into a single neuron in the next layer. The structure and parameter settings of the BPNN, CNN, and LSTM are mainly referred to the corresponding references. The main parameters of the AdaLSTM are given in Table 2. Parameters such as maximum number of iterations and learn rate drop factor are referred to in literature [44]. Other parameters are determined through trial and error. Prediction horizons of 10, 30, and 60 min are performed, respectively. All calculations are carried out in MATLAB-R2021b software (Intel(R) Xeon(R) CPU, 3.10 GHz, and 16 GB RAM).
Table 3 shows comparison results of three performance metrics (Case A, average results of five independent runs). It can be seen that the proposed AdaLSTM model outperforms other models in terms of prediction accuracy for all three prediction horizons. Taking the 30 min prediction as an example, RMSE and MAE of the AdaLSTM model can be reduced, respectively, by 4.15% and 2.40%, and can increase by 0.40% compared with the LSTM model. In addition, it is observed that the prediction accuracy decreases as the prediction horizon increases. Using the real power (“Target”) as the x-axis and the predicted power (“Output”) as the y-axis, Figure 4 shows the scatter plots of four comparison models (30 min prediction). R denotes the multiple correlation coefficient [45]. As shown in Figure 4, the curve of AdaLSTM fitting is closest to line and the corresponding R is the largest, which indicates that the AdaLSTM model has the highest prediction accuracy. Figure 5 shows the radar plot of performance metrics for comparison models (Case A, 30 min prediction). It is also seen that the AdaLSTM model is the best in terms of prediction accuracy.
For Case B, the same five models also perform the prediction tasks and their results are listed in Table 4 (average results of five independent runs). It is seen that the AdaLSTM model still has the highest accuracy under different prediction horizons. Compared with the basic LSTM model (30 min prediction), the RMSE and MAE of AdaLSTM are reduced, respectively, by 2.02% and 14.55%, and increases by 0.15%. Note that in Case B, the unit of PV solar power is kW instead of W in Case A. The results of the two cases demonstrate that the AdaLSTM prediction model has good accuracy and robustness for different PV solar power generation scales.
3.3. Validation of iPSO Based Data Clustering
The performance of the iPSO algorithm is first verified using eight benchmark functions (see Table A1 in Appendix A for function details). The main parameters of iPSO are given in Table 5. Parameters and are referred to in literature [46]. Other parameters such as swarm size and M are determined through trial and error. Comparison results of the convergence graphs between the PSO and iPSO algorithms are shown in Figure 6. The results demonstrate that, for each benchmark function, iPSO is superior to the basic PSO in terms of accuracy. Considering the convergence speed, the iPSO also outperforms the basic PSO for most benchmark functions (two algorithms have similar convergence speeds for f2 and f6). The designed iPSO algorithm is then used to search the optimal clustering centroids for Case A, and the convergence comparison with other algorithms is shown in Figure 7. It can be seen that for cluster centroid searching, the iPSO has the best accuracy with fast convergence speed.
Taking Case A as an example, the global irradiance and PV solar power are selected as the clustering features. The number of clusters K is set from 2 to 10. STDI is used to determine the K value, as shown in Table 6. It is seen that K = 9 corresponds to the largest STDI value. Moreover, after the iPSO-based data clustering with K = 9, the corresponding model (30 min prediction) has the highest prediction accuracy.
3.4. Influence of Clustering on Prediction Accuracy
In this section, the influence of iPSO-based clustering on PV solar power prediction is discussed. Prediction results without clustering, with K-means clustering and with iPSO based clustering, are compared using both Case A and Case B. At the same time, the 30-min data computation time of Case A is chosen to perform a balance analysis between computational burden and computational accuracy. Note that the AdaLSTM prediction model is adopted for each data cluster or for the situation of no clustering.
For Case A and Case B, performance metrics for different clustering methods are compared in Table 7 and Table 8, respectively (average results of five independent runs). It is seen that, for all three prediction horizons, the proposed iPSO-based clustering outperforms others in terms of prediction accuracy. Compared with the commonly used K-means clustering, RMSE and MAE of the prediction model with iPSO-based clustering are reduced, respectively, by 10.75% and 9.90%, and increases by 0.14% (Case A, 30 min prediction). RMSE and MAE are reduced, respectively, by 6.69% and 7.89%, and increases by 0.09% (Case B, 30 min prediction). Prediction curves based on all data in the testing set are compared in Figure 8 (Case A) and Figure 9 (Case B). The curve using measured values is marked as “Real”. It is known that, no matter which case is used, no matter which prediction horizon is addressed, iPSO-based clustering can bring significant improvement in prediction accuracy compared with K-means clustering and no clustering. Moreover, for easier observation of the differences between different clustering methods, Figure 10 and Figure 11 are provided to show just one-day prediction curves for Case A and Case B, respectively. It can be clearly seen that the prediction curve with iPSO-based clustering can track the real curve more accurately.
For the computational burden issue, time consumption for different models is shown in Table 9. It is clearly seen that, due to the data clustering, the training time of the proposed model is shorter than that of the AdaLSTM. Since model training could be conducted offline, the training time of 105.70 s is acceptable in real applications. In addition, the 10.74 s prediction time is also acceptable for 10, 30 or 60 min real-time prediction. These results further demonstrate the effectiveness and applicability of the proposed hybrid prediction model.
4. Discussion
According to the two case studies for multiple prediction horizons, the performance of the proposed PV solar power prediction model is investigated. A brief discussion on the effects of iPSO-based data clustering and AdaLSTM is provided as below.
The proposed iPSO includes two improvements: (1) Particles learn from only the best position; (2) The inertia weight is dynamically adjusted according to the individual fitness. iPSO is utilized to optimize the clustering centroids, which provides a better understanding of the internal patterns and characteristics of the PV dataset. From the results, it is seen that iPSO is generally superior to basic PSO in terms of accuracy and convergence speed. See Figure 6 for benchmark function comparison and Figure 7 for optimal clustering centroid searching. Note that the STDI is used to determine the number of data clusters. On the other hand, three basic LSTM networks are integrated in a serial manner through the AdaBoost algorithm to construct the AdaLSTM model. Prediction results show that AdaLSTM enhances the prediction accuracy in both cases. Compared with the basic LSTM, the RMSEs of AdaLSTM for 10 min, 30 min, and 60 min predictions are reduced, respectively, by 8.17%, 4.15%, and 5.14% (Case A) or 2.03%, 2.01%, and 2.95% (Case B).
By combining iPSO-based clustering and AdaBoost (see Section 2.3 for detailed process), prediction accuracy could be further improved. Compared with K-means clustering, the RMSEs of iPSO clustering for 10 min, 30 min, and 60 min predictions are reduced, respectively, by 6.36%, 10.75%, and 5.18% (Case A) or 10.69%, 6.69%, and 10.39% (Case B). Overall, the RMSE of the proposed hybrid model is reduced, respectively, by 75.22%, 73.80%, 67.60%, 66.30%, 64.85%, and 10.75% compared with persistence, BPNN, CNN, LSTM, AdaLSTM without clustering and AdaLSTM with K-means clustering (Case A, 30 min prediction). For the computational burden (see Table 9), it is known that the training time of the proposed model (105.70 s) is shorter than that of the AdaLSTM (223.68 s), but longer than that of the basic LSTM (70.40 s). Considering the model training could be conducted offline, about 2 min of training time is acceptable in real applications. Moreover, the prediction time of 10.74 s is also acceptable for 10, 30, or 60 min PV solar power prediction.
5. Conclusions
In order to discover the potential distribution patterns of PV data and enhance the accuracy and reliability of solar power prediction, a hybrid prediction model is proposed in this paper based on iPSO-based clustering and the AdaLSTM network. For data clustering, the clustering centroids are optimized by the iPSO algorithm. For each data cluster, the AdaBoost algorithm and multiple basic LSTM networks are combined together to form the AdaLSTM model for model training. Final prediction results are obtained using the KNN method and the corresponding trained AdaLSTM model. Datasets from two PV stations in Australia are adopted for case studies. The results show that the AdaLSTM model outperforms other existing prediction models such as persistence, BPNN, CNN and LSTM. The RMSE and MAE can be reduced, respectively, by 4.15% and 2.40% compared with the best comparison model (Case A, 30 min prediction). Further, the analysis of iPSO in terms of accuracy and convergence speed and the prediction accuracy comparison between no clustering, K-means clustering, and iPSO-based clustering are performed to verify the superiority of the proposed hybrid model. Compared with the commonly used K-means clustering, the RMSE and MAE of the prediction model with iPSO-based clustering are reduced, respectively, by 10.75% and 9.90% (Case A, 30 min prediction). It is worth noting that during the data clustering process, the sum of the Euclidean distances of all data to the nearest centroid is chosen as the fitness function for iPSO, which indicates that unsupervised clustering is performed in this paper. In future research, feature selection of the prediction model will be studied, and the training error will be used to design the fitness function of iPSO to achieve supervised data clustering.
Conceptualization, W.X.; methodology, W.X.; software, J.L.; validation, K.L.; formal analysis, J.L.; data curation, J.L.; writing—original draft preparation, J.L.; writing—review and editing, W.X.; supervision, K.L. and W.X.; funding acquisition, K.L. All authors have read and agreed to the published version of the manuscript.
The data that support the findings of this study are available from the corresponding author upon reasonable request.
The authors declare no conflicts of interest.
AdaBoost | adaptive boosting |
AdaLSTM | adaptive boosting LSTM |
ANN | artificial neural network |
ARMA | autoregressive moving average model |
BPNN | back propagation neural network |
CB | clustering-based |
CHOA | chimp optimization algorithm |
CNN | convolutional neural networks |
DTCWT | dual tree complex wavelet transforms |
EEMD | ensemble empirical mode decomposition |
ELM | extreme learning machine |
iPSO | improved particle swarm optimization |
KNN | k-nearest neighbor |
KSVD | k-singular value decomposition |
LSSVM | least squares support vector machine |
LSSVR | least square support vector regression |
LSTM | long short-term memory |
MAE | mean absolute error |
MHFM | multiscale hybrid forecast model |
MLP | multilayer perceptron |
MLPNN | multilayer perceptron neural network |
MTLBO | modified teaching learning based optimization |
NAR | nonlinear autoregressive |
NNE | neural network ensemble |
PCA | principal components analysis |
PSO | particle swarm optimization |
PV | photovoltaic |
R | multiple correlation coefficient |
| coefficient of determination |
RVFL | random vector functional link |
RMSE | root mean square error |
RNN | recurrent neural network |
RVFLN | random vector functional link network |
SCADA | supervisory control and data acquisition |
STDI | standard deviation based index |
SVM | support vector machine |
VMD | variational mode decomposition |
WRF | weather research and forecasting |
WT | wavelet transform |
| best position of particle i |
| position of the best particle |
| learning probability of particle i |
| population size |
m | refresh pointer |
| inertia weight |
| minimum of inertia weight |
| maximum of inertia weight |
| minimum fitness of all particles |
| average fitness of all particles |
K | number of clusters |
| prediction horizon |
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Figure 4. Prediction output scatter plots of different models (Case A, no clustering).
Figure 5. Radar plot of performance metrics (Case A, no clustering): (a) RMSE, (b) MAE, (c) [Forumla omitted. See PDF.].
Figure 7. Convergence graph of clustering centroid optimization using different algorithms (Case A).
Figure 8. Prediction curves (Case A, 8 February 2020–16 February 2020): (a) 10 min prediction, (b) 30 min prediction, (c) 60 min prediction.
Figure 9. Prediction curves (Case B, 15 February 2020–23 February 2020): (a) 10 min prediction, (b) 30 min prediction, (c) 60 min prediction.
Figure 10. Prediction curves (Case A, from 5:30 to 20:00 on 11 February 2020): (a) 10 min prediction, (b) 30 min prediction, (c) 60 min prediction.
Figure 11. Prediction curves (Case B, from 5:30 to 20:00 on 21 February 2020): (a) 10 min prediction, (b) 30 min prediction, (c) 60 min prediction.
Literature review on PV solar energy prediction.
Ref | Location | Method | Dataset Covering Time | Performance Metrics | Characteristic |
---|---|---|---|---|---|
[ | Colorado | LSTM-MLP | 1 January 2012 to 31 December 2016 | RMSE (62.1618 W/m2) | The role of lag time was significant when the input variables of the LSTM model were small. |
[ | Mediterranean area | ANN | October 2002 to December 2008 | nRMSE (14.9%) | A hybrid approach that utilized a coupled artificial neural network (ANN) and autoregressive moving average (ARMA) predictors could substantially decrease prediction errors. |
[ | Guadeloupe island | MHFM model | January 2012 to December 2012 | nRMSE (4.43–10.24%) | Four types of typical days were identified and datasets were established for each. |
[ | Beijing | Hybrid WT-PSO-SVM | one year with a time step of 10 min | MAPE (4.22%) | The PSO was used to optimize the parameters of the SVM in order to achieve a higher forecasting accuracy. |
[ | Chhattisgarh | MTLBO-ELM | February 2019 to December 2019 | MAPE (8.2091%) | ELM models based on MTLBO optimization outperformed ELM, ELM (randomly fixed weights and biases), and ANN models based on different optimizations. |
[ | Arizona | REVL-CHOA | March 2010 to June 2011 | RMSE (0.00047, 0.05995) | The RFVL-CHOA method was found to be superior and more effective than other optimization models studied for performance prediction. |
[ | Sohag | RVFLN-WWO | NA | | The experimental results uniquely fitted with the predicted results of the proposed artificial intelligence model. |
[ | Dili | WRF-LSTM | January to December 2014 | nRMSE (16.18%) | The combination of the WRF and LSTM methods had better performance and could be applied to simulate other locally relevant weather variables. |
[ | Taiwan | CNN-LSTN | 5 January 2017 to 4 January 2018 | RMSE (0.0472) | The network allowed for time series forecasting using a feature-rich approach, resulting in competitive forecasting performance even with small datasets. |
[ | Yulara soalr system | KSVD-LSTM | 7 March to 7 June 2021 | RMSE (0.3682) | The combined method of KSVD and LSTM achieved highly accurate prediction results. |
[ | Abu Dhabi | CNN-LSTM | July 2019 | RMSE (0.36) | The model predicts both solar irradiance and POA more accurately than what has been reported in the literature. |
[ | Alice Springs | LSTM-Convolutional Network | half year | RMSE (0.621) | The proposed hybrid model outperformed the convolutional LSTM network and had a better prediction effect than the single prediction model. |
[ | Alberta | LSTM-RNN | seven months | MSE (0.00317) | The hybrid model had higher accuracy. |
[ | TUAT | NNE | 2007 to 2008 | MAPE (3.6387) | Compared with traditional networks, neural network ensembles exhibited the highest level of prediction accuracy. |
[ | Folsom | CNN-LSTM-MLP | 1 January 2014 to 31 December 2016 | RMSE (12.53 W/m2) | The model was more accurate and robust than many traditional alternative methods. |
[ | Ames | Kmeans-mlpnn | 08/25–31/2013 | RMSE (32.01 W/m2) | The comparison with the benchmark solar radiation forecast model indicated that the model had superior forecasting capabilities. |
[ | Algeria | Kmeans-NAR | January 1994 to December 1996 | nRMSE (0.1985) | The obtained experimental results showed that the clustering of the input space was an important task to interpret the behavior of the series. |
[ | Begamganj | CB-LSTM | 2013 | nRMSE (19.74) | The performance was improved compared with a single site-specific model. |
[ | Beijing | EEMD-LSSVR-K-LSSVR | 1 January 2009 to 30 June 2017 | nRMSE (2.96%) | The DCE learning method showed promise for predicting solar radiation, with high levels of horizontal and directional accuracy, as well as robustness. |
Main parameters of AdaLSTM.
Parameter | Setting |
---|---|
MaxEpochs | 100 |
Maximum number of iterations | 100 |
Number of hidden neurons | 200 |
Initial learn rate | 0.004 |
Learn rate drop period | 90 |
Learn rate drop factor | 0.2 |
Performance metrics of different models (Case A, no clustering).
Prediction Horizon | Model | RMSE (W) | MAE (W) | |
---|---|---|---|---|
Persistence | 311.89 | 159.80 | 0.9655 | |
BP | 381.46 | 216.80 | 0.9477 | |
10 min | CNN | 328.11 | 186.04 | 0.9616 |
LSTM | 324.06 | 191.53 | 0.9626 | |
AdaLSTM | 297.58 | 155.79 | 0.9685 | |
Persistence | 489.55 | 352.64 | 0.9150 | |
BP | 463.10 | 308.26 | 0.9236 | |
30 min | CNN | 374.39 | 253.01 | 0.9498 |
LSTM | 360.05 | 240.33 | 0.9538 | |
AdaLSTM | 345.12 | 234.57 | 0.9576 | |
Persistence | 834.88 | 654.25 | 0.7516 | |
BP | 698.60 | 477.88 | 0.8261 | |
60 min | CNN | 411.13 | 281.15 | 0.9396 |
LSTM | 384.51 | 271.86 | 0.9471 | |
AdaLSTM | 364.76 | 246.50 | 0.9526 |
Performance metrics of different models (Case B, no clustering).
Prediction Horizon | Model | RMSE (kW) | MAE (kW) | |
---|---|---|---|---|
Persistence | 15.29 | 7.43 | 0.9462 | |
BP | 12.65 | 6.28 | 0.9637 | |
10 min | CNN | 12.37 | 6.65 | 0.9652 |
LSTM | 12.08 | 5.48 | 0.9669 | |
AdaLSTM | 11.84 | 5.40 | 0.9684 | |
Persistence | 22.50 | 15.13 | 0.8835 | |
BP | 15.40 | 8.86 | 0.9461 | |
30 min | CNN | 14.12 | 8.14 | 0.9546 |
LSTM | 12.39 | 7.01 | 0.9651 | |
AdaLSTM | 12.14 | 5.99 | 0.9665 | |
Persistence | 32.82 | 25.55 | 0.7562 | |
BP | 20.97 | 14.17 | 0.8990 | |
60 min | CNN | 15.72 | 8.89 | 0.9437 |
LSTM | 13.63 | 7.76 | 0.9577 | |
AdaLSTM | 13.24 | 7.61 | 0.9601 |
Main parameter of iPSO.
Parameter | Setting |
---|---|
Swarm size | 50 |
| 0.89 |
| 0.7 |
M | 10 |
| 1.5 |
Comparison of indexes for different K (Case A).
K | STDI Value | RMSE (W) | MAE (W) | |
---|---|---|---|---|
2 | 10 | 277.04 | 159.06 | 0.9727 |
3 | 25 | 258.84 | 154.01 | 0.9761 |
4 | 81 | 199.53 | 121.31 | 0.9858 |
5 | 229 | 201.37 | 131.16 | 0.9856 |
6 | 374 | 160.43 | 108.27 | 0.9901 |
7 | 529 | 169.44 | 112.96 | 0.9898 |
8 | 659 | 189.64 | 121.57 | 0.9911 |
9 | 1801 | 126.09 | 88.58 | 0.9943 |
10 | 1488 | 135.95 | 88.21 | 0.9934 |
Performance metrics for different clustering methods (Case A).
Prediction Horizon | Clustering Method | RMSE (W) | MAE (W) | |
---|---|---|---|---|
No clustering | 297.58 | 155.79 | 0.9685 | |
10 min | K-means clustering | 125.07 | 80.95 | 0.9944 |
iPSO based clustering | 117.11 | 79.08 | 0.9951 | |
No clustering | 345.12 | 234.57 | 0.9576 | |
30 min | K-means clustering | 135.94 | 91.20 | 0.9934 |
iPSO based clustering | 121.32 | 82.17 | 0.9948 | |
No clustering | 364.76 | 246.50 | 0.9526 | |
60 min | K-means clustering | 134.86 | 89.99 | 0.9935 |
iPSO based clustering | 127.87 | 87.21 | 0.9941 |
Performance metrics for different clustering methods (Case B).
Prediction Horizon | Clustering Method | RMSE (kW) | MAE (kW) | |
---|---|---|---|---|
No clustering | 11.84 | 5.40 | 0.9684 | |
10 min | K-means clustering | 5.52 | 3.59 | 0.9930 |
iPSO based clustering | 4.93 | 3.29 | 0.9945 | |
No clustering | 12.14 | 5.99 | 0.9665 | |
30 min | K-means clustering | 5.53 | 3.93 | 0.9930 |
iPSO based clustering | 5.16 | 3.62 | 0.9939 | |
No clustering | 13.24 | 7.61 | 0.9601 | |
60 min | K-means clustering | 6.16 | 4.13 | 0.9913 |
iPSO based clustering | 5.52 | 3.82 | 0.9936 |
Time consumption for different models (Case A, 30 min prediction).
Model | Training Time (s) | Prediction Time (s) |
---|---|---|
LSTM | 70.40 | 4.29 |
AdaLSTM | 223.68 | 13.15 |
iPSO clustering + AdaLSTM | 105.70 | 10.74 |
Appendix A
Benchmark functions.
Function | Domain |
---|---|
| [−100, 100] |
| [−10, 10] |
| [−100, 100] |
| [−100, 100] |
| [−32, 32] |
| [−600, 600] |
| |
| [−5, 5] |
|
References
1. Voyant, C.; Notton, G.; Kalogirou, S.; Nivet, M.-L.; Paoli, C.; Motte, F.; Fouilloy, A. Machine learning methods for solar radiation forecasting: A review. Renew. Energy; 2017; 105, pp. 569-582. [DOI: https://dx.doi.org/10.1016/j.renene.2016.12.095]
2. Tawn, R.; Browell, J. A review of very short-term wind and solar power forecasting. Renew. Sustain. Energy Rev.; 2022; 153, 111758. [DOI: https://dx.doi.org/10.1016/j.rser.2021.111758]
3. Huang, X.; Zhang, C.; Li, Q.; Tai, Y.; Shi, J. A Comparison of Hour-Ahead Solar Irradiance Forecasting Models Based on LSTM Network. Math. Probl. Eng.; 2020; 2020, 4251517. [DOI: https://dx.doi.org/10.1155/2020/4251517]
4. Cao, J.; Cao, S. Study of forecasting solar irradiance using neural networks with preprocessing sample data by wavelet analysis. Energy; 2006; 31, pp. 3435-3445. [DOI: https://dx.doi.org/10.1016/j.energy.2006.04.001]
5. Voyant, C.; Muselli, M.; Paoli, C.; Nivet, M.-L. Numerical Weather Prediction (NWP) and hybrid ARMA/ANN model to predict global radiation. Energy; 2012; 39, pp. 341-355. [DOI: https://dx.doi.org/10.1016/j.energy.2012.01.006]
6. Monjoly, S.; André, M.; Calif, R.; Soubdhan, T. Forecast horizon and solar variability influences on the performances of multiscale hybrid forecast model. Energies; 2019; 12, 2264. [DOI: https://dx.doi.org/10.3390/en12122264]
7. Eseye, A.T.; Zhang, J.; Zheng, D. Short-term photovoltaic solar power forecasting using a hybrid Wavelet-PSO-SVM model based on SCADA and Meteorological information. Renew. Energy; 2018; 118, pp. 357-367. [DOI: https://dx.doi.org/10.1016/j.renene.2017.11.011]
8. Olatomiwa, L.; Mekhilef, S.; Shamshirband, S. Global Solar Radiation Forecasting Based on SVM-Wavelet Transform Algorithm. Int. J. Intell. Syst. Appl. (IJISA); 2016; 8, pp. 19-26. [DOI: https://dx.doi.org/10.5815/ijisa.2016.05.03]
9. Sahu, R.K.; Shaw, B.; Nayak, J.R. Short/medium term solar power forecasting of Chhattisgarh state of India using modified TLBO optimized ELM. Eng. Sci. Technol. Int. J.; 2021; 24, pp. 1180-1200. [DOI: https://dx.doi.org/10.1016/j.jestch.2021.02.016]
10. Wang, S.; Wang, Y.; Liu, Y.; Zhang, N. Hourly solar radiation forecasting based on EMD and ELM neural network. Electr. Power Autom. Equip.; 2014; 34, pp. 7-12.
11. Zayed, M.E.; Zhao, J.; Li, W.; Elsheikh, A.H.; Abd Elaziz, M.; Yousri, D.; Zhong, S.; Mingxi, Z. Predicting the performance of solar dish Stirling power plant using a hybrid random vector functional link/chimp optimization model. Sol. Energy; 2021; 222, pp. 1-17. [DOI: https://dx.doi.org/10.1016/j.solener.2021.03.087]
12. Zayed, M.E.; Aboelmaaref, M.M.; Chazy, M. Design of solar air conditioning system integrated with photovoltaic panels and thermoelectric coolers: Experimental analysis and machine learning modeling by random vector functional link coupled with white whale optimization. Therm. Sci. Eng. Prog.; 2023; 44, 102051. [DOI: https://dx.doi.org/10.1016/j.tsep.2023.102051]
13. De Araujo, J.M.S. Combination of WRF model and LSTM network for solar radiation forecasting—Timor leste case study. Comput. Water Energy Environ. Eng.; 2020; 9, pp. 108-144. [DOI: https://dx.doi.org/10.4236/cweee.2020.94009]
14. Hong, Y.-Y.; Martinez, J.J.F.; Fajardo, A.C. Day-ahead solar irradiation forecasting utilizing gramian angular field and convolutional long short-term memory. IEEE Access; 2020; 8, pp. 18741-18753. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.2967900]
15. Ren, Y.; Wang, J.; Yang, C.; Xiao, C.; Li, S. Wind and Solar Integrated Power Prediction Method Research Based on DT-CWT and LSTM. J. Phys. Conf. Ser.; 2021; 1745, 012008. [DOI: https://dx.doi.org/10.1088/1742-6596/1754/1/012008]
16. Wang, J.; Li, J. Short-term Photovoltaic Power Forecasting Using the Combined KSVD and LSTM Method. Proceedings of the 2022 International Conference on Machine Learning, Cloud Computing and Intelligent Mining (MLCCIM 2022); Xiamen, China, 5–7 August 2022; pp. 134-138.
17. Elizabeth Michael, N.; Mishra, M.; Hasan, S.; Al-Durra, A. Short-term solar power predicting model based on multi-step CNN stacked LSTM technique. Energies; 2022; 15, 2150. [DOI: https://dx.doi.org/10.3390/en15062150]
18. Wang, K.; Qi, X.; Liu, H. Photovoltaic power forecasting based LSTM-Convolutional Network. Energy; 2019; 189, 116225. [DOI: https://dx.doi.org/10.1016/j.energy.2019.116225]
19. Zhang, C.; Yan, Z.; Ma, C.; Xu, X. Prediction of direct normal irradiation based on CNN-LSTM model. Proceedings of the 2020 5th International Conference on Multimedia Systems and Signal Processing; Chengdu, China, 8–10 May 2020; pp. 74-80.
20. Kumar, J.; Goomer, R.; Singh, A.K. Long short term memory recurrent neural network (LSTM-RNN) based workload forecasting model for cloud datacenters. Procedia Comput. Sci.; 2018; 125, pp. 676-682. [DOI: https://dx.doi.org/10.1016/j.procs.2017.12.087]
21. Chaouachi, A.; Kamel, R.M.; Nagasaka, K. Neural network ensemble-based solar power generation short-term forecasting. J. Adv. Comput. Intell. Intell. Inform.; 2010; 14, pp. 69-75. [DOI: https://dx.doi.org/10.20965/jaciii.2010.p0069]
22. Liu, J.; Huang, X.; Li, Q.; Chen, Z.; Liu, G.; Tai, Y. Hourly stepwise forecasting for solar irradiance using integrated hybrid models CNN-LSTM-MLP combined with error correction and VMD. Energy Convers. Manag.; 2023; 280, 116804. [DOI: https://dx.doi.org/10.1016/j.enconman.2023.116804]
23. Zhou, Y.; Xue, W.; Liu, J.; Li, K. Photovoltaic power prediction based on SVMD-PCA-EL model. Proceedings of the 2022 41st Chinese Control Conference (CCC); Hefei, China, 25–27 July 2022; pp. 5621-5626.
24. Benmouiza, K.; Cheknane, A. Forecasting hourly global solar radiation using hybrid k-means and nonlinear autoregressive neural network models. Energy Convers. Manag.; 2013; 75, pp. 561-569. [DOI: https://dx.doi.org/10.1016/j.enconman.2013.07.003]
25. Azimi, R.; Ghayekhloo, M.; Ghofrani, M. A hybrid method based on a new clustering technique and multilayer perceptron neural networks for hourly solar radiation forecasting. Energy Convers. Manag.; 2016; 118, pp. 331-344. [DOI: https://dx.doi.org/10.1016/j.enconman.2016.04.009]
26. Malakar, S.; Goswami, S.; Ganguli, B.; Chakrabarti, A.; Roy, S.S.; Boopathi, K.; Rangaraj, A.G. Deep-learning-based adaptive model for solar forecasting using clustering. Energies; 2022; 15, 3568. [DOI: https://dx.doi.org/10.3390/en15103568]
27. Sun, S.; Wang, S.; Zhang, G.; Zheng, J. A decomposition-clustering-ensemble learning approach for solar radiation forecasting. Sol. Energy; 2018; 163, pp. 189-199. [DOI: https://dx.doi.org/10.1016/j.solener.2018.02.006]
28. Kennedy, J.; Eberhart, R. Particle swarm optimization. Int. Conf. Neural Netw.; 1995; 4, pp. 1942-1948.
29. Bai, Q. Analysis of particle swarm optimization algorithm. Comput. Inf. Sci.; 2010; 3, 180. [DOI: https://dx.doi.org/10.5539/cis.v3n1p180]
30. Cai, L.; Hou, Y.; Zhao, Y.; Wang, J. Application research and improvement of particle swarm optimization algorithm. Proceedings of the 2020 IEEE International Conference on Power, Intelligent Computing and Systems (ICPICS); Shenyang, China, 28–30 July 2020; pp. 238-241.
31. Liang, J.J.; Qin, A.K.; Suganthan, P.N.; Baskar, S. Comprehensive learning particle swarm optimizer for global optimization of multimodal functions. IEEE Trans. Evol. Comput.; 2006; 10, pp. 281-295. [DOI: https://dx.doi.org/10.1109/TEVC.2005.857610]
32. El-Zonkoly, A.; Saad, M.; Khalil, R. New algorithm based on CLPSO for controlled islanding of distribution systems. Int. J. Electr. Power Energy Syst.; 2013; 45, pp. 391-403. [DOI: https://dx.doi.org/10.1016/j.ijepes.2012.08.076]
33. Chen, X.; Wang, B.; Yu, M.; Jin, J.; Xu, W. The interpolation of missing wind speed data based on optimized LSSVM model. Proceedings of the 2016 IEEE 8th International Power Electronics and Motion Control Conference (IPEMC-ECCE Asia); Hefei, China, 22–26 May 2016; pp. 1448-1451.
34. Wang, W.; Cheng, J.; Gao, Y.; Yu, X. A scenario reduction method of PV output based on NBSO algorithm and STDI index. J. Phys. Conf. Ser.; 2023; 2496, 012005. [DOI: https://dx.doi.org/10.1088/1742-6596/2496/1/012005]
35. Yu, Y.; Si, X.; Hu, C.; Zhang, J. A review of recurrent neural networks: LSTM cells and network architectures. Neural Comput.; 2019; 31, pp. 1235-1270. [DOI: https://dx.doi.org/10.1162/neco_a_01199]
36. Bai, Y.; Xie, J.; Wang, D.; Zhang, W.; Li, C. A manufacturing quality prediction model based on AdaBoost-LSTM with rough knowledge. Comput. Ind. Eng.; 2021; 155, 107227. [DOI: https://dx.doi.org/10.1016/j.cie.2021.107227]
37. Zhang, M.-L.; Zhou, Z.-H. ML-KNN: A lazy learning approach to multi-label learning. Pattern Recognit.; 2007; 40, 7. [DOI: https://dx.doi.org/10.1016/j.patcog.2006.12.019]
38. DKA Solar Centre. Available online: https://dkasolarcentre.com.au/download?location=alice-springs (accessed on 20 February 2024).
39. DKA Solar Centre. Available online: https://dkasolarcentre.com.au/download?location=yulara (accessed on 20 February 2024).
40. Pedro, H.T.C.; Coimbra, C.F.M. Assessment of forecasting techniques for solar power production with no exogenous inputs. Sol. Energy; 2012; 86, pp. 2017-2028. [DOI: https://dx.doi.org/10.1016/j.solener.2012.04.004]
41. Wang, Z.; Wang, F.; Su, S. Solar irradiance short-term prediction model based on BP neural network. Energy Procedia; 2011; 12, pp. 488-494. [DOI: https://dx.doi.org/10.1016/j.egypro.2011.10.065]
42. Agga, A.; Abbou, A.; Labbadi, M.; El Houm, Y.; Ali, I.H.O. CNN-LSTM: An efficient hybrid deep learning architecture for predicting short-term photovoltaic power production. Electr. Power Syst. Res.; 2022; 208, 107908. [DOI: https://dx.doi.org/10.1016/j.epsr.2022.107908]
43. Abdel-Nasser, M.; Mahmoud, K. Accurate photovoltaic power forecasting models using deep LSTM-RNN. Neural Comput. Appl.; 2019; 31, pp. 2727-2740. [DOI: https://dx.doi.org/10.1007/s00521-017-3225-z]
44. Wei, B.; Li, K.; Zhou, S.; Xue, W.; Tan, G. An instance based multi-source transfer learning strategy for building’s short-term electricity loads prediction under sparse data scenarios. J. Build. Eng.; 2024; 85, 108713. [DOI: https://dx.doi.org/10.1016/j.jobe.2024.108713]
45. Asuero, A.G.; Sayago, A.; González, A.G. The correlation coefficient: An overview. Crit. Rev. Anal. Chem.; 2006; 36, pp. 41-59. [DOI: https://dx.doi.org/10.1080/10408340500526766]
46. Lin, A.; Sun, W.; Yu, H.; Wu, G.; Tang, H. Adaptive comprehensive learning particle swarm optimization with cooperative archive. Appl. Soft Comput.; 2019; 77, pp. 533-546. [DOI: https://dx.doi.org/10.1016/j.asoc.2019.01.047]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Due to the increasing integration of photovoltaic (PV) solar power into power systems, the prediction of PV solar power output plays an important role in power system planning and management. This study combines an optimized data clustering method with a serially integrated AdaLSTM network to improve the accuracy and robustness of PV solar power prediction. During the data clustering process, the Euclidean distance-based clustering centroids are optimized by an improved particle swarm optimization (iPSO) algorithm. For each obtained data cluster, the AdaLSTM network is utilized for model training, in which multiple LSTMs are serially combined together through the AdaBoost algorithm. For PV power prediction tasks, the inputs of the testing set are classified into the nearest data cluster by the K-nearest neighbor (KNN) method, and then the corresponding AdaLSTM network of this cluster is used to perform the prediction. Case studies from two real PV stations are used for prediction performance evaluation. Results based on three prediction horizons (10, 30 and 60 min) demonstrate that the proposed model combining the optimized data clustering and AdaLSTM has higher prediction accuracy and robustness than other comparison models. The root mean square error (RMSE) of the proposed model is reduced, respectively, by 75.22%, 73.80%, 67.60%, 66.30%, and 64.85% compared with persistence, BPNN, CNN, LSTM, and AdaLSTM without clustering (Case A, 30 min prediction). Even compared with the model combining the K-means clustering and AdaLSTM, the RMSE can be reduced by 10.75%.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer