This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
The probability screen is a typical nonlinear, non-Gaussian, time-varying multiinput and multioutput vibration system, manifesting different vibration signal outputs in varying operational conditions. Screening efficiency is an important indicator for gauging the overall performance of the screening system. However, numerous factors influence the system’s screening efficiency, including vibration frequency, amplitude, inclination, and various operational conditions, alongside structural characteristics, material properties, and other relevant factors [1]. Currently, some scholars use simulation software to simulate the screening structure characteristics by establishing mechanical models. This approach enables the exploration of how screening parameters impact screening efficiency [2]. Li and Huang [3], Shi and Huang [4], Tang [5], and Zheng and Huang [6] delved into the influence of kernel parameters in least squares support vector machine (LS-SVM) on the prediction screening efficiency for probability screens. Their research also encompassed screening efficiency prediction through SVM based on higher-order cumulant auto-regressive (AR) models and examined the effect of vibration parameters on the system’s screening efficiency by studying the Wigner higher-order spectrum characteristics of vibration signal time series. In these studies, feature extraction and machine learning techniques for probability screen time series were relatively straightforward, primarily focusing on features such as AR model coefficients and fractal dimensions of time series or time series as feature vectors that have large dimensions or single input. Besides, SVM was predominantly chosen as the machine learning method. Consequently, ample room remains for advancing time series feature extraction, classification prediction methodology, and prediction efficiency enhancement within the domain of probability screen vibration systems. Selecting the feature extraction and prediction algorithm suitable for the system’s screening efficiency, along with the corresponding model parameter optimization, continues to be a pivotal area of research in the realm of probability screen studies.
In recent years, a growing body of research has explored time series feature extraction and prediction algorithms, with many scholars achieving promising results in their research on state feature extraction, classification prediction, equipment fault diagnosis, state recognition monitoring, target detection, and other aspects through the combination of empirical mode decomposition energy entropy (EMD-EE) of signals and machine learning methods such as SVM [7, 8]. Additionally, the wavelet packet energy entropy (WP-EE) of a signal has emerged as a characteristic parameter of the state, often paired with optimization methods like the gray wolf optimizer-SVM, back propagation neural network (BPNN), LSSVM, and particle swarm optimization. This approach has been applied to the fields of mechanical equipment, medical health, and power systems, for tasks like fault diagnosis, feature parameter extraction, and state recognition research [9, 10, 11]. Generalized regressive neural network (GRNN) is a radial basis function network based on mathematical statistics. It has robust nonlinear mapping ability and rapid learning, making it particularly well-suited for scenarios with limited sample sizes [12]. GRNN models have been widely used in state prediction, pattern recognition, fault diagnosis, and other areas. For instance, based on a study [13] comparing several different time series feature extraction methods, BPNN and GRNN were selected to predict and compare the damping efficiency of the magnetorheological system, with GRNN model prediction achieving favorable results. Kernel principal component analysis (KPCA) is a nonlinear method to achieve data dimensionality reduction or feature extraction, effectively reduce vector dimensions, and reconstruct feature vectors [14]. When combined with machine learning, KPCA can reduce the dimensions of feature vectors through parameter optimization, while improving computational efficiency in prediction and classification tasks, all while ensuring prediction accuracy. Despite these advancements, there is a dearth of literature concerning the integration of WP-EE with GRNN [15]. Furthermore, there is a notable absence of research literature on the combination of WP-EE and GRNN applied to probability screens, as well as the potential of WP-EE and GRNN-KPCA.
This paper focuses on the experimental prototype of the probability screen as its research subject. It conducts tests and obtains the time series and screening efficiency of the system vibration signal under various operational conditions, calculates the energy entropy of the decomposed signal by wavelet packet decomposition (WPD) under different decomposition layers and wavelet basis functions, and computes the energy entropy with EMD and VMD. This study also uses a GRNN model to predict the screening efficiency of its single input characteristics and conducts a comparative analysis of the prediction performance between d-GRNN, d-BPNN, d-Weighted LSSVM, WP-d-GRNN, WP-EE-BPNN, EMD-EE-GRNN, variational mode decomposition energy entropy (VMD-EE)-GRNN, and WP-EE-GRNN. By analyzing the prediction effect of single-input WP-EE-GRNN, the WPD decomposition scheme is extracted for feature vector reconstruction, and the second feature reconstruction is then performed through KPCA dimensionality reduction. On the basis of setting different WPD parameters, GRNN parameters, and KPCA parameters, the KPCA-WP-EE-GRNN precalculation method is established to predict the screening efficiency, analyze the prediction effect, and select the optimal parameter combination scheme suitable for the probability screen system’s prediction algorithm.
2. Basic Theory
2.1. WP-EE
WPD involves projecting a time series signal into the space of the base function and decomposing the subsignals of high- and low-frequency components through a series of filters with different center frequencies but the same bandwidth, which can decompose the signal in multiple layers. WPD is a decomposition method without redundancy and omission [9, 10, 11].
For a time series signal x (t), given the scale function φ (t) and wavelet function ∅ (t), the two-scale equation of wavelet packet transform is as follows [11, 12, 16]:
Information entropy can reflect the uncertainty of a signal or system or the complexity of a random signal [8]. The information entropy of the total energy of each subsignal obtained by WPD is called WP-EE [8]. Different WP-EE will be obtained under different WPD layers and wavelet basis functions. The WP-EE of a signal is then expressed as follows [9, 11]:
2.2. GRNN
GRNN is a radial basis function network based on mathematical statistics. Its theoretical basis is nonlinear regression analysis. Its main structure includes four layers: input, mode, summation, and output layers. The smoothing factor Spread is an important parameter of GRNN models. The model’s output can be typically changed by adjusting the smoothing factor value [13].
2.3. KPCA
KPCA is based on the principle of kernel functions. It involves projecting the input space into the high-dimensional space through nonlinear mapping and then conducting principal component analysis on the mapped data within the high-dimensional space. This method possesses robust nonlinear processing ability. Typically, KPCA is applied by adjusting kernel function parameters σ2 to obtain different feature vectors and assess the contribution rate of each principal component. The principal components with high contribution rates are selected as the features for analysis, thereby accomplishing dimensionality reduction [14].
3. Data Source and Processing
3.1. Probability Screen Structure and Vibration Test
This experiment uses the probability screen experimental machine (Figure 1) developed by Huang Yijian’s team. Its structure comprises a screen, feed inlet, discharge outlet, screen body, vibration excitation motor, support, and other components. The experimental test system consists of a computer, a data acquisition card (PCI6014), a piezoelectric acceleration sensor (HK9103), a charge amplifier (HK9205), and a LabVIEW software environment. Factors affecting screening efficiency during operation encompass material properties, screen surface inclination, feed speed, screen amplitude, vibration frequency, and more. This experiment mainly considers the impact of screen amplitude A and excitation frequency f on screening efficiency. The vibration test is conducted at A = 3, 4, 5, 6, and 7 mm, f = 15, 20, 25, 30, 35, and 40 Hz, and the two parameters are combined in 30 working states, respectively. The signal sampling frequency is set to fs = 1,000 Hz, and the time series of vibration acceleration in the Z direction and corresponding screening efficiency are obtained. The screen size in the experiment measures 1.0 mm × 1.0 mm, and the selected material has a particle size of 0.6 mm [1, 5].
[figure(s) omitted; refer to PDF]
3.2. WP-EE Characteristics of Time Series
For the detected vibration signal time series, 1,024 test values from the stable signal section are chosen for analysis under each working condition. For the extracted time series analysis segment signals, WPD is applied across decomposition layers 1–6. The wavelet basis functions used include Daubechies wavelets (db1–5), Biological wavelets (bior2.2, 1.3, 1.5), Cofflet wavelets (coif1–5), Symlets wavelets (sym1–5), Fejer–Korovkin orthogonal wavelets (fk14, 8, 6, 4), and Discrete Meyer wavelets (dmey). WPD is employed to extract and decompose the wavelet coefficients of each sub-signal, followed by the calculation of the information entropy of its total energy. Figure 2 shows the WP-EE of the vibration signal of the probability screen at A = 3 mm and f = 15 Hz. It illustrates that WP-EE exhibits a consistent trend in its variation with the number of decomposition layers s. A greater number of decomposition layers s results in higher WP-EE values, reflecting increased information about the complexity of the state signal. For the same number of layers s, WP-EE values differ significantly across various wavelet basis functions. The energy entropy values corresponding to Daubechies wavelets are relatively large, followed by the energy entropy values corresponding to Biorthogonal wavelets, and the energy entropy values corresponding to the other several basis functions are relatively close. Furthermore, the WP-EE values calculated under various operational conditions exhibit substantial numerical variations. This disparity effectively captures the distinctions in signal complexity across various working conditions, serving as a significant characteristic of the system’s state.
[figure(s) omitted; refer to PDF]
4. Prediction Algorithm Design
4.1. KPCA-GRNN Prediction Algorithm Based on WP-EE
Considering the distinct characteristics of WP-EE in different states, it is proposed to obtain the energy entropy reconstruction feature vector through the WPD of time series and use the KPCA-GRNN model to predict and analyze the probability screening efficiency. The steps of the algorithm are as follows:
(1) Extract and analyze the signal. Detect the time series under different operating conditions and screening efficiency and select a stable signal segment for analysis.
(2) Compute the WP-EE. WPD is applied on the time series of each state analysis signal considering different layers and wavelet basis functions. The wavelet decomposition coefficients are obtained for each node along with the total energy and its information entropy for each coefficient resulting from the time series decomposition under different states.
(3) Set up training and test sample sets. The WP-EE of time series in different states is taken as the single input eigenvalue, and the corresponding screening efficiency is taken as the state output value to form 30 sample sets. The first N (N ranging from 16 to 25) samples constitute the training set, and the last five samples represent the test set.
(4) GRNN model prediction. Set the training sample number N and smoothing factor Spread range of the GRNN model, predict the screening efficiency with the GRNN model, and obtain the mean absolute value of the relative error between the predicted and experimental values of five test samples, which is called the absolute average error
For the same number of training samples N, the minimum absolute average error obtained from different m (Spread) values is designated as the prediction error
The average of each prediction error R for different training sample numbers N (N ranging from 16 to 25) is termed the average prediction error
(5) The WP-EE eigenvector of the reconstructed features is optimized. Compare and analyze the prediction accuracy of WP-EE under different decomposition levels and wavelet basis functions. Select WP-EE for the reconstruction feature vector when the average prediction error
(6) KPCA and secondary feature vector reconstruction. KPCA is performed on the reconstructed energy entropy feature vector, and the kernel function parameters σ2 are adjusted. Determine the contribution rate of each principal component, take the first several principal components with a high contribution rate, and reduce the dimensions to reconstruct the feature vector twice.
(7) KPCA-GRNN prediction and optimal prediction parameter combination determination. Use different parameter values for σ2. Input the second reconstruction feature vector after dimensionality reduction to GRNN prediction according to step (4). Select the best prediction result and the corresponding parameter combination scheme based on the prediction result analysis.
In order to clearly illustrate the above algorithm process of the KPCA-GRNN prediction algorithm based on WP-EE, the algorithm flowchart is drawn in Figure 3.
[figure(s) omitted; refer to PDF]
4.2. Comparison Prediction Algorithm Design
For comparative analysis, the GRNN, BPNN, and weighted LSSVM probability screening prediction methods based on time-series box dimension feature extraction in reference [18] were used to compare the prediction accuracy of screening efficiency. Then, the WP-d-GRNN, WP-EE-BPNN, EMD-EE-GRNN, and VMD-EE-GRNN prediction algorithms with single feature input are designed and compared with the WP-EE-GRNN prediction algorithm with single feature input. The WP-d-GRNN, WP-EE-BPNN, and EMD-EE-GRNN prediction algorithm are carried out according to steps (1)–(4) in Section 4.1, where the EE in steps (2) and (3) is replaced with box dimension (d) for the WP-d-GRNN, the GRNN in steps (4) is replaced with BPNN for WP-EE-BPNN, the WP-EE in steps (2) and (3) is replaced with EMD-EE for the EMD-EE-GRNN. The algorithm flow of the EMD-EE-GRNN prediction algorithm is illustrated in Figure 4. The VMD-EE-GRNN prediction algorithm follows steps (1) and (4) in Section 4.1, while in steps (2) and (3), the VMD parameter penalty factor M value is initially set, followed by configuring the range of the decomposition number K, and obtaining the energy entropy of the corresponding decomposition signal using VMD, instead of WP-EE. After step (4), an additional step (5) is introduced: obtain the prediction error corresponding to different K values, select the K value with the lowest average prediction error under varying N values, and then define the M range. Subsequently, repeat steps (2)–(4) to obtain the minimum prediction error of the VMD-EE-GRNN model across different M values. The flowchart of the VMD-EE-GRNN prediction algorithm is illustrated in Figure 5. It is worth noting that the time series data selected for all the three comparison algorithms are the same, and the parameter range set during the GRNN model prediction within the dashed box is consistent. Finally, the prediction performance is assessed by comparing the prediction error of the three algorithms.
[figure(s) omitted; refer to PDF]
5. Prediction Result Analysis
5.1. Analysis of Prediction Results of Three Algorithms with Single Feature Input
The EMD-EE is obtained by calculating the modal components, envelope energy, and total energy information entropy of the time series across 30 operational conditions using EMD, followed by prediction with the GRNN model. In the case of VMD, the procedure begins with setting M = 99 and K between 1 and 15. The energy information entropy of each mode component, envelope energy, and total energy are computed after applying VMD, and VMD-EE is then derived. After the initial prediction using the GRNN model, the value of K that yields the smallest average prediction error is determined, and K is then set accordingly. With M = 1 : 10 : 200, the energy entropy is then computed after VMD, and a second prediction is carried out using the GRNN model. For the WP-EE of 30 states obtained by WPD under the decomposition levels s = 1–6 and the DB3 wavelet basis function applied to the time series data, the GRNN model is employed for prediction. The energy entropy obtained by three different methods is used as a single input characteristic value, and the screening efficiency is used as the output value to construct 30 samples. The first N samples, with N ranging from 16 to 25, constitute the training sample set, and the last five samples are used as the test set. Different Spread values are set accordingly. GRNN is then used for prediction, with the prediction error R determined for various numbers N of training samples. A graphical representation of the R–N curve is displayed in Figure 6, showcasing the prediction performance of WP-d-GRNN, WP-EE-BPNN, EMD-EE-GRNN, VMD-EE-GRNN, and WP-EE-GRNN when decomposing layers 1–6 and using DB3 as the basis function. Chen and Huang [18] used three methods, d-GRNN, d-BPNN, and d-Weighted LSSVM, to predict screening efficiency under different training sample sizes. The predicted results are listed in Table 1.
[figure(s) omitted; refer to PDF]
Table 1
Prediction results of screening efficiency under different training samples.
The number of training samples | Minimum prediction error R (%) based on box-dimension d [18] | Minimum prediction error R (%) based on EE | |||||||
d - GRNN | d -BPNN | d-Weighted-LSSVM | WP-d-GRNN | WP-EE-BPNN | EMD-EE-GRNN | VMD-EE-GRNN | WP-EE-GRNN | KPCA-WP-EE-GRNN | |
16 | 5.35 | 5.41 | 6.38 | 3.90 | 4.59 | 5.77 | 4.57 | 2.48 | 1.62 |
17 | 5.67 | 5.21 | 6.62 | 2.43 | 4.57 | 5.48 | 4.69 | 1.80 | 1.64 |
18 | 5.41 | 5.51 | 6.20 | 1.75 | 4.15 | 5.48 | 5.13 | 1.45 | 1.42 |
19 | 5.85 | 5.29 | 6.42 | 1.79 | 4.24 | 5.48 | 5.05 | 1.34 | 0.78 |
20 | 5.84 | 5.20 | 6.05 | 1.81 | 3.43 | 5.48 | 5.53 | 1.99 | 0.71 |
21 | 5.75 | 5.46 | 5.49 | 2.85 | 2.83 | 5.48 | 2.66 | 1.92 | 0.76 |
22 | 5.46 | 4.20 | 5.33 | 3.44 | 3.05 | 5.71 | 2.66 | 1.92 | 0.95 |
23 | 5.55 | 3.28 | 4.14 | 3.22 | 1.85 | 5.72 | 2.66 | 1.65 | 0.95 |
24 | 5.10 | 3.45 | 4.40 | 1.77 | 1.57 | 5.70 | 2.66 | 1.51 | 0.92 |
25 | 6.04 | 3.62 | 4.80 | 3.03 | 1.57 | 5.72 | 2.66 | 2.03 | 0.92 |
Average | 5.60 | 4.66 | 5.58 | 2.60 | 3.18 | 5.60 | 3.83 | 1.81 | 1.07 |
From the data in Table 1, it can be seen that the minimum prediction error values of the d-GRNN method for screening efficiency under different sample sizes are between 5.10% and 6.04%, with an average value of 5.60%. The minimum prediction error value of the d-BPNN method for screening efficiency under different sample sizes is 3.28%–5.51%, with an average value of 4.66%. The minimum prediction error values of the d-weighted LSSVM method for screening efficiency under different sample sizes are all between 4.14% and 6.62%, with an average value of 5.58%.
Figure 6 provides several noteworthy insights. First, the EMD-EE-GRNN prediction error R demonstrates a slight upward trend with varying N (16–25), ranging from 5.48% to 5.7%, with an average prediction error Rav of 5.60%, reaching its minimum of 5.48% when N = 17–21. Second, in VMD-EE-GRNN predictions, when M = 99, the average prediction error varies from 5% to 7% for different values of K, with the lowest average prediction error Rav = 5.15% found at K = 4. With K = 4, the VMD-EE-GRNN prediction yields an average error ranging from 4.5% to 7% under different M and N values, with the smallest average prediction error of 4.57% achieved for M = 50 and the minimum prediction error dropping to 2.66% within N = 21 to 25. While VMD-EE-GRNN offers smaller average and minimum prediction errors compared to EMD-EE-GRNN, the prediction process is more complex. Third, the WP (DB3)-d-GRNN prediction error R demonstrates a slight upward trend with varying N (16–25), ranging from 1.75% to 3.90%, with an average prediction error Rav of 2.60%, reaching its minimum of 1.75% when N = 18 and s = 3. Fourth, the WP (DB3)-EE-BPNN prediction error R demonstrates a slight upward trend with varying N (16–25), ranging from 1.57% to 4.59%, with an average prediction error Rav of 3.18%, reaching its minimum of 1.57% when N = 24–25 and s = 3. Finally, the error R for WP-EE (DB3)-GRNN prediction is mostly stable or declining with N but varies significantly across different layers, with prediction errors averaging between 1.90% and 6.17%. Notably, when s = 6, the prediction error is relatively stable and small, and Rav = 1.91%. When N = 21 and s = 3, the prediction error reaches the minimum error value of 1.34%.
Overall, the WP-EE-GRNN prediction algorithm consistently demonstrates higher prediction accuracy compared to d-GRNN, d-BPNN, d-Weighted LSSVM, WP-d-GRNN, WP-EE-BPNN, EMD-EE-GRNN, and VMD-EE-GRNN, showcasing its superior predictive capabilities.
5.2. KPCA-GRNN Prediction Results Based on WP-EE
5.2.1. Analysis of WP-EE-GRNN Prediction Results
Different wavelet basis functions, including Daubechies wavelets (db1–5), Coiflet wavelets (coif1–5), Symlets wavelets (sym1–5), Biometric wavelets Reverser (bior2.2, 1.3, 1.5), Biorthogonal (rbio2. 2, 1.3, 1.5), Fejer-Korovkin orthometric wavelets (fk14,8,6,4), and Discrete Meyer wavelets (dmey), were selected, resulting in 26 types of WPD when S = 1–6. This process yields 156 groups of WP-EE. By adjusting the Spread value of GRNN model parameters predicting with GRNN, prediction errors were determined for various numbers of training samples N. The average prediction errors corresponding to N values between 7 and 25 were calculated, and the relationship curve between the average prediction error and the number of layers s was visualized under each wavelet basis function, as illustrated in Figure 7.
[figure(s) omitted; refer to PDF]
Figure 7 reveals that the average prediction error value Rav with the number of decomposition layers s presents an obvious nonlinear relationship, but the changing trend of the relationship curve under different wavelet basis functions remains relatively consistent. Notably, the average prediction error is generally small when s = 3 and s = 6, with Rav ranging between 1.8% and 6.7%.
5.2.2. WP-EE Reconstruction Feature Vector-GRNN Prediction
To further improve the prediction accuracy, the 14 WP-EE reconstructions with small average prediction error (1.8%–2.1%) are taken as the state eigenvectors, and the screening efficiency forms a 30 × 15 sample set. The previous N = 7–25 samples are the training set, and the last five samples are the test set. The optimization prediction is carried out by setting distinct GRNN model parameter Spread values. The R–N curve is presented in Figure 8. It is evident that the prediction error after reconstruction decreases with the increase in the number of training samples N, with error values ranging between 1.25% and 2.29%. The change is relatively stable, and the average prediction error Rav is 1.703%. When N = 24 and 25, the prediction error achieves a minimum of 1.264%. In comparison to using a single WP-EE input without reconstruction, the prediction error is significantly reduced, and the sensitivity to the number of training samples is substantially diminished. However, since the feature vector comprises 14-dimensions, the prediction requires a significant amount of computational time.
[figure(s) omitted; refer to PDF]
5.2.3. KPCA Dimensionality Reduction and Feature Vector Secondary Reconstruction
The WP-EE feature vector reconstruction process expands the original 1-dimensional vector to a 14-dimensional vector, significantly increasing the computational time required for prediction. To reduce the dimensions while improving prediction efficiency and enhancing prediction accuracy, the KPCA method is adopted, along with the RBF kernel function. The kernel function parameters σ2 are calculated and adjusted, taking into consideration the contribution rate of each principal component. Subsequently, the first n eigenvectors with the highest contribution rate are selected for secondary reconstruction, effectively achieving dimensionality reduction. Parameter adjustments involve varying σ2 and extracting 1, 2, 3, 4, and 5 principal components for further feature vector reconstruction to optimize predictions using the GRNN model.
5.2.4. KPCA-WP-EE-GRNN Prediction Results
To evaluate the KPCA-WP-EE-GRNN prediction algorithm, we extracted the first n (ranging from 1 to 5) principal components and reconstructed the resulting n feature vectors obtained by KPCA. Subsequently, we established as ample set and adjusting the KPCA kernel function parameters σ2 = 0.05 : 0.2 : 2.85, the GRNN model parameter Spread = 0.001 : 0.05 : 10, and the number of training samples N = 7–25. Following the KPCA-WP-EE-GRNN prediction algorithm process, we calculated the average prediction error for different N values and different n numbers of principal components. Additionally, we employed 14 eigenvectors as input and maintained consistent parameter settings for σ2 and Spread for GRNN prediction. Comparing the average prediction error before and after the second reconstruction, we plotted the R–N curve (as depicted in Figure 8) to analyze the variance in prediction outcomes. We also considered principal component contribution rates, prediction average errors, minimum errors, and the optimal σ2 for the prediction algorithm. Prediction results of screening efficiency under different training samples, such as average prediction error under different N (Rav), the minimum prediction error under different N (Rmin), and the minimum RRMSE under different N (RRMSEmin). The minimum root mean square error (RMSE) under different N (RMSEmin) and Pearson correlation coefficient, the relevant parameters related to the algorithm before and after KPCA dimensionality reduction and reconstruction, such as the optimal Spread and the optimal number of training samples N, are listed in Table 2.
Table 2
Prediction results and optimal parameters related to the algorithm before and after KPCA dimensionality reduction and reconstruction.
Feature reconstruction | Prediction results | Model parameters with minimum error | ||||||
Average prediction error under different N Rav (%) | The minimum prediction error under different N Rmin (%) | The minimum RRMSE under different N RRMSEmin (%) | The minimum RMSE under different N RMSEmin (%) | Pearson correlation coefficient | KPCA parameter σ2 | Spread value | Number of training samples | |
14 energy entropy | 1.703 | 1.264 | 1.274 | 1.051 | 0.988 | — | 0.001 | 24, 25 |
Extracting 5 principal components | 1.685 | 0.979 | 1.089 | 0.907 | 0.993 | 0.85 | 0.101 | 25 |
Extracting 4 principal components | 1.666 | 1.004 | 1.103 | 0.924 | 0.994 | 2.25 | 0.101 | 24, 25 |
Extracting 3 principal components | 1.614 | 0.996 | 1.219 | 1.030 | 0.995 | 1.05 | 0.151 | 25 |
Extracting 2 principal components | 1.434 | 0.708 | 1.166 | 0.836 | 0.997 | 0.85 | 0.051 | 19 |
Extracting 1 principal components | 1.703 | 1.301 | 1.454 | 1.209 | 0.996 | 0.25 | 0.051 | 25 |
Tables 1 and 2 and Figure 8 reveal that the R–N curve, representing the input of the secondary reconstruction feature vector after the principal component dimensionality reduction extracted by KPCA, is basically consistent with the trend before dimensionality reduction. It also exhibits a stepwise decline as N increases. However, there are notable improvements in Rav, Rmin, RRRMSEmin, RRMSEmin, and Pearson correlation coefficient compared to before dimensionality reduction. The prediction outcomes display a certain relationship with the number of extracted principal components. Particularly, when extracting 2, 3, 4, and 5 principal components, there is a significant reduction in Rav, Rmin, RRMSEmin, RMSEmin, and Pearson correlation coefficient compared to the results before dimensionality reduction. Extracting two principal components yields the lowest Rav at 1.434%, Rmin at 0.708%, and RMSEmin at 0.836%, marking the lowest values achieved, while the Pearson correlation coefficient at 0.997, marking the closest to 1. In this scenario, the corresponding model parameters are as follows: KPCA parameters σ2 = 0.85, the optimal parameter of GRNN model Spread = 0.051, and an optimal number of training samples N = 19. These parameters constitute the optimal parameter combination scheme for the model. The application of KPCA dimensionality reduction and secondary reconstruction of feature vectors not only reduces prediction time but also enhances prediction accuracy.
6. Conclusion
Through the design of the KPCA-GRNN prediction algorithm and comparative analysis with WP-EE-based algorithms, several key conclusions can be drawn regarding feature extraction from probability screen vibration signal and screening efficiency prediction:
(1) Compared with d-GRNN, d-BPNN, d-WeightedLSSVM, WP-d-GRNN, WP-EE-BPNN, EMD-EE-GRNN, and VMD-EE-GRNN algorithms, the WP-EE-GRNN algorithm exhibits the highest prediction accuracy for screening efficiency in general. Furthermore, the prediction error for both algorithms either decreases or remains stable as the number of training samples increases.
(2) Using the WP-EE-GRNN algorithm, the relationship between the predicted average error value and the number of training samples N and decomposition layers s demonstrates a clear nonlinear pattern. Notably, most of the predicted average errors are minimal when the number of layers s = 3 or 6.
(3) The optimized feature vector for GRNN prediction reconstruction, derived from WP-EE under different decomposition levels and basis functions, outperforms GRNN with a single energy entropy input feature. However, it does entail increased computational time.
(4) The WP-EE-GRNN prediction algorithm, incorporating KPCA dimensionality reduction and secondary reconstruction, not only achieves superior prediction accuracy compared to pre-KPCA dimensionality reduction but also reduces prediction time and improves prediction efficiency.
(5) The KPCA-GRNN prediction algorithm based on WPD energy entropy demonstrates a high level of prediction accuracy and facilitates the selection of an optimal model parameter combination for screening efficiency prediction in probability screens.
Thanks for the strong support, guidance and help of Huang Yijian’s team from Huaqiao University.
Acknowledgments
This study was funded by the Guiding Science Project of Fujian Province (2018H0031 and 2021H0059); team construction funds for advanced materials and laser processing of Putian University; Fujian Province Key Laboratory of CNC Machine Tools and Intelligent Manufacturing (KJTPT2019ZDSYS02020063).
[1] Q. Tang, Y. J. Huang, "Analysis of probability screen efficiency using bispectrum estimation based on AR model," Journal of Huaqiao University (Natural Science), vol. 32 no. 3, pp. 253-257, 2011.
[2] J. Qiao, C. Duan, Y. Zhao, H. Jiang, H. Diao, "Study on screening efficiency of banana vibrating screen based on 3D DEM simulation," pp. 1265-1275, DOI: 10.1007/978-981-10-1926-5_130, .
[3] Z. B. Li, Y. J. Huang, "Analysis of screen vibration signals of probability sieve using AR bispectrum and its diagonal slices," Mechanical Science and Technology for Aerospace Engineering, vol. 31 no. 1, pp. 113-117, DOI: 10.13433/j.cnki.1003-8728.2012.01.019, 2012.
[4] Z. Z. Shi, Y. J. Huang, "Research on screening efficiency based on AR model of high-order cumulant LS-SVM," China Mechanical Engineering, vol. 22 no. 16, pp. 1965-1969, 2011.
[5] Q. Tang, Study on Time Frequency Characteristics of Wigner Higher Order Spectrum and Its Application in Screening Operation, 2011.
[6] G. X. Zheng, Y. J. Huang, "Since the synchronous performance testing research," Mechanical Design and Manufacturing, vol. 28 no. 7, 2010.
[7] Z. Yang, C. Kong, Y. Wang, X. Rong, L. Wei, "Fault diagnosis of mine asynchronous motor based on MEEMD energy entropy and ANN," Computers & Electrical Engineering, vol. 92,DOI: 10.1016/j.compeleceng.2021.107070, 2021.
[8] X. Chen, Y. Yang, Z. Cui, J. Shen, "Vibration fault diagnosis of wind turbines based on variational mode decomposition and energy entropy," Energy, vol. 174 no. 3, pp. 1100-1109, DOI: 10.1016/j.energy.2019.03.057, 2019.
[9] X. Liu, J. Li, B. Shi, G. Ding, F. Dong, Z. Zhang, "Intelligent detection technology for leakage bag of baghouse based on distributed optical fiber sensor," Optical Fiber Technology, vol. 52,DOI: 10.1016/j.yofte.2019.101947, 2019.
[10] Y. Hao, L. Zhu, B. Yan, S. Qin, D. Cui, H. Lu, "Milling chatter detection with WPD and power entropy for Ti-6Al-4V thin-walled parts based on multi-source signals fusion," Mechanical Systems and Signal Processing, vol. 177,DOI: 10.1016/j.ymssp.2022.109225, 2022.
[11] X. J. Zhang, Y. H. Ding, L. Y. Huang, X. F. Cheng, Automatic Classification Method of MEG Based on Wavelet Packet and Energy Entropy, 2016.
[12] Z. Zhang, S. Wang, J. Fu, "Application of improved GRNN algorithm for task man-hours prediction in metro project," Signal and Information Processing, Networking and Computers, vol. 917, pp. 1421-1430, DOI: 10.1007/978-981-19-3387-5_169, 2023.
[13] C. Yi-ze, C. Qing-tang, "State prediction of MR system by VMD-GRNN based on fractal dimension," Advances in Mechanical Engineering, vol. 14 no. 12,DOI: 10.1177/16878132221145899, 2022.
[14] F. Yang, Z. Ma, M. Xie, "Image classification with parallel KPCA-PCA network," Computational Intelligence, vol. 38 no. 2, pp. 397-415, DOI: 10.1111/coin.12503, 2022.
[15] H. Chen, P. D. S. Assala, Y. Cai, P. Yang, "Intelligent transient overvoltage location in distribution systems using wavelet packet decomposition and general regression neural networks," IEEE Transactions on Industrial Informatics, vol. 12 no. 5, pp. 1726-1735, DOI: 10.1109/TII.2016.2520909, 2016.
[16] S. G. Chickaramanna, S. T. Veerabhadrappa, P. M. Shivakumaraswamy, S. N. Sheela, S. K. Keerthana, U. Likith, L. Swaroop, V. Meghana, "Classification of arrhythmia using machine learning algorithm," Revue d’intelligence Artificielle, vol. 36 no. 4, pp. 529-534, 2022.
[17] X. W. Zeng, W. M. Zhao, H. K. Shi, Z. R. Li, "Selection of wavelet basis function in process of time-frequency analysis of earthquake signals using wavelet packet transform," Journal of Seismological Research, vol. 33 no. 4, pp. 323-328, 2010.
[18] Q. Chen, Y. Huang, "Prediction of comprehensive dynamic performance for probability screen based on AR model-box dimension," Journal of Measurements in Engineering, vol. 11 no. 4, pp. 525-535, DOI: 10.21595/jme.2023.23522, 2023.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2024 Qingtang Chen and Yijian Huang. This is an open access article distributed under the Creative Commons Attribution License (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. https://creativecommons.org/licenses/by/4.0/
Abstract
The screening system is a nonlinear and non-Gaussian complex system. To better characterize its attributes and improve the prediction accuracy of screening efficiency, this study involves the acquisition of the vibration signals and screening efficiency data under various operational conditions. Subsequently, empirical mode decomposition energy entropy (EMD-EE), variational mode decomposition energy entropy (VMD-EE), and wavelet packet energy entropy (WP-EE) features are extracted from the time series vibration signals, and three single input energy entropy-generalized regressive neural network (GRNN) prediction accuracy models are established and compared. Furthermore, we introduce the kernel principal component analysis (KPCA)-WP-EE feature reconstruction-GRNN prediction algorithm. This approach involves reconstructing the feature vector by optimizing WP-EE-GRNN prediction under varying parameters. The parameterized GRNN model is then predicted and analyzed through secondary reconstruction involving KPCA dimensionality reduction features. The results show that WP-EE-GRNN achieves superior prediction accuracy compared to box dimension (d)-GRNN, box dimension-back propagation neural network (BPNN), and d-weighted least squares support vector machine, WP-d-GRNN, WP-EE-BPNN, EMD-EE-GRNN, and VMD-EE-GRNN. Additionally, the WP-EE feature reconstruction-GRNN algorithm exhibits higher prediction accuracy than the single-input WP-EE-GRNN algorithm. The WP-EE-GRNN prediction algorithm using KPCA dimensionality reduction and secondary reconstruction not only achieves higher prediction accuracy than prior to KPCA dimensionality reduction but also improves prediction efficiency. Following the extraction of two core principal components, model parameters when KPCA’s σ2 = 0.85, the optimal parameter of GRNN model Spread = 0.051, and the optimal number of training samples N = 19, the average prediction error is 1.434%, the minimum prediction error reaching 0.708%, the minimum root mean square error reaching 0.836% and Pearson correlation coefficient marking the closest to 1, these result all representing the optimum achievable values. The budget model selects the optimal parameter combination scheme for the system.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer