1. Introduction
As one of the key technologies of real-time traffic signal control, traffic assignment, route guidance, and other functions in the intelligent transportation system, short-term traffic flow prediction has always been the research focus. Its forecasting accuracy plays a decisive role in improving the performance of the intelligent transportation system [1]. For pursuing higher accuracy, a variety of spatio-temporal forecasting methods have been developed [2,3]. Among them, the temporal forecasting methods are widely used and have attracted more and more attention in the recent decades. Generally, these methods are roughly divided into three categories, i.e., statistical theoretical models, intelligent models, and hybrid models.
Statistical methods mainly include time-series models (e.g., autoregressive integrated moving average (ARIMA), seasonal autoregressive integrated moving average (SARIMA), etc.) [4,5,6], Kalman filtering model [7,8], and history average model [9]. Among them, the time series model has been widely applied in the prediction of traffic volume data. For example, Kumar et al. [4] developed a SARIMA model to predict the traffic flow, in which the order of model was determined by autocorrelation function and partial autocorrelation function. The forecasting results showed that the proposed model had satisfactory forecasting accuracy. Zhao et al. [5] proposed a short-term traffic flow forecasting model combined the ARIMA model and the space-time characteristics of the expressway network to improve forecasting accuracy. Wang et al. [6] adopted an ARIMA model to forecast the traffic time-series data, and the satisfactory results could be obtained. Generally, the statistical theoretical models are simple, convenient and easy to apply. However, those models usually overlook the interferences of random factor, strong non-stationarity and nonlinearity hidden in the traffic data.
Unlike the statistical methods, the intelligent models usually perform better in explaining the nonlinear relationship between the input and output. These models include artificial neural network (ANN) [10,11,12,13,14], support vector machine (SVM) [15,16] and least square SVM (LSSVM) [17]. Wang et al. [16] proposed a brand-new model integrated the wavelet function and the SVM model to forecast the target data, which could improve the forecasting results. Luo et al. [18] presented a hybrid optimization algorithm combined particle swarm optimization (PSO) and genetic algorithm to find the optimal parameters of LSSVM, which could effectively improve the model’s accuracy and convergence speed. Shang et al. [19] introduced the proportion coefficient to combine the advantages of Gaussian kernel function and polynomial function. The forecasting result showed the built model was effective and practicable. Obviously, these intelligent models do not contain some special model architectures and have highly adaptable, especially for the nonlinear data. However, they may suffer from the problems of slow convergence speed and over-fitting.
To obtain more accurate and stable prediction, many scholars have introduced a variety of hybrid models which could combine the advantages of different models. The hybrid models can be commonly divided into four types: decomposition-based methods, weighting-based methods, parameter optimization-based methods, and error correction-based methods [20]. In recent years, the decomposition-based methods have become the research focus [21]. This kind of hybrid model could use the data processing models to address the nonlinear and non-stationary features in the data, and thus the forecasting accuracy could be enhanced.
The widely used decomposition algorithms have wavelet decomposition, empirical mode decomposition (EMD), and ensemble empirical mode decomposition (EEMD), etc. Among them, wavelet decomposition [22] is a multi-scale signal analysis method to tackle non-stationary signals. However, its performance usually relies on the selection of wavelet base functions. On the other hand, EMD can filter the signal adaptively [22]. By this method, different features in the original sequence can be filtered out step by step, and the corresponding subseries can be regarded as intrinsic mode functions (IMFs). Unfortunately, its decomposition process may suffer from the problems of model mixing and end effect. By adding many Gaussian white noise samples in EMD, EEMD has been developed [23]. However, it has the problem in the determination of noise amplitude and ensemble number. Nevertheless, these decomposition algorithms still have been successfully applied in the traffic flow prediction. For example, Duo et al. [24] proposed a hybrid forecasting method of short-term traffic volume based on EMD and the improved SVM. The forecasting results verified that EMD could improve accuracy significantly. Tang et al. [25] adopted a new hybrid model for traffic volume prediction by using the combination of EEMD and SVM. The results showed this model had superior performance over the single SVM. Tian et al. [26] presented a hybrid prediction model based on the improved complete EEMD (ICEEMDAN) algorithm, the kernel online sequential extreme learning machine (KOSELM), and the ARIMA model. The forecasting accuracy had been improved significantly. Despite these applications, the decomposition-based methods still go through various challenges.
To further enhance the accuracy of the traffic volume prediction, it is necessary to find new methods to deal with short-term traffic volume data. This paper proposes a novel time varying filtering based empirical mode decomposition (TVF-EMD) algorithm, which vividly describes the time-varying characteristics of data and overcomes the occurrence of mode mixing [27]. Specifically, TVF-EMD is firstly adopted to decompose the short-term traffic volume data and obtain multiple subsequences. Secondly, the LSSVM model is adopted for each subsequence to perform the final prediction. On this basis, five evaluation indexes including the mean absolute error, mean relative percentage error, root mean square error, root mean square relative error and equal coefficient are used to systematically evaluate the forecasting results. Meanwhile, the comparison of the proposed method with other forecasting models including EMD-LSSVM, LSSVM, and ARIMA is conducted. Finally, some conclusions are provided.
The rest of this paper is organized as follows: In Section 2, TVF-EMD and LSSVM are briefly discussed. Simultaneously, the structure and procedure of the proposed method are described in detail; In Section 3, two case studies are performed and the effectiveness of the proposed method is analyzed and discussed; In Section 4, some conclusions are summarized.
2. Methods
TVF-EMD is a data decomposition algorithm, which can be used to reduce the nonlinear and non-stationary components in short-time traffic volume data. On the other hand, LSSVM could perform well in describing short-time traffic volume data with nonlinear and non-stationary characteristics. This paper simultaneously combines the advantages of these two models and builds a new hybrid forecasting model, i.e., TVF-EMD-LSSVM. In order to better understand this method, the specific illustration of its notations is summarized in Appendix A.
2.1. Time Varying Filtering Based Empirical Mode Decomposition
EMD is an adaptive signal processing method that can decompose the signal into a series of IMFs and a non-zero mean residual [28], the expression is shown in Equation (1):
x(t)=∑i=1Nimfi(t)+r(t)
whereimfi(t)is theithimf,i=1,2,…,N . The EMD screening process can be divided into five steps, as shown in Appendix B.
As an IMF, the following conditions should be satisfied: (i) the number of zeros and poles must either be equal or differ at most by one; (ii) the local mean value of the upper and lower envelopes is zero. However, the above requirements have two limitations: (i) in the actual screening process, it is too rigid for stopping criterion; (ii) the second requirement of IMF may not be valid at a low sampling rate [27]. Thus, the model mixing occurs during decomposition. Aiming to overcome the weakness of EMD, Li et al. [27] proposed a TVF-EMD screening method to solve the above problems by developing local narrow-band signal. The local narrowband signal is not only similar to the IMF but also provides a Hilbert spectrum with physical significance. The filtering process of this method is completed by time-varying filtering, which is divided into three steps: (i) estimation of the local cut-off frequency; (ii) calculation of the local mean function; (iii) judgement of the residual signal.
2.1.1. Estimation of the Local Cut-Off Frequency
In TVF-EMD method, B-spline approximation filter is chosen as a time-varying filter, which adopts polynomial splines to approximate the signal and can be represented as:
gmn(t)=[pmn*x]↓m∗bmn(t)
where[.]↓mis the down-sampling operation;pmnis a pre-filter andpmn=[([bmn×bmn]↓m)-1]↑m×bmn,bmn(t)=βn(t/m);βn(t)denotes B-spline function;nstands for B-spline order;mrepresents the node;tis time;∗represents convolution operation.
According to Equation (2), the nodem determines the local cut-off frequency of the B-spline time-varying filter. In practice, the nodes cannot be known. As a result, it is necessary to estimate the local cut-off frequency from the input signal. Then, the B-spline time-varying filter is constructed. The specific process is provided in Appendix C.
2.1.2. Calculation of Local Mean Function
After obtaining the local cut-off frequencyφbis′(t), the signalsh(t)can be obtained by
h(t)=cos[∫φbis′(t)dt]
Taking the extreme time point({tmin},{tmin})ofh(t)as nodem, the time-varying filter can be constructed by B-spline approximation, and the cutoff frequency of the filter is consistent withφbis′(t). Subsequently, the B-spline approximation filter is performed on the input signal and the result is recorded asm(t).
2.1.3. Judgement of the Residual Signal
Since the definition of local narrow-band signal is closely related to the instantaneous bandwidth, TVF-EMD has formulated the relative criteria to check the instantaneous narrow-band signal, namely,
θ(t)=BLoughlin(t)φavg(t)
For a given bandwidthξthreshold, ifθ(t)≤ξ, the signal can be viewed as a narrow-band signal. Here, the weighted average instantaneous frequencyφavg(t)and Loughlin instantaneous bandwidthBLcan be calculated by:
φavg(t)= a12(t)φ1′(t)+ a22(t)φ2′(t) a12(t)+ a22(t)
BL(t)=a ′12(t)+a ′22(t) a12(t)+ a22(t)+ a12(t) a22(t)(φ1′(t)−φ2′(t))2( a12(t)+ a22(t))2
2.2. Least Square Support Vector Machine
After decomposing by TVF-EMD, the LSSVM model is built for each subseries. LSSVM has great improvement over the SVM model. The inequality constraints in the standard SVM algorithm are replaced by the equality constraints. On the conditions, the quadratic programming problem is transformed into the problem of solving linear equations [29].
Considering a set of dataD=(xi,yi),i=1,⋯,k, wherexi∈Rgis input andgis the dimension ofxi which can be determined by minimizing the root mean square error of the values output by the training part [20];yi∈Ris corresponding output. Assuming that the training part{(x1,y1),(x2,y2),⋯(xk−g,yk−g)}is composed ofk−gdata sets and the corresponding output isyi=x(i+g),i=1,2,⋯,k−g. Thus, the regression function can be written as follow:
f(x)=ωTψ(x)+d
whereψ(·)denotes a non-linear function;ωrepresents a weight vector;dis an offset. The parametersωanddcan be obtained by optimizing the following function:
{minω,d,qJ1(ω,q)=μEW+ςED=12μωTω+12ς∑i=1k−gqi2s.t.yi=ωTψ(x)+d+qi
whereqidenotes error variable;μandςdenote variable parameters;EW=12ωTω;ED=12∑i=1k−gqi2=12∑i=1k−g[yi−ωTψ(x)+d]2. To solve the above optimization problems, the Lagrange function is constructed as shown in Equation (9).
L(ω,d,q,α)=J(ω,q)−∑i=1k−gαi{ωTψ(x)+d+q−yi}
whereαiis the Lagrange multiplier. According to the Karush-Kuhn-Tucker (KKT) conditions, the optimal solution can be calculated by:
{∂L∂ω=0→ω=∑i=1k−gαiψ(xi)∂L∂d=0→∑i=1k−gαi=0∂L∂qi=0→αi=γqi∂L∂αi=0→ωTψ(xi)+d+qi−yi=0
whereγ=ζ/μdenotes the penalty coefficient. After eliminatingqiandω, the original optimization problem becomes
[0LTLΩij+1γI][dα]=[0Y]
whereΩij=ψ(xi)Tψ(xj)=K(xi, xj);L=[1,⋯,1]T;Y=[y1,⋯yN]T. Finding outαanddthrough Equation (11), the LSSVM regression model becomes:
f(x)=∑i=1k−gαiK(x,xi)+d
whereK(x,xi)is the kernel function which needs to meet Mercer’s conditions. Generally, the kernel functions include RBF kernel function, sigmoid kernel function and polynomial kernel function, etc. The RBF kernel function is also called the Gaussian kernel function. It has strong nonlinear learning ability with fewer parameters, which is the most effective kernel function. Therefore, the RBF kernel function is selected in this paper. It can be expressed as
K(x,xi)=exp[−‖x−xi‖2/(2σ2)],σ>0
whereσdenotes the kernel function parameter. When applying the LSSVM model with RBF kernel function, the selections of the parameterσand the penalty coefficientsγdetermine the model’s learning and generalization capabilities. Thus, it’s vital to search for the most suitable parameters.
2.3. The Proposed Method
Based on the above discussions, a novel hybrid model which combines the TVF-EMD model and LSSVM model can be developed to improve the forecasting accuracy. First, the TVF-EMD method is presented to deal with the non-stationary and nonlinear traffic volume series. After that, multiple subsequences called narrow-band signals are obtained. Then, the LSSVM model is established for each subsequence. Finally, the prediction results of the subsequences are accumulated to generate the lasted forecasting results. The specific process of TVF-EMD-LSSVM model is shown in Figure 1, and the steps are shown as follows:
Step 1: Preprocess the original traffic volume data with the errors data and missing data to get the experimental data;
Step 2: Decompose the data into several subsequences({cj(1),…,cj(k)},j=1…M+1)by TVF-EMD algorithm;
Step 3: Divide each subseries into two parts, including training parts{x′(1),…,x′(k)}and test parts{x′(k+1),…,x′(k+N)};
Step 4: Establish the LSSVM model to predict thek+1th datac^j(k+1)of subsequences, and sum up to get the forecasting valuex^(k+1);
Step 5: After updating the training set data to{x′(2),…,x′(k+1)}, repeat step 2 to step 4 to obtain the prediction results. Continue to predict one step ahead until the prediction task is completed.
3. Case Study 3.1. Data Description
The data collection A (including 2016 samples) was measured from the intersection entrance A of an arterial road in the main urban area of Chongqing and the location is shown in Figure 2. The statistical interval was 5 min, as shown in Figure 3. Two-thirds of the data were used to train the model, and the rest were used to test the performance of the built model. Table 1 summarizes the characteristics of data collection A. It could be observed that this dataset had strong volatility.
3.2. Data Processing
There are many factors affecting prediction accuracy, such as data quality, data characteristics, and model selection, etc. However, the quality of traffic volume data is one of the main factors [26]. Therefore, the processing of the abnormal data including missing data and erroneous data appears to be crucial in traffic volume prediction [30]. To repair abnormal data, the adjacent completion method is adopted and its function is shown in Equation (14):
x(t)=[x(t−w)+x(t−w−1)+⋯+x(t−1)]/w
wherewdenotes the number of data to be repaired.
3.3. Evaluation Criteria
In order to analyze and evaluate the forecasting performance of the proposed model, five commonly used evaluation indexes including mean absolute error (MAE), mean relative percentage error (MRPE), root mean square error (RMSE), root mean square relative error (RMSRE) and equal coefficient (EC) were used in the study [26,29]. Their specific definitions are given by:
MAE=1n∑i=1n|yi−y^i|
MRPE=1n∑i=1n|yi−y^iyi|
RMSE=1n∑i=1n(yi−y^i)2
RMSRE=1n∑i=1n(yi−y^iyi)2
EC=1-∑i=1n(yi−y^i)2∑i=1n(yi)2+∑i=1n(y^i)2
The smaller values of MAE, MRPE, RMSE, and RMSRE indicate the higher accuracy. The closer to one the EC value is, the higher accuracy the prediction is. 3.4. Prediction Results and Analysis
3.4.1. TVF-EMD-LSSVM Model Prediction
According to the forecasting process of the proposed model in Section 2.3, TVF-EMD is used to decompose the experimental data A into 10 subsequences, as shown in Figure 4.
By constructing training and test sets for each subsequence, the LSSVM model is built to predict them. The dimension parameter was determined by minimizing the root mean square error of the output value in the training part [20]. Moreover, the optimal penalty coefficient and kernel function parameters of each subsequence were determined by the optimization function. Finally, the traffic volume prediction value was obtained by accumulating the forecasting results of the subsequences.
3.4.2. Comparison and Analysis of Forecasting Results
To illustrate the performance of the proposed method, three additional forecasting models including ARIMA model, LSSVM model, and EMD-LSSVM model were used to perform the performance comparison. The processes of the LSSVM model, ARIMA model, and EMD-LSSVM model were similar to the forecasting progress in Section 3.4.1. The evaluation indexes of four different models are shown in Table 2 and the corresponding prediction results are shown in Figure 5. From these comparisons, some main observations are provided below:
-
Compared with the other three involved models, the proposed model had better forecasting performance, where its error indexes of MAE, MRPE, RMSE, RMSRE, and EC were 1.721, 3.969%, 2.974, 6.797%, and 0.9956, respectively. Specifically, in Figure 4, the red line represents the prediction result of the proposed model, while the blue line represents the true value. Their comparison indicates the proposed method could well capture the time-varying characteristics of the actual situation. From Table 2, the forecasting accuracy of the proposed model was higher than the EMD-LSSVM model with the reductions in terms of the five indexes MAE, MRPE, RMSE, RMSRE, and EC by 2.654, 5.991%, 2.831, 8.464%, and 0.0174, respectively. The reason could be that the TVF-EMD algorithm uses time-varying filtering technology, which could describe the time-varying characteristics of the data. Simultaneously, it can improve the imperfection of the model mixing in the EMD algorithm.
- Compared with the single models, the decomposition-based forecasting methods had the higher forecasting accuracy. For example, five error indexes in terms of MAE, MRPE, RMSE, RMSRE, and EC of the LSSVM model were 8.131, 17.871%, 10.801, 27.674%, and 0.9336, respectively, which presents the evident accuracy reduction in comparison with those of the proposed method. Compared with EMD-LSSVM model, these indexes were reduced by 3.756, 7.911%, 4.996, 12.413%, and 0.0306, respectively. The reason for these phenomena could be attributed to high non-stationarity and nonlinear characteristics embedded in the original data, which could be effectively addressed by the decomposition methods.
- The MAE, MRPE, RMSE, RMSRE, and EC of the ARIMA model were 8.284, 17.977%, 11.01, 27.25% and 0.9322. Compared with LSSVM, these indexes were reduced by 0.153, 0.106%, 0.209 0.424, and 0.0014, respectively. The reason could be attributed to that the nonlinear features hidden in the original data were more significant than those of linear one, which leads to the conclusion that the linear ARIMA model cannot capture the characteristics well. Therefore, it owns the lowest forecasting accuracy.
To further test the stability of the proposed model, another group of data (data collection B) was used. These data were measured from the intersection entrance B of an arterial road in Chongqing (including 2016 samples), as shown in Figure 2 and Figure 6. Table 3 provided the relevant information of them. For simplicity, only the error indexes are given in Table 4. The intuitively results are shown in Figure 7. From Table 4 and Figure 7, the main results we
- TVF-EMD was better than EMD in dealing with data nonlinearity and non-stationary. The forecasting result proves that the forecasting accuracy of TVF-EMD based method was higher than EMD based method.
- The hybrid models could take advantage of the superiority each component model. The results display that the forecasting accuracy of the hybrid models was higher than that of the single models.
- The ARIMA model usually presented the high performance for the data with significant linear features. However, for short-term traffic volume data with high nonlinear characteristics, the LSSVM model may have better forecasting performance.
Data Resource | Mean | Variance | Maximum | Minimum | Skewness | Kurtosis | Non-Stationarity |
---|---|---|---|---|---|---|---|
A | 70.184 | 20,710 | 188 | 3 | 0.0183 | 1.6495 | Strong |
Model | MAE | MRPE | RMSE | RMSRE | EC |
---|---|---|---|---|---|
LSSVM | 8.131 | 17.871% | 10.801 | 27.674% | 0.9336 |
EMD-LSSVM | 4.375 | 9.96% | 5.805 | 15.261% | 0.9642 |
TVF-EMD-LSSVM | 1.721 | 3.969% | 2.974 | 6.797% | 0.9816 |
ARIMA | 8.284 | 17.977% | 11.01 | 27.25% | 0.9322 |
Data Resource | Mean | Variance | Maximum | Minimum | Skewness | Kurtosis | Non-Stationarity |
---|---|---|---|---|---|---|---|
B | 68.8284 | 2200.5 | 235 | 2 | 0.2587 | 2.0441 | Strong |
Model | MAE | MRPE | RMSE | RMSRE | EC |
---|---|---|---|---|---|
LSSVM | 9.281 | 20.415% | 14.787 | 33.153% | 0.8201 |
EMD-LSSVM | 5.93 | 13.228% | 8.405 | 20.675% | 0.8983 |
TVF-EMD-LSSVM | 0.898 | 2.653% | 1.20 | 5.089% | 0.9855 |
ARIMA | 9.584 | 20.364% | 15.762 | 32.269% | 0.808 |
Author Contributions
Conceptualization, Y.W. and L.Z.; investigation, Y.X.; methodology, Y.W. and X.W.; software, L.Z.; validation, Y.W., L.Z. and S.L.; writing-original draft, Y.W.; writing-review and editing, Y.W. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by Science and Technology Research Project Fund of Chongqing Education Commission (Grant NO. KJ1705136, KJ1600512).
Conflicts of Interest
The authors declare no conflict of interest.
Appendix A. Notation Illustration
Parameters and variables | υ | Dimension | |
ξ | Bandwidth | ||
A(t) | Instantaneous amplitude | x(t) | Input signal |
φ'(t) | Instantaneous frequency | ai(t) | Amplitude of the i-th component |
φbis'(t) | Local cut-off frequency | [.]↓m | Down-sampling operation |
βn(t) | B-spline function | n | B-spline order |
t | Time | ∗ | Convolution operation |
φavg(t) | Average instantaneous frequency | BL | Instantaneous bandwidth |
ω | Weight vector | d | Offset |
μ | Variable Parameter | ς | Variable Parameter |
γ | Penalty coefficient | σ | Kernel function parameters |
φ(t) | Instantaneous phase | ψ(·) | Non-linear function |
φi(t) | Phase of the i-th Component | qi | Error variable |
pmn | Pre-filter | αi | Lagrange multiplier |
m | Node | xi | Input |
Appendix B. The Screening Process of EMD
Step 1: Find the local maximum and minimum of thex(t);
Step 2: Calculate the average valuem(t)of the upper line and lower envelope linem(t)=(u(t)+l(t))/2. The upper envelopeu(t)and the lower envelopel(t)are obtained by using the cubic spline function;
Step 3: Extracth(t)=x(t)-m(t)and judge whether theh(t)is satisfied the conditions of IMF. If not, viewh(t)as the original sequence and repeat the above steps;
Step 4: Afterntimes of screening, an IMFhn(t)which satisfies the conditions of IMF recorded asc1(t)=hn(t)can be received and calculate the residual componentr1(t)=x(t)-c1(t);
Step 5: Repeat the above steps forr1(t)to get all the IMFs.
Appendix C. The Construction of the B-Spline Time-Varying Filter
Step 1: The Hilbert transform is used to calculate the instantaneous amplitudeA(t)and instantaneous frequencyφ'(t)of the input signalx(t).
A(t)=x2+x^(t)2
φ'(t)=d(arctan[x^(t)/x(t)])/dt
wherex^(t)denotes the Hilbert transform of the signal.
Step 2: Determine the maximum value{tmax}and minimum{tmin}value ofA(t). For multicomponent signals, the analytical signal can be expressed as the sum of two signals.
z(t)=A(t)ejφ(t)=a1 ejφ1(t)+a2 ejφ2(t)
whereφ(t)stands for instantaneous phase,φ(t)=arctan[x^(t)/x(t)].
Therefore, the following equations can be obtained.
A2(t)=a12(t)+a22(t)+2a1(t)a2(t)·cos[φ1(t)-φ2(t)]
φ'(t)=( φ'1(t)(a12(t)+a1(t)a2(t)cos[φ1(t)-φ2(t)])+ φ'2(t)(a22(t)+a1(t)a2(t)cos[φ1(t)-φ2(t)]))1A2(t)+1A2(t)( a'1(t)+a2(t)sin[φ1(t)-φ2(t)]- a'2(t)+a1(t)sin[φ1(t)-φ2(t)])
In Equations (A4) and (A5),ai(t)andφi(t)are the amplitude and phase of theith component respectively. Assuming that the local minimum valueA(t)is obtained attmin, it satisfies the Equation (A6).
cos[φ1(tmin)-φ2(tmin)]=-1
Then, Equations (A7) and (A8) can be obtained by substituting Equation (A6).
A(tmin)=|a1(tmin)-a2(tmin)|
φ'(tmin)A2(tmin)=φ1'(tmin) a12(tmin)-φ1'(tmin)a1(tmin)a2(tmin)+φ2'(tmin)[ a22(tmin)-a1(tmin)a2(tmin)]
Simultaneously,A(tmin)denotes a local minimum ofA(t), letA'(tmin)=0, the Equation (A9) can be acquired.
a1'(tmin)-a2'(tmin)=0
Thus, the minimum value ofA(t)can be obtained by solving Equations (A4)-(A9). Similarly, the maximum value ofA(t)can be determined, too.
Step 3: Calculatea1(t)and let,
β1(t)=|a1(t)-a2(t)|β2(t)=a1(t)+a2(t)
Thus, the Equation (A11) can be obtained from Equation (A5).
β1(tmin)=A(tmin)=|a1(tmin)-a2(tmin)|β2(tmax)=A(tmax)=a1(tmax)+a2(tmax)
Becausea1(t)anda2(t)change slowly,β1(t)andβ2(t)can be acquired by interpolation in point setA({tmin})andA({tmax})respectively.a1(t)anda2(t)can be gained by Equation (A11).
a1(t)=[β1(t)+β2(t)]/2a2(t)=[β2(t)-β1(t)]/2
Step 4: Calculateφ1'(t)andφ2'(t), let,
η1(t)=φ1'(t)[ a12(t)-a1(t)a2(t)]+φ2'(t)[ a22(t)-a1(t)a2(t)]η2(t)=φ1'(t)[ a12(t)+a1(t)a2(t)]+φ2'(t)[ a22(t)+a1(t)a2(t)]
From Equation (A5), we have
η1(tmin)=φ'(tmin)A2(tmin)=φ1'(tmin)[ a12(tmin)-a1(tmin)a2(tmin)]+φ2'(tmin)[ a22(tmin)-a1(tmin)a2(tmin)]η2(tmax)=φ'(tmax)A2(tmax)=φ1'(tmax)[ a12(tmax)+a1(tmax)a2(tmax)]+φ2'(tmax)[ a22(tmax)+a1(tmax)a2(tmax)]
Sincea1(t),a2(t),φ1'(t)andφ2'(t)changes slowly,η1(t)andη2(t)can be received by interpolation in point setφ'({tmin})A2({tmin})andφ'({tmax})A2({tmax}). Thus,φ1'(t)andφ2'(t)can be calculated by solving Equation (A13).
φ1'(t)=η1(t)2 a12(t)-2a1(t)a2(t)+η2(t)2 a12(t)+2a1(t)a2(t)φ2'(t)=η1(t)2 a22(t)-2a1(t)a2(t)+η2(t)2 a22(t)+2a1(t)a2(t)
Step 5: Calculate the local cut-off frequencyφbis'(t)as follows:
φbis'(t)=φ1'(t)+φ2'(t)2=η2(t)-η1(t)4a1(t)a2(t)
Step 6: Rearrangeφbis'(t)to solve the problem of signal intermittence.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2020. This work is licensed under http://creativecommons.org/licenses/by/3.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Short-term traffic flow prediction is important to realize real-time traffic instruction. However, due to the existing strong nonlinearity and non-stationarity in short-term traffic volume data, it is hard to obtain a satisfactory result through the traditional method. To this end, this paper develops an innovative hybrid method based on the time varying filtering based empirical mode decomposition (TVF-EMD) and least square support vector machine (LSSVM). Specifically, TVF-EMD is firstly used to deal with the implied non-stationarity in the original data by decomposing them into several different subseries. Then, the LSSVM models are established for each subseries to capture the linear and nonlinear characteristics embedded in the original data, and the corresponding prediction results are superimposed to obtain the final one. Finally, case studies based on two groups of data measured from an arterial road intersection are employed to evaluate the performance of the proposed method. The experimental results indicate it outperforms the other involved models. For example, compared with the LSSVM model, the average improvements by the proposed method in terms of the indexes of mean absolute error, mean relative percentage error, root mean square error and root mean square relative error are 7.397, 15.832%, 10.707 and 24.471%, respectively.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer