Baoyu Xu 1 and Hongjun Zhang 1 and Zhiteng Wang 1 and Huaixiao Wang 2 and Youliang Zhang 3
Academic Editor:Peter Dabnichki
1, PLA University of Science and Technology, Nanjing 210007, China
2, Bengbu Automobile NCO Academy, Bengbu 233011, China
3, Logistics Academy, Beijing 100858, China
Received 30 August 2015; Accepted 15 October 2015; 2 November 2015
This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
Artificial neural network (ANN) is a new information processing and computer system which is based on the modern neuroscience research and is formed by abstracting, simplifying, and simulating of biological structure. The features of ANN are as follows. ANN can fully approximate to any complex nonlinear relations; all quantitative or qualitative information keeps in storage equipotentially in each neuron of network so it is with very strong robustness and fault tolerance; ANN is a kind of system which can emulate and adapt to the unknown system and is able to deal with the quantitative and qualitative knowledge at the same time [1]. The main research of ANN is how to make computer simulate and realize the self-learning and mathematical thinking ability of human to mine the inner relation from limited samples. It is mainly through studying and storing the data relations, the inference rules, and probability distribution of known sample to deduce and reveal the potential information between variables in the unknown data sample [2].
BP neural network is currently the most popular neural network model in application [3]. BP algorithm was proposed by Rumelhart and Mcllelland in 1986 which well solved the weight adjustment problems of nonlinear continuous function in the field of multilayer feedforward neural network. It is a typical error back propagation algorithm [4, 5]. Selection of activation function, design of structure parameter, and improvement of network defect have been researched a lot since the emergence of the BP neural network. In 1973, Grossberg found sigmoid function is very similar to the work situation of biological neurons so he began to explore the relationship between the features of sigmoid function and the stability of the neural network. He prompted the function to become the most commonly used activation function of BP neural network. Since then many scholars have done improvements on the limitations of the sigmoid function. In order to solve the problem that BP algorithm can obtain larger gradient values in the whole domain, many scholars try to combine activation functions with different characteristics in different intervals together to set a larger derivative value at the needful position to make up for the inadequacy of single activation function [6]. In 1991, Kung and Hu [7] proposed FARM approximate simplified method which uses Frobenius norm according to the ideas of the global optimization and gradual optimization to delete the hidden layer units and determine the number of hidden layer nodes. Fahlman [8] found that if we adopt the sigmoid function as the activation function, the derivative of error of weight will become very small when the output value of BP network is close to 0 and 1. Salomon [9] created two BP neural networks with exactly the same initial parameters and structure and put forward a new method of network learning. The learning rate [figure omitted; refer to PDF] is increased and decreased, respectively, to update the weight of two networks and network is set with fast error drop as the starting point of the next update. The effect is obvious. Dan Foresee and Hagan [10] proposed BFGS quasi-Newton method which not only can avoid calculation of the second derivative but also keeps the advantage of fast convergence of the Newton algorithm. Riedmiller and Braun [11] proposed elastic BP (RPROP) method in 1993. RPROP method introduced probability of resilient update value to modify weights directly which can reduce influence of the network structure parameters in the whole learning process and avoid unforeseen gradient error convergence of fuzzy network performance.
At present, the research based on ANN to carry out quantitative structure-activity relationship research has gradually been applied to various fields. Wang et al. [12] utilized neural network to study the quantitative structure-activity relationship of angiotensin converting enzyme inhibitors. Zhao et al. [13] established the model of antitumor activity of emodin derivatives on the basis of neural network. Cui et al. [14] applied the principal component analysis and neural network method to study the quantitative structure-activity relationship of the nitrobenzene and its homologue. Gonzalez-Diaz et al. [15] established some quantitative structure-activity relationship models of synthetic compounds by neural network and verified the reliability of the model. The model can be used to design new drugs. Prado-Prado et al. [16] studied quantitative structure-activity relationship of parasite drug resistance between different kinds of parasites through neural network. Ramirez-Galicia et al. [17] established quantitative structure-activity relationship model of amoeba drug resistance through multiple linear regression, stepwise regression analysis, and neural network. The result shows that the three-dimensional structure of the model is very important.
Although BP algorithm has become the most widely used artificial neural network, the BP neural network has the following defects.
[figure omitted; refer to PDF] It falls into local minimum value easily: the BP neural network learning algorithm based on gradient descent method is easy to fall into local extremum points and saddle point if error function is not strictly convex function and there are multiple points whose gradient is zero. Then, the network can not converge and is unable to get the optimal solution of the problem so the optimal network connection weights and threshold parameters can not be obtained [18].
[figure omitted; refer to PDF] The speed of error convergence is slow: convergence speed of BP neural network is decided by two aspects: one is learning rate and the other one is the size of derivative of related excitation function [19]. It is usually not easy to choose the size of the learning rate. If the learning rate is too large, the oscillation or even no convergence phenomenon will happen in the process of training. The learning rate should be a small positive number. If the learning rate is too small, the product of learning rate and negative gradient vector will become smaller. Then, the renewal speed of weights and thresholds will be affected. In addition, the size of derivative of the activation function also affects the rate of convergence. In the flat area of error curved surface, the gradient of error function is small to make update speed of weights and threshold of network slow. The network needs to pass through much iteration to be out of the flat area so the convergence speed of the network becomes slow.
[figure omitted; refer to PDF] Structure of network is not easy to determine: the determination of the structure of BP neural network usually refers to determination of the number of hidden layers and the number of neurons of hidden layer. Especially after Kolmogorov et al. prove the single hidden layer of BP neural network approximation theorem, how to select the number of neurons in single hidden layer neural network has always been number one of the hot and key problems. Generally, the number of neurons of input layer and output layer is easy to determine according to identification objects. The number of neurons of hidden layer is difficult to determine, and it directly affects topology structure and performance of BP neural network. It has been proved in theory that the three-layer BP neural network can approximate nonlinear functions of arbitrary precision which solves the problem of determination of the number of hidden layers in the BP neural network. If the number of the neurons of hidden layer is too small, the network can not meet the requirements of learning and the approximation performance; if the number of the neurons of hidden layer is too much, there will be adverse phenomenon in the network and it will make the hardware implementation and software calculation complicate at the same time. At present, there is no unified and complete determination theoretical framework to determine the structure of the BP neural network. The experience or grope through lots of experiments is the usual manner to estimate and adjust the structure of neural network.
In order to further improve the efficiency of the BP neural network and overcome the shortage of it, a lot of research has been conducted. Sun et al. established improved BP neural network prediction model and quantitatively researched related parameters [20]. The accuracy of the model was improved to some extent. Xiao et al. proposed BP neural network with rough set for short term load forecasting. The accuracy of prediction of BP neural network was further improved [21].
Genetic algorithm imitates evolutionary and genetic rule of biology and is a mathematical algorithm which can solve problems to find the global optimization. Due to the strong macroscopic search ability and good global optimization performance of genetic algorithm, many scholars try to use genetic algorithm to optimize the connection weights, structure, learning rules, and so forth of BP neural network. Xiao et al. [22] applied genetic algorithm to constitute GA-ANN method to optimize the optimization problems in complex engineering. The method not only takes advantage of the nonlinear mapping, network reasoning, and predicting function of neural network but also uses the characteristics of global optimization of genetic algorithm. It can be widely applied to many complex engineering problems whose objective function is difficult to express by the form of explicit function of decision variables. Yang et al. [23] applied genetic algorithm to select parameters which overcame the restriction of symmetric weight matrix of traditional fluid neural network and broadened the application fields of this intelligent exploration method. Ge [24] applied genetic algorithm to optimize the controller parameters of structure of the neural network and applied the controller to control object with pure lag. The experiment proved that the control system optimized by genetic algorithm was with good static performance and dynamic performance. Li et al. [25] combined genetic algorithm with traditional DBD algorithm to propose a new algorithm to optimize the BP neural network, making network with large scale be able to converge fast and be out of the trap of local minimum. This algorithm is less sensitive to the selection of network parameters and obtains good effect in the application of missile comprehensive test.
In order to further overcome the shortage of the model based on genetic algorithm to optimize the BP neural network, the model and algorithm of BP neural network based on expanded multichain quantum optimization are proposed. The structure of neural network is effectively optimized. The model can overcome a series of problems of basic genetic algorithm, such as the slow convergence speed, premature convergence, and bad computational stability to further improve the performance of the BP neural network controller.
2. BP Neural Network Model and Algorithm
The main idea of BP algorithm is to divide learning process into two stages, that is, positive communication of signal and back propagation of error. In the stage of positive communication, input information is from the input layer to output layer through the hidden layer [26]. The output signal forms on the output side. The weights of network are fixed in the process of the signal transmission forward. The state of neurons of each layer only affects state of neurons of the next layer. If the desired output can not be achieved in the output layer, then the error signal will back propagate. In the back propagation stage, the error signal which failed to meet the accuracy requirement spreads forward step by step and the error is shared by all units of each layer. The connection weights are adjusted dynamically according to the error signal. The weight value between neurons keeps correcting through the cycle of forward and back adjustment. The learning stops when the error of the output signal meets the requirement of precision [27].
2.1. Topology Structure of BP Neutral Network
The simplest BP neural network is with three layers as is shown in Figure 1. It includes input layer, hidden layer, and output layer. The number of nodes in the input layer is equal to the dimension of input vector [figure omitted; refer to PDF] . The number of nodes in the output layer is equal to the type of output module [figure omitted; refer to PDF] . The number of nodes in the hidden layer [figure omitted; refer to PDF] is associated with specific application which is usually selected by test.
Figure 1: Topology structure of BP neutral network.
[figure omitted; refer to PDF]
The connection weight matrix between input layer nodes and hidden layer nodes is [figure omitted; refer to PDF] which is defined as formula (1). Similar to it, the connection weight matrix between hidden layer nodes and the output nodes can be defined as [figure omitted; refer to PDF] . The threshold matrix of nodes of hidden layer and output layer is [figure omitted; refer to PDF] and [figure omitted; refer to PDF] , respectively: [figure omitted; refer to PDF]
2.2. Learning Algorithm of BP Network
The learning algorithm of BP network can be divided into two stages that are forward and back propagation. It is a gradient descent algorithm which can make error of per connection weights of the neural network reduce. At the beginning of learning, assign the random numbers in the range of [figure omitted; refer to PDF] to the connection weight matrices [figure omitted; refer to PDF] , [figure omitted; refer to PDF] and threshold matrices of the hidden layer and output layer nodes [figure omitted; refer to PDF] , [figure omitted; refer to PDF] .
2.2.1. Forward Operation
First, enter the learning samples [figure omitted; refer to PDF] , where [figure omitted; refer to PDF] is the input vector of the learning samples and [figure omitted; refer to PDF] is the corresponding output vector: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] is the number of nodes of input layer; [figure omitted; refer to PDF] is the number of nodes of output layer.
The data forward propagates to the output layer through the input layer and hidden layer. The produced weight value of output pattern classification is the learning result. The following steps are mainly included.
Step 1 (calculation of output value of the hidden layer nodes).
The input value of nodes in hidden layer is [figure omitted; refer to PDF] where [figure omitted; refer to PDF] is the number of nodes of input layer, [figure omitted; refer to PDF] is the number of nodes of hidden layer, [figure omitted; refer to PDF] is connection weight, and [figure omitted; refer to PDF] is component of input vector.
The output value of node [figure omitted; refer to PDF] is [figure omitted; refer to PDF] where [figure omitted; refer to PDF] is the threshold of node [figure omitted; refer to PDF] . The activation function [figure omitted; refer to PDF] used is the sigmoid function given by Rumelhart: [figure omitted; refer to PDF]
Step 2 (calculation of the output value of the nodes of output layer).
The input values of nodes of output layer [figure omitted; refer to PDF] are [figure omitted; refer to PDF]
The output value of node [figure omitted; refer to PDF] is [figure omitted; refer to PDF] where [figure omitted; refer to PDF] is threshold of node of output layer [figure omitted; refer to PDF] and [figure omitted; refer to PDF] is activation function defined by type (5).
2.2.2. The Back Operation
Calculate the error of output value and expected value of output layer. The error back propagates from the output layer through the hidden layer to the input layer. The connection weight values are modified. Steps are as follows.
Step 1 (calculate the output error of nodes of output layer).
The error between the learning value [figure omitted; refer to PDF] of node of output layer [figure omitted; refer to PDF] and the output value of learning samples [figure omitted; refer to PDF] is [figure omitted; refer to PDF]
Step 2 (test learning error).
[figure omitted; refer to PDF] is the allowed maximum learning error, which is in the interval [figure omitted; refer to PDF] set by the user.
If [figure omitted; refer to PDF] , enter the next learning sample. Otherwise, adjust the weights of network and reenter the original learning sample.
The learning process is ended when all learning samples meet the aforementioned conditions.
Step 3 (calculate the learning error of nodes of output layer).
The learning error of node of output layer [figure omitted; refer to PDF] is [figure omitted; refer to PDF]
Step 4 (calculate learning error of nodes of hidden layer).
Learning error of node of hidden layer [figure omitted; refer to PDF] is [figure omitted; refer to PDF]
Step 5 (revise the value of the connection weights matrix [figure omitted; refer to PDF] ).
Set the weight value at time [figure omitted; refer to PDF] as adjusted new weight value; then [figure omitted; refer to PDF] where [figure omitted; refer to PDF] is learning rate and [figure omitted; refer to PDF] is momentum factor. Both [figure omitted; refer to PDF] and [figure omitted; refer to PDF] are in the scope [figure omitted; refer to PDF] . Using [figure omitted; refer to PDF] can accelerate the learning speed and be helpful to overcome the local minimum problem of common BP algorithm.
Step 6 (revise the value of the connection weight matrix [figure omitted; refer to PDF] ).
Consider [figure omitted; refer to PDF]
Step 7 (revise threshold [figure omitted; refer to PDF] ).
Threshold of nodes of output layer is [figure omitted; refer to PDF]
Step 8 (revise threshold [figure omitted; refer to PDF] ).
Threshold of nodes of hidden layer [figure omitted; refer to PDF] is [figure omitted; refer to PDF]
2.3. Run the BP Neural Network
Run the BP neural network after learning and carry out pattern classification for the input vector [figure omitted; refer to PDF] . Only the forward operation of BP learning algorithm is used. The process is as follows.
Step 1.
Assign the value revised by BP learning algorithm to the connection weight matrices [figure omitted; refer to PDF] and [figure omitted; refer to PDF] and threshold matrices [figure omitted; refer to PDF] and [figure omitted; refer to PDF] .
Step 2.
Input the input vector [figure omitted; refer to PDF] which needs to be recognized.
Step 3.
Apply formulas (3), (4), and (5) to calculate the output of hidden layer.
Step 4.
Use formulas (5), (6), and (7) to calculate the output of the output layer, that is, the classification result of input vector [figure omitted; refer to PDF] .
3. The Quantum Optimal Model Based on Multichain Expanded Coding Scheme
3.1. Principle of Quantum Coding Based on Multichain Expanded Scheme
In quantum computation, the smallest unit of information is represented by quantum bit whose state can be expressed as [figure omitted; refer to PDF] where [figure omitted; refer to PDF] and [figure omitted; refer to PDF] satisfy the normalization condition [figure omitted; refer to PDF]
The complex numbers [figure omitted; refer to PDF] and [figure omitted; refer to PDF] satisfying formulas (15) and (16) are called probability amplitude of a quantum bit. Quantum bit can also be expressed by probability amplitude as [figure omitted; refer to PDF] . According to the property of probability amplitude, a quantum bit [figure omitted; refer to PDF] can be expressed by Figure 2.
Figure 2: Probability amplitude of quantum bit.
[figure omitted; refer to PDF]
Obviously, in Figure 2, [figure omitted; refer to PDF] and [figure omitted; refer to PDF] ; therefore, quantum bit can be represented as [figure omitted; refer to PDF]
The coding scheme is [figure omitted; refer to PDF] where [figure omitted; refer to PDF] , [figure omitted; refer to PDF] ; [figure omitted; refer to PDF] ; [figure omitted; refer to PDF] ; [figure omitted; refer to PDF] is the size of the population; [figure omitted; refer to PDF] is the number of quantum bit.
Obviously, in Figure 3, [figure omitted; refer to PDF] can be decomposed into [figure omitted; refer to PDF] with the same way of formula (17). Combine formulas (17) and (18); quantum bit can be represented as the following form: [figure omitted; refer to PDF]
Figure 3: Decomposition of probability amplitude of quantum bit.
[figure omitted; refer to PDF]
Formula (19) is to decompose [figure omitted; refer to PDF] of formula (17) into two variables, so it also meets the condition of formula (16); however, it describes the feature from the view of two-dimensional space. Only one variable [figure omitted; refer to PDF] is not conducive to describing the dynamic behavior of the quantum objectively, comprehensively, and vividly. [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) is used to replace the described component of hypotenuse according to the characteristics of formulas (17) and (18). Formula (19) can be transformed to [figure omitted; refer to PDF]
Formula (9) is equivalent to translating [figure omitted; refer to PDF] in formula (1) to the representation of quantum bit that is the variable in two-dimensional space which is translated to three-dimensional space. The encoding scheme based on three genes of three-dimensional space of the two angle variables is formed: [figure omitted; refer to PDF]
Similarly, [figure omitted; refer to PDF] can be added to extend [figure omitted; refer to PDF] to form the encoding scheme based on four chains of four-dimensional space: [figure omitted; refer to PDF]
Increase of the supporting angle constantly like this can conclude the encoding scheme based on [figure omitted; refer to PDF] chain of [figure omitted; refer to PDF] -dimensional space, where [figure omitted; refer to PDF] is the number of angle variables of multiple chain code.
It can be concluded from Figure 4 that the [figure omitted; refer to PDF] chain coding scheme is [figure omitted; refer to PDF]
Figure 4: [figure omitted; refer to PDF] chain gene code.
[figure omitted; refer to PDF]
3.2. Transform of Solution Space
The process of the quantum genetic optimization based on multichain code is limited within the unit space [figure omitted; refer to PDF] . Therefore, transformation between unit space and solution space of the optimization problem is required. Suppose that the [figure omitted; refer to PDF] th solution variable of the optimization problem is [figure omitted; refer to PDF] and the value of the [figure omitted; refer to PDF] th quantum bit of the [figure omitted; refer to PDF] th chromosome [figure omitted; refer to PDF] is [figure omitted; refer to PDF] , where [figure omitted; refer to PDF] is the number of the angle variable of multichain code. The corresponding transformation formula of solution space is [figure omitted; refer to PDF]
3.3. Update of Quantum Chromosome
The update of the quantum rotation gate [figure omitted; refer to PDF] makes each chromosome in the current population close to the optimal chromosome. During this process, better chromosome may be generated to make the population evolve. However, different coding scheme needs different quantum rotation gate to make the quantum bit rotate a corresponding angle [figure omitted; refer to PDF]
Set [figure omitted; refer to PDF] to be the [figure omitted; refer to PDF] th quantum bit of contemporary optimal chromosome, and [figure omitted; refer to PDF] is the [figure omitted; refer to PDF] th quantum bit of the [figure omitted; refer to PDF] th chromosome of contemporary population [figure omitted; refer to PDF]
[figure omitted; refer to PDF] The rule of determining of rotation angle [figure omitted; refer to PDF] is as follows. If [figure omitted; refer to PDF] , the direction is [figure omitted; refer to PDF] ; if [figure omitted; refer to PDF] , the direction can be either positive or negative.
[figure omitted; refer to PDF] The rule of determining of rotation angle [figure omitted; refer to PDF] is as follows. If [figure omitted; refer to PDF] , the direction is [figure omitted; refer to PDF] ; if [figure omitted; refer to PDF] , the direction can be either positive or negative.
Set the maximal evolutionary generation to be [figure omitted; refer to PDF] , and the size of the [figure omitted; refer to PDF] th quantum rotation angle of the [figure omitted; refer to PDF] th generation is [figure omitted; refer to PDF] , where [figure omitted; refer to PDF] is the maximum rotation angle. The changing type of the rotation angle can ensure the angle decreases with the increase of generation to rotate around the optimal solution.
3.4. Mutation of Quantum Chromosome
The mutation of quantum chromosome is realized by quantum Not gate [figure omitted; refer to PDF] . The quantum bit with the amplitude of [figure omitted; refer to PDF] is changed to [figure omitted; refer to PDF] through the role of quantum Not gate. Consider [figure omitted; refer to PDF]
The specific form of [figure omitted; refer to PDF] can be determined by the method of undetermined coefficients. Diversity of population can be increased though the quantum Not gate which can prevent local convergence and avoid premature phenomenon in the process of evolution.
3.5. Algorithm Description
The main realization steps of multichain quantum genetic algorithm are as follows.
Step 1 (population initialization).
Generate population with [figure omitted; refer to PDF] initial individuals [figure omitted; refer to PDF] according to formula (12) randomly.
Step 2 (transformation of solution space).
Map the multiple approximate solutions of each chromosome from unit space [figure omitted; refer to PDF] to the solution space of optimization problem [figure omitted; refer to PDF] to get the solution set [figure omitted; refer to PDF] .
Step 3.
Calculate the fitness of the approximate solution to obtain the contemporary optimal solution Best [figure omitted; refer to PDF] and contemporary optimal chromosome Best [figure omitted; refer to PDF] .
Step 4.
Make Best [figure omitted; refer to PDF] be the optimal solution [figure omitted; refer to PDF] ; make Best [figure omitted; refer to PDF] be the global optimal chromosome [figure omitted; refer to PDF] .
Step 5.
Do the iteration cycle, [figure omitted; refer to PDF] , obtaining new species by updating and mutation.
Step 6.
Transform solution space of [figure omitted; refer to PDF] to get the solution of optimal problem [figure omitted; refer to PDF] .
Step 7.
Evaluate [figure omitted; refer to PDF] to access the contemporary optimal solution Best [figure omitted; refer to PDF] and optimal chromosome Best [figure omitted; refer to PDF] .
Step 8.
If [figure omitted; refer to PDF] , update the contemporary optimal solution [figure omitted; refer to PDF] and update the contemporary optimal chromosome [figure omitted; refer to PDF] at the same time to prevent the degradation phenomenon. Otherwise, [figure omitted; refer to PDF] , [figure omitted; refer to PDF] . Such control can ensure the convergence to the optimal value.
Step 9.
Go back to Step [figure omitted; refer to PDF] .
4. The Model of BP Neural Network Based on Multichain Code Quantum Optimization Algorithm
4.1. Code
Connect each weight and threshold of BP network in order to form a long string of an array of real number. The string is as a chromosome. The decoding value of individual is the corresponding weights and thresholds.
4.2. Fitness Function
The purpose of using multichain quantum algorithm to optimize BP network is to simplify the structure of the network to minimize the error of network. Therefore, the fitness function is [figure omitted; refer to PDF] where [figure omitted; refer to PDF] is the global error of network; [figure omitted; refer to PDF] is the number of training samples; [figure omitted; refer to PDF] is the number of units of output layer; [figure omitted; refer to PDF] is the desired output value; [figure omitted; refer to PDF] is the actual output value.
The combination of multichain quantum optimization algorithm and neural network is realized by using multichain quantum optimization algorithm to optimize the parameters of the neural network. The specific ways are as follows.
(1) Use the algorithm to optimize the initial weights and thresholds of neural network.
(2) Use the algorithm to optimize structure parameters of BP network, the main of which is to optimize the unit number of hidden layers of network.
(3) Use the algorithm to optimize the selection the learning rate and momentum factor of BP network.
The first optimization way is adopted to realize the optimization of neural network in this paper. The specific implementation process is as below.
The basic idea of using multichain quantum optimization algorithm to optimize the initial weight and threshold value of BP network is as follows. The multichain quantum optimization algorithm is applied to ensure the appropriate initial weight and threshold value to find the optimal solution of BP algorithm.
The specific implementation steps of MCQA-BP (BP neural network based on multichain quantum optimization) algorithm are as follows.
(1) Initial population with [figure omitted; refer to PDF] individuals is produced by multichain quantum optimization algorithm. Each individual in the population composed the chromosome strings with the weight, threshold, and the number of units of hidden layer of BP network.
(2) Enter the neural network module. Apply BP algorithm to do the learning process and apply multichain quantum optimization algorithm to do the fitness inspection. If the fitness is unqualified, update the chromosome until the fitness meets the requirements. The individuals which meet requirements are updated by multichain quantum optimization algorithm to calculate fitness. The unqualified individuals are updated until they meet the accuracy requirement. Then end the algorithm.
5. Simulation Test
5.1. Test Case 1
Apply MCQA-BP algorithm to diagnose fault state of machine. The data sample after normalization of machine status is shown in Table 1. The algorithm adopts network with three layers. The function of hidden layer is sigmoid tangent function; the function of output layer is sigmoid logarithmic function. The formula [figure omitted; refer to PDF] is used to calculate the number of neurons of hidden layer, where [figure omitted; refer to PDF] is the number of neurons of input layer; [figure omitted; refer to PDF] is the number of neurons of hidden layer. Since there are 15 input parameters and 3 output parameters, the structure of BP neural network is 15-31-3. That is, the number of nodes of input layer is 15, the number of nodes of hidden layer is 31, and the number of nodes of output layer is 3. The number of weight values is [figure omitted; refer to PDF] and the number of thresholds is [figure omitted; refer to PDF] . The total number of all the parameters that need to be optimized is 592. The training frequency of network is 1000; the goal of training is 0.01. The minimum value of error of network is as the fitness function. The state of machine failure is divided into 3 types which are, respectively, normal: [figure omitted; refer to PDF] ; crack: [figure omitted; refer to PDF] ; defect: [figure omitted; refer to PDF] . In order to test the trained network, three sets of new data are presented in Table 2.
Table 1: The training sample data.
Sample order | Characteristic value of sample | State |
1 | 0.2286, 0.1292, 0.072, 0.1592, 0.1335, 0.0733, 0.1159, 0.094, 0.0522, 0.1345, 0.009, 0.126, 0.3619, 0.069, 0.1828 | Normal |
2 | 0.209, 0.0947, 0.1393, 0.1387, 0.2558, 0.09, 0.0771, 0.0882, 0.0393, 0.143, 0.0126, 0.167, 0.245, 0.0508, 0.1328 | Normal |
3 | 0.0442, 0.088, 0.1147, 0.0563, 0.3347, 0.115, 0.1453, 0.0429, 0.1818, 0.0378, 0.0092, 0.2251, 0.1516, 0.0858, 0.067 | Normal |
4 | 0.2603, 0.1715, 0.0702, 0.2711, 0.1491, 0.133, 0.0968, 0.1911, 0.2545, 0.0871, 0.006, 0.1793, 0.1002, 0.0789, 0.0909 | Crack |
5 | 0.369, 0.2222, 0.0562, 0.5157, 0.1872, 0.1614, 0.1425, 0.1506, 0.131, 0.05, 0.0078, 0.0348, 0.0451, 0.0707, 0.088 | Crack |
6 | 0.0359, 0.1149, 0.123, 0.546, 0.1977, 0.1248, 0.0624, 0.0832, 0.164, 0.1002, 0.0059, 0.1503, 0.1837, 0.1295, 0.07 | Crack |
7 | 0.1759, 0.2347, 0.1829, 0.1811, 0.2922, 0.0655, 0.0774, 0.02273, 0.2056, 0.0925, 0.0078, 0.1852, 0.3501, 0.168, 0.2668 | Defect |
8 | 0.0724, 0.1909, 0.134, 0.2409, 0.2842, 0.045, 0.0824, 0.1064, 0.1909, 0.1586, 0.0116, 0.1698, 0.3644, 0.2718, 0.2494 | Defect |
9 | 0.2634, 0.2258, 0.1165, 0.1154, 0.1074, 0.0657, 0.061, 0.2623, 0.2588, 0.1155, 0.005, 0.0978, 0.1511, 0.2273, 0.322 | Defect |
Table 2: The test sample data.
Sample order | Characteristic value of sample | State |
1 | 0.2101, 0.095, 0.1298, 0.1359, 0.2601, 0.1001, 0.0753, 0.089, 0.0389, 0.1451, 0.0128, 0.159, 0.2452, 0.0512, 0.1319 | Normal |
2 | 0.2593, 0.18, 0.0711, 0.2801, 0.1501, 0.1298, 0.1001, 0.1891, 0.2531, 0.0875, 0.0058, 0.1803, 0.0992, 0.0802, 0.1002 | Crack |
3 | 0.2599, 0.2235, 0.1201, 0.0071, 0.1102, 0.0683, 0.0621, 0.2597, 0.2602, 0.1167, 0.0048, 0.1002, 0.1521, 0.2281, 0.3205 | Defect |
Use BP, GA-BP, DCQA-BP (BP neural network based on double-chain quantum optimization algorithm), and MCQA-BP (BP neural network based on multichain quantum optimization algorithm, where the number of chain is 4). In GA-BP and DCQA-BP, the population number is set to be 50; the maximal evolutionary generation is set to be 30. The 10 times average values of the performance of algorithms are shown in Table 3.
Table 3: The experiment results of the four algorithms.
Algorithm | Error | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | Average value |
BP | Error 1 | 0.9455 | 1.0252 | 1.1493 | 1.0368 | 0.8473 | 0.9781 | 1.0185 | 1.0967 | 0.9795 | 1.0955 | 1.0172 |
Error 2 | 1.9559 | 1.9604 | 2.0556 | 1.9821 | 1.8676 | 1.9592 | 1.9683 | 2.0947 | 1.9839 | 1.9868 | 1.9815 | |
| ||||||||||||
GA-BP | Error 1 | 0.8828 | 0.3893 | 0.5888 | 0.6893 | 0.8178 | 0.7885 | 0.3947 | 0.5838 | 0.5635 | 0.8384 | 0.6537 |
Error 2 | 1.6535 | 1.421 | 1.3859 | 1.5857 | 1.8434 | 1.6883 | 1.4641 | 1.6759 | 1.5937 | 1.8758 | 1.6187 | |
| ||||||||||||
DCQA-BP | Error 1 | 0.6463 | 0.4027 | 0.5394 | 0.4859 | 0.6673 | 0.6375 | 0.3375 | 0.5473 | 0.5532 | 0.6391 | 0.5456 |
Error 2 | 1.5874 | 1.3867 | 1.3589 | 1.4845 | 1.5838 | 1.5423 | 1.5984 | 1.5873 | 1.4839 | 1.5868 | 1.52 | |
| ||||||||||||
MCQA-BP | Error 1 | 0.5263 | 0.3017 | 0.4194 | 0.3559 | 0.5273 | 0.4275 | 0.5175 | 0.6273 | 0.3332 | 0.3291 | 0.4365 |
Error 2 | 1.5174 | 1.3167 | 1.4389 | 1.3545 | 1.4438 | 1.3123 | 1.5284 | 1.4373 | 1.3239 | 1.3268 | 1.4 |
Apply the test method of mean difference of two normal populations ( [figure omitted; refer to PDF] -test) to analyze the error of BP, GA-BP, DCQA-BP, and MCQA-BP.
The data of error of the four methods can be considered to be samples from normal population [figure omitted; refer to PDF] . The means of the sample of the four methods are [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] , respectively, and the variances are [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] . The [figure omitted; refer to PDF] -statistic is introduced to be the test statistic: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , [figure omitted; refer to PDF] .
The form of rejection region is [figure omitted; refer to PDF] .
Consider [figure omitted; refer to PDF] ; then [figure omitted; refer to PDF] .
So the rejection region is [figure omitted; refer to PDF] .
The hypotheses [figure omitted; refer to PDF] : [figure omitted; refer to PDF] and [figure omitted; refer to PDF] : [figure omitted; refer to PDF] need to be tested.
The mean and variance of Error 1 of BP, GA-BP, DCQA-BP, and MCQA-BP are as follows: [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] ; [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] ; [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] ; [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] . Then, [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] .
Consider [figure omitted; refer to PDF] . So reject [figure omitted; refer to PDF] ; that is, Error 1 of BP is larger than the GA-BP with probability of more than 99.5%.
Consider [figure omitted; refer to PDF] . So reject [figure omitted; refer to PDF] ; that is, Error 1 of GA-BP is larger than the DCQA-BP with probability of more than 90%.
Consider [figure omitted; refer to PDF] . So reject [figure omitted; refer to PDF] ; that is, Error 1 of DCQA-BP is larger than the MCQA-BP with probability of more than 90%.
The mean and variance of Error 2 of BP, GA-BP, DCQA-BP, and MCQA-BP are as follows: [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] ; [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] ; [figure omitted; refer to PDF] , [figure omitted; refer to PDF] ; [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] . Then, [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , and [figure omitted; refer to PDF] .
Consider [figure omitted; refer to PDF] . So reject [figure omitted; refer to PDF] ; that is, Error 2 of BP is larger than the GA-BP with probability of more than 99.5%.
Consider [figure omitted; refer to PDF] . So reject [figure omitted; refer to PDF] ; that is, Error 2 of GA-BP is larger than the DCQA-BP with probability of more than 90%.
Consider [figure omitted; refer to PDF] . So reject [figure omitted; refer to PDF] ; that is, Error 2 of DCQA-BP is larger than the MCQA-BP with probability of more than 99.5%.
The simulation results show that MCQA-BP is significantly superior to BP, GA-BP, and DCQA-BP. The evolutionary process of BP, GA-BP, DCQA-BP, and MCQA-BP is shown in Figure 5.
Figure 5: The evolutionary process of BP, GA-BP, DCQA-BP, and MCQA-BP.
[figure omitted; refer to PDF]
It can be seen that the fitting results of different models are roughly the same. The trend of these algorithms is the same. It shows that the method of this paper is scientific and effective. The fitting error of the model of this paper is less than results of other models. So the model in this paper is better than other models.
For the reason that MCQA-BP with super parallel and ultra-high speed which can overcome a series of problems existing in the BP and GA-BP such as slow convergence speed, premature convergence, and bad computational stability optimizes the structure of the neural network effectively, the computational cost of the MCQA-BP is lower than the method of BP, GA-BP, and DCQA-BP in test case [figure omitted; refer to PDF] and test case [figure omitted; refer to PDF] .
5.2. Test Case 2
The training data is from the results of numerical simulation of 1365 kinds of target board damage under the kinetic energy rod collision of penetration cases. 15 kinds of destruction values predicted by neural network are randomly chosen and shown in Table 4. In the listed penetration value forecast under collision situation, the fitting results of different models are different. The errors between predicted values and the actual values of different models are distinguished from each other. The error of fitting result of MCQA-BP model is less than results of other models. The efficiency of MCQA-BP is further proved.
Table 4: Contrast between 15 kinds of simulation of damage index value and predicted value.
Order number | Value of laboratory | Prediction value of BP | Prediction value of GA-BP | Prediction value of MCQA-BP | Error value of BP | Error value of GA-BP | Error value of MCQA-BP |
1 | 0.5753 | 0.6743 | 0.6642 | 0.5912 | 17.21% | 15.45% | 2.76% |
2 | 0.3735 | 0.3574 | 0.3671 | 0.3719 | 4.31% | 1.71% | 0.43% |
3 | 1.0842 | 1.1218 | 1.1189 | 1.0901 | 3.47% | 3.20% | 0.54% |
4 | 1.2436 | 1.2120 | 1.1203 | 1.2392 | 2.54% | 9.91% | 0.35% |
5 | 1.8643 | 1.9212 | 1.9200 | 1.8765 | 3.05% | 2.99% | 0.65% |
6 | 0.607 | 0.5716 | 0.5825 | 0.5923 | 5.83% | 3.23% | 2.42% |
7 | 0.5394 | 0.6067 | 0.5923 | 0.5532 | 12.48% | 9.81% | 2.56% |
8 | 1.1865 | 1.2077 | 1.2065 | 1.1965 | 1.79% | 1.69% | 0.84% |
9 | 1.3987 | 1.3844 | 1.3865 | 1.3899 | 1.02% | 0.87% | 0.63% |
10 | 1.4826 | 1.4863 | 1.4861 | 1.4830 | 0.25% | 0.24% | 0.03% |
11 | 0.5436 | 0.5627 | 0.5512 | 0.5311 | 3.51% | 1.40% | 2.30% |
12 | 0.0817 | 0.0780 | 0.079 | 0.0813 | 4.53% | 3.30% | 0.49% |
13 | 0.9045 | 0.8847 | 0.8859 | 0.8899 | 2.19% | 2.06% | 1.61% |
14 | 1.4768 | 1.4950 | 1.4854 | 1.4801 | 1.23% | 0.58% | 0.22% |
15 | 1.3164 | 1.3490 | 1.3387 | 1.3214 | 2.48% | 1.69% | 0.38% |
It can be seen from the table that the prediction effect of MCQA-BP is better than BP network and GA-BP on the same sample data. The error range of MCQA-BP network is less than that of BP neural network and GA-BP neural network under the same learning times. The high efficiency of MCQA-BP network model is fully illustrated.
6. Conclusions
A kind of super parallel ultra-fast BP neural network model and algorithm based on multichain quantum optimization algorithm are proposed. The algorithm makes full use of local time domain feature of the BP network and global optimization search capability of multichain quantum optimization algorithm to enhance the intelligent search ability of network. It overcomes the disadvantages of BP network to improve the effectiveness of the optimization and accelerate search efficiency and convergence speed. The model controller based on MCQA-BP network can evaluate damage effect of penetration target board of kinetic energy well. The antijamming ability of the model is fine. The MCQA-BP network controller is effective as can be seen from the actual simulation results.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
[1] S. Yu, K. Zhu, F. Diao, "A dynamic all parameters adaptive BP neural networks model and its application on oil reservoir prediction," Applied Mathematics and Computation , vol. 195, no. 1, pp. 66-75, 2008.
[2] S. S. Haykin Neural Networks and Learning Machines , vol. 3, Pearson Education, Upper Saddle River, NJ, USA, 2009.
[3] Y. Wang, D. Niu, L. Ji, "Short-term power load forecasting based on IVL-BP neural network technology," Systems Engineering Procedia , vol. 4, pp. 168-174, 2012.
[4] X. You, X. Cao, "Study of liquid lithium coolant interaction based on BP neural network optimized by genetic algorithm," Journal of Fusion Energy , vol. 34, no. 4, pp. 918-924, 2015.
[5] D. E. Rumelhart, G. E. Hinton, R. J. Williams, "Learning representations by back-propagating errors," Nature , vol. 323, pp. 533-536, 1986.
[6] W. Jia, D. Zhao, T. Shen, S. Ding, Y. Zhao, C. Hu, "An optimized classification algorithm by BP neural network based on PLS and HCA," Applied Intelligence , vol. 43, no. 1, pp. 176-191, 2015.
[7] S. Y. Kung, Y. H. Hu, "A Frobenius approximation reduction method (FARM) for determining optimal number of hidden units," in Proceedings of the Seattle International Joint Conference on Neural Networks (IJCNN '91), vol. 2, pp. 163-168, IEEE, Seattle, Wash, USA, July 1991.
[8] S. E. Fahlman, "An empirical study of learning speed in back-propagation networks,", no. CMU-CS-88-162, Carnegie-Mellon University, 1988.
[9] R. Salomon, "Adaptive Regelung der Lernratebei back-propagation," Technische Universitat Berlin , no. FB 20, Institut für Softwaretechnik und Theoretische Informatik, 1989.
[10] F. Dan Foresee, M. T. Hagan, "Gauss-Newton approximation to Bayesian learning," in Proceedings of the International Conference on Neural Networks (IJCNN '97), vol. 3, pp. 1930-1935, IEEE, Houston, Tex, USA, June 1997.
[11] M. Riedmiller, H. Braun, "A direct adaptive method for faster backpropagation learning: the RPROP algorithm," in Proceedings of the IEEE International Conference on Neural Networks (IJCNN '93), vol. 1, pp. 586-591, IEEE, San Francisco, Calif, USA, March-April 1993.
[12] H. Wang, B. Chen, S. Yao, "Quantitative structure-activity relationship modeling of angiotensin converting enzyme inhibitors by back propagation artificial neural network," Chinese Journal of Analytical Chemistry , vol. 34, no. 12, pp. 1674-1678, 2006.
[13] C. B. Zhao, Z. L. Wang, X. H. Guo, L. H. Li, "QSAR model of emodin derivatives against the cell line of MCF-7 on artificial neural network," Journal of Shaanxi University of Technology (Natural Science Edition) , vol. 23, no. 4, pp. 64-67, 2007.
[14] X. Cui, Z. Zhang, X. Yuan, Z. Su, S. Liu, "QSAR studies on nitrobenzene derivatives with PCA-ANN," Computers and Applied Chemistry , vol. 22, no. 11, pp. 1038-1040, 2005.
[15] H. Gonzalez-Diaz, I. Bonet, C. Teran, E. De Clercq, R. Bello, M. M. Garcia, L. Santana, E. Uriarte, "ANN-QSAR model for selection of anticancer leads from structurally heterogeneous series of compounds," European Journal of Medicinal Chemistry , vol. 42, no. 5, pp. 580-585, 2007.
[16] F. J. Prado-Prado, X. Garcia-Mera, H. Gonzalez-Diaz, "Multi-target spectral moment QSAR versus ANN for antiparasitic drugs against different parasite species," Bioorganic & Medicinal Chemistry , vol. 18, no. 6, pp. 2225-2231, 2010.
[17] G. Ramirez-Galicia, H. Martinez-Pacheco, R. Garduño-Juarez, O. Deeb, "Exploring QSAR of antiamoebic agents of isolated natural products by MLR, ANN, and RTO," Medicinal Chemistry Research , vol. 21, no. 9, pp. 2501-2516, 2012.
[18] Y. Zhang, G. Ruan, "Bernoulli neural network with weights directly determined and with the number of hidden- layer neurons automatically determined," Advances in Neural Networks-ISNN 2009 , vol. 5551, of Lecture Notes in Computer Science, pp. 36-45, Springer, Berlin, Germany, 2009.
[19] F. M. Ham, I. Kostanic Principles of Neuro Computing for Science & Engineering , McGraw-Hill, New York, NY, USA, 2001.
[20] Z. Sun, X. Wang, J. Zhang, H. Yang, "Prediction and control of equiaxed α in near-β forging of TA15 Ti-alloy based on BP neural network: for purpose of tri-modal microstructure," Materials Science and Engineering A , vol. 591, pp. 18-25, 2014.
[21] Z. Xiao, S.-J. Ye, B. Zhong, C.-X. Sun, "BP neural network with rough set for short term load forecasting," Expert Systems with Applications , vol. 36, no. 1, pp. 273-279, 2009.
[22] Z. Xiao, R. Xu, X. Gong, "A practical method for solving the optimal problems in complicated engineering," Journal of Hydraulic Engineering , vol. 2, pp. 23-27, 1999.
[23] Z. Yang, G. Jiang, H. Wen, "A study on the application of fluid neuron network under nonstandard condition," Systems Engineering-Theory & Practice , vol. 19, no. 8, pp. 54-58, 1999.
[24] H. Ge, "Training the controller based on neural network with genetic algorithm," Journal of South China Normal University (Natural Science Edition) , vol. 1, pp. 33-38, 2000.
[25] J. Li, G. Hu, X. Sun, "Application of a modified BP algorithm of neural networks in missile-testing expert system," Computer Applications and Software , vol. 22, no. 1, pp. 100-102, 2005.
[26] Z. Xie, Q. Du, F. Ren, X. Zhang, S. Jamiesone, "Improving the forecast precision of river stage spatial and temporal distribution using drain pipeline knowledge coupled with BP artificial neural networks: a case study of Panlong River, Kunming, China," Natural Hazards , vol. 77, no. 2, pp. 1081-1102, 2015.
[27] S. Wang, D. Chen, P. Hou, X. Wang, Z. Wang, M. Wei, "Determination of the sodium methylparaben content based on spectrum fluorescence spectral technology and GA-BP neural network," Spectroscopy and Spectral Analysis , vol. 35, no. 6, pp. 1606-1610, 2015.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2015 Baoyu Xu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Abstract
The model and algorithm of BP neural network optimized by expanded multichain quantum optimization algorithm with super parallel and ultra-high speed are proposed based on the analysis of the research status quo and defects of BP neural network to overcome the defects of overfitting, the random initial weights, and the oscillation of the fitting and generalization ability along with subtle changes of the network parameters. The method optimizes the structure of the neural network effectively and can overcome a series of problems existing in the BP neural network optimized by basic genetic algorithm such as slow convergence speed, premature convergence, and bad computational stability. The performance of the BP neural network controller is further improved. The simulation experimental results show that the model is with good stability, high precision of the extracted parameters, and good real-time performance and adaptability in the actual parameter extraction.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer