Erik Cuevas 1,2 and Jorge Galvez 1 and Salvador Hinojosa 1 and Omar Avalos 1 and Daniel Zaldivar 1,2 and Marco Perez-Cisneros 3
Academic Editor:Xin-She Yang
1, Departamento de Electronica, Universidad de Guadalajara, CUCEI, Avenida Revolucion 1500, 44430 Guadalajara, JAL, Mexico
2, Centro Universitario Azteca, Unidad de Investigacion, Avenida Juarez 340, 44280 Guadalajara, JAL, Mexico
3, CUTONALA, Avenida Nuevo Periferico 555, Ejido San Jose Tateposco, 48525 Tonala, JAL, Mexico
Received 18 August 2014; Accepted 30 September 2014; 25 December 2014
This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
System identification is a complex optimization problem which has recently attracted the attention in the field of science and engineering. System identification is important in the disciplines of control systems [1], communication [2], signal processing [3], and image processing [4].
In a system identification configuration, an optimization algorithm attempts to iteratively determine the adaptive model parameters to get an optimal model for an unknown plant by minimizing some error function between the output of the candidate model and the output of the plant. The optimal model or solution is attained when such error function is effectively reduced. The adequacy of the estimated model depends on the adaptive model structure, the optimization algorithm, and also the characteristic and quality of the input-output data [5].
Systems or plants can be better modeled through infinite impulse response (IIR) models because they emulate physical plants more accurately than their equivalent FIR (finite impulse response) models [6]. In addition, IIR models are typically capable of meeting performance specifications using fewer model parameters. However, IIR structures tend to produce multimodal error surfaces whose cost functions are significantly difficult to minimize [7]. Hence, in order to identify IIR models, a practical, efficient, and robust global optimization algorithm is necessary to minimize the multimodal error function.
Traditionally, the least mean square (LMS) technique and its variants [8] have been extensively used as optimization tools for IIR model identification. The wide acceptance of such gradient based optimization techniques is due to the low complexity and simplicity of implementation. However, the error surface for the IIR model is mostly multimodal with respect to the filter coefficients. This may result in leading traditional gradient-descent approaches into local optima [9].
The difficulties associated with the use of gradient based optimization methods for solving several engineering problems have contributed to the development of alternative solutions. Evolutionary computation techniques (ECT) such as the particle swarm optimization (PSO) [10], artificial bee colony (ABC) [11], electromagnetism-like method (EM) [12], cuckoo search (CS) [13], and flower pollination algorithm (FPA) [14] have received much attention regarding their potential as global optimization methods in real-world applications. Inspired by the evolution process and survival of the fittest in the biological world, ECT are search methods that are different from traditional optimization methods. They are based on a collective learning process within a population of candidate solutions. The population in ECT is usually arbitrarily initialized, and each iteration (also called a generation) evolves towards better and better solution regions by means of randomized processes where several operators are applied to each candidate solution. ECT have been applied to many engineering optimization problems and have proven to be effective for solving some specific problems, including multimodal optimization, dynamic optimization, noisy optimization, and multiobjective optimization [15-17]. Hence, they are becoming increasingly popular tools to solve various hard optimization problems.
As an alternative to gradient based techniques, the problem of IIR modelling has also been handled through evolutionary computation techniques. In general, they have been demonstrated to yield better results than those based on gradient algorithms with respect to accuracy and robustness [9]. Such approaches have produced several robust IIR identification systems by using different evolutionary computation techniques such as PSO [18], ABC [19], EM [20], and CS [21], whose results have been individually reported.
ECT are often designed to meet the requirements of particular problems because no single optimization algorithm can solve all problems competitively [22]. Therefore, when new alternative algorithms are proposed, their relative efficiency must be appropriately evaluated. Many efforts [23-25] have also been devoted to comparing ECT to each other. Typically, such comparisons have been based on synthetic numerical benchmark problems with most studies verifying if one algorithm outperforms others over a given set of benchmarks functions overlooking any statistical test. However, few comparative studies of various ECT considering the application context are available in the literature. Therefore, it is very important to discuss and compare the performance of ECT methods from an application point of view.
This paper presents the comparison of various evolutionary computation optimization techniques that are applied to IIR model identification. In the comparison, special attention is paid to recently developed algorithms such as the cuckoo search (CS) and the flower pollination algorithm (FPA), including also popular approaches as the particle swarm optimization (PSO), the artificial bee colony (ABC) optimization, and the electromagnetism-like optimization (EM) algorithm. Results over several models with different ranges of complexity are presented and validated within a statistically significant framework.
The rest of this paper is organized as follows: Section 2 presents a review of the evolutionary computation techniques that are employed in the comparison whereas Section 3 discusses the IIR system identification problem. In Section 4 all experimental results are depicted with some concluding remarks being drawn in Section 5.
2. Evolutionary Computation Techniques (ECT)
In the real world, many optimization problems can be considered as black box challenges. Often, less information is available about an optimization problem itself unless the information emerges from function evaluations. In the worst case, nothing is known about the characteristics of the fitness function, for example, whether it is unimodal or multimodal.
On the other hand, ECT are used to estimate the solution to complex optimization problems since they adapt easily to black-box formulations and extremely ill-behaved functions. ECT are based on a collective learning process within a population of candidate solutions. The population in ECT is usually arbitrarily initialized while each iteration (also called a generation) evolves towards better solution regions by means of randomized processes with several operators being applied to each candidate solution. ECT have been applied to many engineering optimization problems ensuring an effective solution for some specific problems, including multimodal optimization, dynamic optimization, noisy optimization, multiobjective optimization, and others [15-17].
Therefore, ECT are becoming increasingly popular tools to solve various hard optimization problems. This section presents a brief description of five evolutionary computation techniques: swarm optimization (PSO), artificial bee colony (ABC) optimization and electromagnetism-like optimization (EM), cuckoo search (CS), and flower pollination algorithm (FPA), which have been all employed in our comparative study.
2.1. Particle Swarm Optimization (PSO)
PSO, proposed by Kennedy and Eberhart in 1995 [10], is a population-based stochastic optimization technique that is inspired on the social behavior of bird flocking or fish schooling. The algorithm searches for the optimum using a group or swarm formed by possible solutions of the problem, which are called particles. From the implementation point of view, in the PSO operation, a population [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) of [figure omitted; refer to PDF] particles (individuals) evolves from the initial point [figure omitted; refer to PDF] to a total gen number of iterations [figure omitted; refer to PDF] . Each particle [figure omitted; refer to PDF] [figure omitted; refer to PDF] represents a [figure omitted; refer to PDF] -dimensional vector [figure omitted; refer to PDF] where each dimension corresponds to a decision variable of the optimization problem at hand. The quality of each particle [figure omitted; refer to PDF] (candidate solution) is evaluated by using an objective function [figure omitted; refer to PDF] whose final result represents the fitness value of [figure omitted; refer to PDF] . During the evolution process, the best global position [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) seen so far is stored with the best position [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) being reached by each particle. Such positions are computed by considering a minimization problem as follows: [figure omitted; refer to PDF] In this work, the modified PSO version proposed by Lin et al. in [26] has been implemented. Under such approach, the new position [figure omitted; refer to PDF] of each particle [figure omitted; refer to PDF] is calculated by using the following equations: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] is called the inertia weight that controls the impact of the current velocity on the updated velocity ( [figure omitted; refer to PDF] ). [figure omitted; refer to PDF] and [figure omitted; refer to PDF] are the positive acceleration coefficients that rule the movement of each particle towards the positions [figure omitted; refer to PDF] and [figure omitted; refer to PDF] , respectively. [figure omitted; refer to PDF] and [figure omitted; refer to PDF] are uniformly distributed random numbers that are chosen within the interval [figure omitted; refer to PDF] .
2.2. Artificial Bee Colony (ABC)
The artificial bee colony (ABC) algorithm, proposed by Karaboga [11], is an ECT inspired by the intelligent foraging behavior of a honeybee swarm. In the ABC operation, a population [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) of [figure omitted; refer to PDF] food locations (individuals) is evolved from the initial point [figure omitted; refer to PDF] to a total gen number of iterations [figure omitted; refer to PDF] . Each food location [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) represents a [figure omitted; refer to PDF] -dimensional vector [figure omitted; refer to PDF] where each dimension corresponds to a decision variable of the optimization problem to be solved. After initialization, an objective function evaluates each food location to assess whether it represent an acceptable solution (nectar-amount) or not. Guided by the values of such an objective function, the candidate solution [figure omitted; refer to PDF] is evolved through different ABC operations (honeybee types). In the main operator, each food location [figure omitted; refer to PDF] generates a new food source [figure omitted; refer to PDF] in the neighborhood of its present position as follows: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] is a randomly chosen food location, satisfying the condition [figure omitted; refer to PDF] . The scale factor [figure omitted; refer to PDF] is a random number between [figure omitted; refer to PDF] . Once a new solution [figure omitted; refer to PDF] is generated, a fitness value representing the profitability associated with a particular solution [figure omitted; refer to PDF] is calculated. The fitness value for a minimization problem can be assigned to a candidate solution [figure omitted; refer to PDF] by the following expression: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] represents the objective function to be minimized. Once the fitness values are calculated, a greedy selection process is applied between [figure omitted; refer to PDF] and [figure omitted; refer to PDF] . If [figure omitted; refer to PDF] is better than [figure omitted; refer to PDF] , then the candidate solution [figure omitted; refer to PDF] is replaced by [figure omitted; refer to PDF] ; otherwise, [figure omitted; refer to PDF] remains.
2.3. Electromagnetism-Like (EM) Algorithm
The EM algorithm, proposed by Ilker et al. [12] is a simple and population-based search algorithm which has been inspired by the electromagnetism phenomenon. In EM, individuals emulate charged particles which interact to each other based on the electromagnetism laws of repulsion and attraction. The method utilizes [figure omitted; refer to PDF] , [figure omitted; refer to PDF] -dimensional points [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , where each point [figure omitted; refer to PDF] is a [figure omitted; refer to PDF] -dimensional vector containing the parameter values to be optimized [figure omitted; refer to PDF] whereas [figure omitted; refer to PDF] denotes the iteration (or generation) number. The initial population [figure omitted; refer to PDF] (being [figure omitted; refer to PDF] ) is taken from uniformly distributed samples of the search space. We denote the population set at the [figure omitted; refer to PDF] th generation by [figure omitted; refer to PDF] , because members of [figure omitted; refer to PDF] change with [figure omitted; refer to PDF] . After the initialization of [figure omitted; refer to PDF] , EM continues its iterative process until a stopping condition (e.g., the maximum number of generations, [figure omitted; refer to PDF] ) is met. An iteration of EM consists of three steps. In the first step each point in [figure omitted; refer to PDF] moves to a different location by using the attraction-repulsion mechanism of the electromagnetism theory. In the second step, points moved by the electromagnetism principle are further moved locally by a local search procedure. Finally, in the third step, in order to generate the new population [figure omitted; refer to PDF] , a greedy selection process selects best points between those produced by the local search procedure and the originals. Both the attraction-repulsion mechanism and the local search in EM are responsible for driving the members [figure omitted; refer to PDF] of [figure omitted; refer to PDF] to the proximity of the global optimum.
2.4. Cuckoo Search (CS) Method
CS is one of the latest nature-inspired algorithms that has been developed by Yang and Deb [13]. CS is based on the brood parasitism of some cuckoo species. In addition, this algorithm is enhanced by the so-called Levy flights [27], rather than by simple isotropic random walks. From the implementation point of view of the CS operation, a population [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) of [figure omitted; refer to PDF] eggs (individuals) is evolved from the initial point [figure omitted; refer to PDF] to a total gen number of iterations [figure omitted; refer to PDF] . Each egg [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) represents a [figure omitted; refer to PDF] -dimensional vector [figure omitted; refer to PDF] where each dimension corresponds to a decision variable of the optimization problem to be solved. The quality of each egg [figure omitted; refer to PDF] (candidate solution) is evaluated by using an objective function [figure omitted; refer to PDF] whose final result represents the fitness value of [figure omitted; refer to PDF] . Three different operators define the evolution process of CS: (A) Levy flight, (B) the replacing of nests operator for constructing new solutions, and (C) the elitist selection strategy.
(A) The Levy Flight . One of the most powerful features of cuckoo search is the use of Levy flights to generate new candidate solutions (eggs). Under this approach, a new candidate solution [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) is produced by perturbing the current [figure omitted; refer to PDF] with a change of position [figure omitted; refer to PDF] . In order to obtain [figure omitted; refer to PDF] , a random step [figure omitted; refer to PDF] is generated by a symmetric Levy distribution. For producing [figure omitted; refer to PDF] , Mantegna's algorithm [28] is employed as follows: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] and [figure omitted; refer to PDF] are [figure omitted; refer to PDF] -dimensional vectors and [figure omitted; refer to PDF] . Each element of [figure omitted; refer to PDF] and [figure omitted; refer to PDF] is calculated by considering the following normal distributions: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] represents the Gamma distribution. Once [figure omitted; refer to PDF] has been calculated, the required change of position [figure omitted; refer to PDF] is computed as follows: [figure omitted; refer to PDF] where the product [figure omitted; refer to PDF] denotes entrywise multiplications whereas [figure omitted; refer to PDF] is the best solution (egg) seen so far in terms of its fitness value. Finally, the new candidate solution [figure omitted; refer to PDF] is calculated by using [figure omitted; refer to PDF]
(B) Replacing Some Nests by Constructing New Solutions. Under this operation, a set of individuals (eggs) are probabilistically selected and replaced with a new value. Each individual [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) can be selected with a probability [figure omitted; refer to PDF] . In order to implement this operation, a uniform random number [figure omitted; refer to PDF] is generated within the range [figure omitted; refer to PDF] . If [figure omitted; refer to PDF] is less than [figure omitted; refer to PDF] , the individual [figure omitted; refer to PDF] is selected and modified according to (5); otherwise [figure omitted; refer to PDF] remains with no change. This operation can be resumed by the following model: [figure omitted; refer to PDF] where rand is a random number normally distributed whereas [figure omitted; refer to PDF] and [figure omitted; refer to PDF] are random integers from 1 to [figure omitted; refer to PDF] .
(C) The Elitist Selection Strategy. After producing [figure omitted; refer to PDF] either by the operator A or by the operator B, it must be compared with its past value [figure omitted; refer to PDF] . If the fitness value of [figure omitted; refer to PDF] is better than [figure omitted; refer to PDF] , then [figure omitted; refer to PDF] is accepted as the final solution; otherwise, [figure omitted; refer to PDF] is retained. This procedure can be resumed by the following statement: [figure omitted; refer to PDF] The elitist selection strategy denotes that only high-quality eggs (best solutions near to the optimal value) which are the most similar to the host bird's eggs have the opportunity to develop (next generation) and become mature cuckoos.
2.5. Flower Pollination Algorithm (FPA)
The flower pollination algorithm (FPA), proposed by Yang [14], is an ECT inspired by the pollination process of flowers. In FPA, individuals emulate a set of flowers or pollen gametes which behaves based on biological laws of the pollination process. From the implementation point of view, in the FPA operation, a population [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) of [figure omitted; refer to PDF] flower positions (individuals) is evolved from the initial point [figure omitted; refer to PDF] to a total gen number of iterations [figure omitted; refer to PDF] . Each flower [figure omitted; refer to PDF] ( [figure omitted; refer to PDF] ) represents a [figure omitted; refer to PDF] -dimensional vector [figure omitted; refer to PDF] where each dimension corresponds to a decision variable of the optimization problem to be solved. In FPA, a new population [figure omitted; refer to PDF] is produced by considering two operators: local and global pollination. A probabilistic global pollination factor [figure omitted; refer to PDF] is associated with such operators. In order to decide which operator should be applied to each current flower position [figure omitted; refer to PDF] , a uniform random number [figure omitted; refer to PDF] is generated within the range [figure omitted; refer to PDF] . If [figure omitted; refer to PDF] is less than [figure omitted; refer to PDF] , the global pollination operator is applied to [figure omitted; refer to PDF] . Otherwise, the local pollination operator is considered.
Global Pollination Operator. Under this operator, the original position [figure omitted; refer to PDF] is displaced to a new position [figure omitted; refer to PDF] according to the following model: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] is the global best position seen so far whereas [figure omitted; refer to PDF] controls the length of the displacement. The [figure omitted; refer to PDF] value is generated by a symmetric Levy distribution according to (5)-(6).
Local Pollination Operator. In the local pollination operator, the current position [figure omitted; refer to PDF] is perturbed to a new position [figure omitted; refer to PDF] as follows: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] and [figure omitted; refer to PDF] are two randomly chosen flower positions, satisfying the condition [figure omitted; refer to PDF] . The scale factor [figure omitted; refer to PDF] is a random number between [figure omitted; refer to PDF] .
3. IIR Model Identification (Problem Formulation)
System identification is the mathematical representation of an unknown system by using only input-output data. In a system identification configuration, an optimization algorithm attempts to iteratively determine the adaptive model parameters to get an optimal model for the unknown plant based on minimizing some error function between the output of the candidate model and the actual output of the plant.
The use of infinite impulse response (IIR) models for identification is preferred over their equivalent FIR (finite impulse response) models since the former produce more accurate models of physical plants for real world applications [6]. In addition, IIR models are typically capable of meeting performance specifications using fewer model parameters. Figure 1 represents an IIR identification model of any arbitrary system.
Figure 1: Adaptive IIR model for system identification.
[figure omitted; refer to PDF]
An IIR system can be represented by the transfer function: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] and [figure omitted; refer to PDF] are the number of numerator and denominator coefficients of the transfer function, respectively, and, [figure omitted; refer to PDF] and [figure omitted; refer to PDF] are the pole and zero parameters of the IIR model ( [figure omitted; refer to PDF] ). Equation (13) can be written as difference equation of the form: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] and [figure omitted; refer to PDF] represent the [figure omitted; refer to PDF] th input and output of the system, respectively. Therefore, the set of unknown parameters that models the IIR system is represented by [figure omitted; refer to PDF] . Considering that the number of unknown parameters of [figure omitted; refer to PDF] is [figure omitted; refer to PDF] , the search space [figure omitted; refer to PDF] of feasible values for [figure omitted; refer to PDF] is [figure omitted; refer to PDF] .
According to the block diagram of Figure 1, the output of the plant is [figure omitted; refer to PDF] whereas the output of the IIR filter is [figure omitted; refer to PDF] . The output difference between the actual system and its model yields the error [figure omitted; refer to PDF] . Hence, the problem of IIR model identification can be considered as a minimization problem of the function [figure omitted; refer to PDF] stated as the following: [figure omitted; refer to PDF] where [figure omitted; refer to PDF] is the number of samples used in the simulation.
The aim is to minimize the cost function [figure omitted; refer to PDF] by adjusting [figure omitted; refer to PDF] . The optimal model [figure omitted; refer to PDF] or solution is attained when the error function [figure omitted; refer to PDF] reaches its minimum value, as follows: [figure omitted; refer to PDF]
4. Experimental Results
In the comparison study, a comprehensive set of experiments has been used to test the performance of each evolutionary computation technique. The set considers the use of IIR models with different orders. Such experimental set has been carefully selected to assure compatibility between similar works reported in the literature [18-21]. In the comparison, five ETC have been considered: PSO, ABC, EM, CS, and FPA.
The parameter setting for each evolutionary computation algorithm that is used in the comparison is described as follows.
(1) PSO: the parameters are set to [figure omitted; refer to PDF] , [figure omitted; refer to PDF] ; besides, the weight factor decreases linearly from 0.9 to 0.2 [18].
(2) ABC: the algorithm has been implemented using the guidelines provided by its own reference [19], using the parameter [figure omitted; refer to PDF] .
(3) EM: particle number = 50, [figure omitted; refer to PDF] , [figure omitted; refer to PDF] , [figure omitted; refer to PDF] . Such values, according to [12, 20] represent the best possible configuration.
(4) CS: according to [13, 21], the parameters are set to [figure omitted; refer to PDF] and the number of generations [figure omitted; refer to PDF] .
(5) FPA: the probabilistic global pollination factor [figure omitted; refer to PDF] is set to 0.8. Under such value, the algorithm presents the best performance according to [14].
For all algorithms, the population size has been set to 25 ( [figure omitted; refer to PDF] ) whereas the maximum iteration number has been configured to 3000 generations ( [figure omitted; refer to PDF] ).
The results are divided into two sections. In the first set, the performance of each ETC for each identification experiment is presented. In the second set, the results are analyzed from a statistical point of view by using the Wilcoxon test.
4.1. IIR Model Identification Results
The results are reported considering three experiments that include [figure omitted; refer to PDF] a second-order plant with a first-order IIR model; [figure omitted; refer to PDF] a second-order plant with a second-order IIR model; and finally, [figure omitted; refer to PDF] a high-order plant with a high-order model. Each case is discussed below.
( 1) A Plant with a Second-Order System and a First-Order IIR Model (First Experiment). In this experiment, each algorithm is applied to identify a second-order plant through a first-order IIR model. Under such conditions, the unknown plant [figure omitted; refer to PDF] and the IIR model [figure omitted; refer to PDF] hold the following transfer functions: [figure omitted; refer to PDF] In the simulations, it has been considered a white sequence of 100 samples [figure omitted; refer to PDF] for the input [figure omitted; refer to PDF] . Since a reduced order model is employed to identify a plant of a superior order, [figure omitted; refer to PDF] is multimodal [19]. The error surface [figure omitted; refer to PDF] is shown in Figure 2.
Multimodal error surface [figure omitted; refer to PDF] for the first experiment: (a) tridimensional figure and (b) contour.
(a) [figure omitted; refer to PDF]
(b) [figure omitted; refer to PDF]
The performance evaluation over 30 different executions is reported in Table 1 considering the following indexes: the best parameter values (ABP), the average [figure omitted; refer to PDF] value (AV), and the standard deviation (SD). The best parameter values (ABP) report the best model parameters obtained during the 30 executions while the average [figure omitted; refer to PDF] value (AV) indicates the average minimum value of [figure omitted; refer to PDF] , considering the same number of executions. Finally, the standard deviation (SD) reports the dispersion from the average [figure omitted; refer to PDF] value regarding 30 executions.
Table 1: Performance results of the first experiment.
Algorithms | ABP | AV | SD | |
[figure omitted; refer to PDF] | [figure omitted; refer to PDF] | |||
PSO | 0.9125 | [figure omitted; refer to PDF] | 0.0284 | 0.0105 |
ABC | 0.1420 | [figure omitted; refer to PDF] | 0.0197 | 0.0015 |
EM | 0.9034 | 0.3030 | 0.0165 | 0.0012 |
CS | 0.9173 | [figure omitted; refer to PDF] | 0.0101 | [figure omitted; refer to PDF] |
FPA | 0.9364 | [figure omitted; refer to PDF] | 0.0105 | [figure omitted; refer to PDF] |
According to Table 1, the CS algorithm provides better results than PSO, ABC, and EM. In particular, the results show that CS maintains a considerable precision (the lowest AV value) and more robustness (smallest SD value). Nevertheless, the CS performance is similar to the FPA algorithm. On the other hand, the worst performance is reached by the PSO algorithm. Such a fact corresponds to its difficulty (premature convergence) to overcome local minima in multimodal functions.
( 2) A Plant with Second-Order System and Second-Order IIR Model (Second Experiment). In the second experiment, the performance for each algorithm is evaluated at the identification of a second-order plant through a second-order IIR model. Therefore, the unknown plant [figure omitted; refer to PDF] and the IIR model [figure omitted; refer to PDF] hold the following transfer functions: [figure omitted; refer to PDF] For the simulations, the input [figure omitted; refer to PDF] that is applied to the system and to the IIR model simultaneously has been configured as a white sequence with 100 samples. Since the order of the model [figure omitted; refer to PDF] is equal to the order of the to-be-identified system [figure omitted; refer to PDF] , only one global minimum exists in [figure omitted; refer to PDF] [19]. The results of this experiment over 30 different executions are reported in Table 2.
Table 2: Performance results of the second experiment.
Algorithms | ABP | AV | SD | ||
[figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | |||
PSO | [figure omitted; refer to PDF] | 0.4925 | 0.9706 | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] |
ABC | [figure omitted; refer to PDF] | 0.6850 | 0.2736 | 0.3584 | 0.1987 |
EM | [figure omitted; refer to PDF] | 0.4802 | 1.0091 | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] |
CS | [figure omitted; refer to PDF] | 0.4900 | 1.000 | 0.000 | 0.000 |
FPA | [figure omitted; refer to PDF] | 0.4900 | 1.000 | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] |
The results in Table 2 show that PSO, ABC, EM, CS, and FPA have similar values in their performance. The evidence shows that evolutionary algorithms maintain a similar average performance when they face unimodal low-dimensional functions [29, 30]. In particular, the test remarks that the small difference in performance is directly related to a better exploitation mechanism included in CS and FPA.
( 3) A Superior-Order Plant and a High-Order Model (Third Experiment) . Finally, the performance for each algorithm is evaluated at the identification of a superior-order plant through a high-order IIR model. Therefore, the unknown plant [figure omitted; refer to PDF] and the IIR model [figure omitted; refer to PDF] hold the following transfer functions: [figure omitted; refer to PDF] Since the plant is a sixth-order system and the IIR model a fourth-order system, the error surface [figure omitted; refer to PDF] is multimodal just as it is in the first experiment. A white sequence with 100 samples has been used as input. The results of this experiment over 30 different executions are reported in Tables 3 and 4. Table 3 presents the best parameter values (ABP) whereas Table 4 shows the average [figure omitted; refer to PDF] value (AV) and its standard deviation (SD).
Table 3: The best parameter values (ABP) for the second experiment.
Algorithms | ABP | ||||||||
[figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | |
PSO | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] |
ABC | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] |
EM | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] |
CS | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] |
FPA | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] |
Table 4: The average [figure omitted; refer to PDF] value (AV) and the standard deviation (SD).
Algorithms | AV | SD |
PSO | 5.8843 | 3.4812 |
ABC | 7.3067 | 4.3194 |
EM | 0.0140 | 0.0064 |
CS | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] |
FPA | 0.0018 | 0.0020 |
According to the AV and SD indexes in Table 4, the CS algorithm finds better results than PSO, ABC, EM, and FPA. The results show that CS presents better precision (AV value) and robustness (SD value). These results also indicate that CS, FPA, and EM are able to identify the sixth-order plant under different accuracy levels. On the other hand, PSO and ABC obtain suboptimal solutions whose parameters weakly model the unknown system.
4.2. Statistical Analysis
In order to statistically validate the results, a nonparametric statistical significance-proof which is known as Wilcoxon's rank sum test for independent samples [31, 32] has been conducted over the "the average [figure omitted; refer to PDF] value" (AV) data of Tables 1, 2, and 4 with a 5% significance level. The test has been conducted considering 30 different executions for each algorithm. Table 5 reports the [figure omitted; refer to PDF] values produced by Wilcoxon's test for the pairwise comparison of the "the average [figure omitted; refer to PDF] value" of four groups. Such groups are formed by CS versus PSO, CS versus ABC, CS versus EM, and CS versus FPA. As a null hypothesis, it is assumed that there is no significant difference between averaged values of the two algorithms. The alternative hypothesis considers a significant difference between the AV values of both approaches.
Table 5: [figure omitted; refer to PDF] -values produced by Wilcoxon's test comparing CS vs PSO, ABC, EM and FPA over the "The average [figure omitted; refer to PDF] values (AV)" from Tables 1, 2 and 4.
CS vs | PSO | ABC | EM | FPA |
First experiment | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | 0.7870 |
Second experiment | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | 0.3313 |
Third experiment | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | [figure omitted; refer to PDF] | 0.1011 |
For the case of PSO, ABC, and EM, all [figure omitted; refer to PDF] values reported in Table 5 are less than 0.05 (5% significance level) which is a strong evidence against the null hypothesis. Therefore, such evidence indicates that CS results are statistically significant and that it has not occurred by coincidence (i.e., due to common noise contained in the process). On the other hand, since the [figure omitted; refer to PDF] values for the case of CS versus FPA are more than 0.05, there is not statistical difference between both. Therefore, it can be concluded that the CS algorithm is better than PSO, ABC, and EM in the application of IIR modeling for system identification. However, CS presents the same performance as FPA and therefore there is not statistical evidence that CS surpasses the FPA algorithm.
5. Conclusions
This paper presents a comparison study between five evolutionary algorithms for the IIR-based model identification. Under this research, the identification task is considered as an optimization problem. In the comparison, special attention is paid to recently developed algorithms such as the cuckoo search (CS) and the flower pollination algorithm (FPA), also including popular approaches such as the particle swarm optimization (PSO), the artificial Bee colony optimization (ABC), and the electromagnetism-like (EM) optimization algorithm.
The comparison has been experimentally evaluated over a test suite of three benchmark experiments that produce multimodal functions. The experiment results have demonstrated that CS outperforms PSO, ABC, and EM in terms of both the accuracy (AV values) and robustness (SD values), within a statistically significant framework (Wilcoxon test). However, there is not statistical evidence that CS surpasses the FPA performance.
The remarkable performance of CS and FPA is explained by two different features: (i) operators (such as Levy flight) that allow a better exploration of the search space, increasing the capacity to find multiple optima, and (ii) their exploitation operators that allow a better precision of previously found solutions.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
[1] X. Zhou, C. Yang, W. Gui, "Nonlinear system identification and control using state transition algorithm," Applied Mathematics and Computation , vol. 226, pp. 169-179, 2014.
[2] M. Albaghdadi, B. Briley, M. Evens, "Event storm detection and identification in communication systems," Reliability Engineering and System Safety , vol. 91, no. 5, pp. 602-613, 2006.
[3] P. Frank Pai, B.-A. Nguyen, M. J. Sundaresan, "Nonlinearity identification by time-domain-only signal processing," International Journal of Non-Linear Mechanics , vol. 54, pp. 85-98, 2013.
[4] H.-C. Chung, J. Liang, S. Kushiyama, M. Shinozuka, "Digital image processing for non-linear system identification," International Journal of Non-Linear Mechanics , vol. 39, no. 5, pp. 691-707, 2004.
[5] J. Na, X. Ren, Y. Xia, "Adaptive parameter identification of linear SISO systems with unknown time-delay," Systems & Control Letters , vol. 66, no. 1, pp. 43-50, 2014.
[6] O. Kukrer, "Analysis of the dynamics of a memoryless nonlinear gradient IIR adaptive notch filter," Signal Processing , vol. 91, no. 10, pp. 2379-2394, 2011.
[7] T. Mostajabi, J. Poshtan, Z. Mostajabi, "IIR model identification via evolutionary algorithms-a comparative study," Artificial Intelligence Review , 2013.
[8] C. Dai, W. Chen, Y. Zhu, "Seeker optimization algorithm for digital IIR filter design," IEEE Transactions on Industrial Electronics , vol. 57, no. 5, pp. 1710-1718, 2010.
[9] W. Fang, J. Sun, W. Xu, "A new mutated quantum-behaved particle swarm optimizer for digital IIR filter design," Eurasip Journal on Advances in Signal Processing , vol. 2009, 2009.
[10] J. Kennedy, R. C. Eberhart, "Particle swarm optimization," in Proceedings of the IEEE International Conference on Neural Networks, vol. 4, pp. 1942-1948, December 1995.
[11] D. Karaboga, "An idea based on honey bee swarm for numerical optimization,", no. TR06, Computer Engineering Department, Engineering Faculty, Erciyes University, 2005.
[12] B. Ilker, S. Birbil, F. Shu-Cherng, "An electromagnetism-like mechanism for global optimization," Journal of Global Optimization , vol. 25, no. 3, pp. 263-282, 2003.
[13] X.-S. Yang, S. Deb, "Cuckoo search via Levy flights," in Proceedings of the World Congress on Nature and Biologically Inspired Computing (NABIC '09), pp. 210-214, December 2009.
[14] X. S. Yang, "Flower pollination algorithm for global optimization," Unconventional Computation and Natural Computation , vol. 7445, of Lecture Notes in Computer Science, pp. 240-249, Springer, Berlin, Germany, 2012.
[15] C. Ahn Advances in Evolutionary Algorithms: Theory, Design and Practice , Springer, New York, NY, USA, 2006.
[16] R. Chiong, T. Weise, Z. Michalewicz Variants of Evolutionary Algorithms for Real-World Applications , Springer, New York, NY, USA, 2012.
[17] M. Oltean, "Evolving evolutionary algorithms with patterns," Soft Computing , vol. 11, no. 6, pp. 503-518, 2007.
[18] S. Chen, B. L. Luk, "Digital IIR filter design using particle swarm optimisation," International Journal of Modelling, Identification and Control , vol. 9, no. 4, pp. 327-335, 2010.
[19] N. Karaboga, "A new design method based on artificial bee colony algorithm for digital IIR filters," Journal of the Franklin Institute , vol. 346, no. 4, pp. 328-348, 2009.
[20] E. Cuevas, D. Oliva, "IIR filter modeling using an algorithm inspired on electromagnetism," Ingenieria Investigacion y Tecnologia , vol. 14, no. 1, pp. 125-138, 2013.
[21] A. P. Patwardhan, R. Patidar, N. V. George, "On a cuckoo search optimization approach towards feedback system identification," Digital Signal Processing , vol. 32, pp. 156-163, 2014.
[22] D. H. Wolpert, W. G. Macready, "No free lunch theorems for optimization," IEEE Transactions on Evolutionary Computation , vol. 1, no. 1, pp. 67-82, 1997.
[23] E. Elbeltagi, T. Hegazy, D. Grierson, "Comparison among five evolutionary-based optimization algorithms," Advanced Engineering Informatics , vol. 19, no. 1, pp. 43-53, 2005.
[24] D. Shilane, J. Martikainen, S. Dudoit, S. J. Ovaska, "A general framework for statistical performance comparison of evolutionary computation algorithms," Information Sciences , vol. 178, no. 14, pp. 2870-2879, 2008.
[25] V. Osuna-Enciso, E. Cuevas, H. Sossa, "A comparison of nature inspired algorithms for multi-threshold image segmentation," Expert Systems with Applications , vol. 40, no. 4, pp. 1213-1219, 2013.
[26] Y.-L. Lin, W.-D. Chang, J.-G. Hsieh, "A particle swarm optimization approach to nonlinear rational filter modeling," Expert Systems with Applications , vol. 34, no. 2, pp. 1194-1199, 2008.
[27] I. Pavlyukevich, "Levy flights, non-local search and simulated annealing," Journal of Computational Physics , vol. 226, no. 2, pp. 1830-1844, 2007.
[28] R. N. Mantegna, "Fast, accurate algorithm for numerical simulation of Levy stable stochastic processes," Physical Review E , vol. 49, no. 5, pp. 4677-4683, 2007.
[29] E. Cuevas, M. Gonzalez, D. Zaldivar, M. Perez-Cisneros, G. Garcia, "An algorithm for global optimization inspired by collective animal behavior," Discrete Dynamics in Nature and Society , vol. 2012, 2012.
[30] E. Cuevas, M. Cienfuegos, D. Zaldivar, M. Perez-Cisneros, "A swarm optimization algorithm inspired in the behavior of the social-spider," Expert Systems with Applications , vol. 40, no. 16, pp. 6374-6384, 2013.
[31] S. Garcia, D. Molina, M. Lozano, F. Herrera, "A study on the use of non-parametric tests for analyzing the evolutionary algorithms' behaviour: a case study on the CEC'2005 special session on real parameter optimization," Journal of Heuristics , vol. 15, no. 6, pp. 617-644, 2009.
[32] D. Shilane, J. Martikainen, S. Dudoit, S. J. Ovaska, "A general framework for statistical performance comparison of evolutionary computation algorithms," Information Sciences , vol. 178, no. 14, pp. 2870-2879, 2008.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2014 Erik Cuevas et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Abstract
System identification is a complex optimization problem which has recently attracted the attention in the field of science and engineering. In particular, the use of infinite impulse response (IIR) models for identification is preferred over their equivalent FIR (finite impulse response) models since the former yield more accurate models of physical plants for real world applications. However, IIR structures tend to produce multimodal error surfaces whose cost functions are significantly difficult to minimize. Evolutionary computation techniques (ECT) are used to estimate the solution to complex optimization problems. They are often designed to meet the requirements of particular problems because no single optimization algorithm can solve all problems competitively. Therefore, when new algorithms are proposed, their relative efficacies must be appropriately evaluated. Several comparisons among ECT have been reported in the literature. Nevertheless, they suffer from one limitation: their conclusions are based on the performance of popular evolutionary approaches over a set of synthetic functions with exact solutions and well-known behaviors, without considering the application context or including recent developments. This study presents the comparison of various evolutionary computation optimization techniques applied to IIR model identification. Results over several models are presented and statistically validated.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer





