I JSC 040201
I JSC 040201
I JSC 040201
2, May 2013
ABSTRACT
Particle Swarm Optimizer (PSO) is such a complex stochastic process so that analysis on the stochastic behavior of the PSO is not easy. The choosing of parameters plays an important role since it is critical in the performance of PSO. As far as our investigation is concerned, most of the relevant researches are based on computer simulations and few of them are based on theoretical approach. In this paper, theoretical approach is used to investigate the behavior of PSO. Firstly, a state of PSO is defined in this paper, which contains all the information needed for the future evolution. Then the memory-less property of the state defined in this paper is investigated and proved. Secondly, by using the concept of the state and suitably dividing the whole process of PSO into countable number of stages (levels), a stationary Markov chain is established. Finally, according to the property of a stationary Markov chain, an adaptive method for parameter selection is proposed.
KEYWORDS
Markov chain, Memory-less property, Order Statistics, Particle Swarm Optimizer, Percentile, Stationary Markov chain
1. INTRODUCTION
The Particle Swarm Optimizer (PSO), introduced by Kennedy and Eberhart, is a stochastic optimization technique likened to the behavior of a flock of birds or the sociological behavior of a group of people [10], [11]. PSO is a population based optimization technique. The population is called a swarm. PSO is motivated from the simulation of social behavior of the group of individuals and therefore different from other evolutionary computational methods. It generates new values for all particles (individuals) in the swarm (population). PSO memorizes the previous individual and social (swarm) experience and uses them to help generate new particles. This idea for generation of new individuals differentiates PSO from other population-based algorithms. PSO is a high performance optimizer that possesses several desirable attributes, including the fact that the basic algorithm is very easy to understand and implement. It is similar in some ways to genetic algorithm (GA) and evolutionary algorithms. But it requires very less computational bookkeeping and fewer lines of code. Hence, it gains increasing popularity with its wide applicability and effectiveness in performing difficult optimization tasks. Among other applications, it has been applied to many problems, including multi-objective problems, mini-max problems, integer programming problems, errors-in-variables problems, and numerous
DOI: 10.5121/ijsc.2013.4201 1
engineering applications. It also shares the ability of the genetic algorithm to handle arbitrary nonlinear cost functions, but with a much simpler implementation. It clearly demonstrates good possibilities for widespread use in electromagnetic optimization [3]. Though PSO has the advantage of being easy, it suffer from the curse of dimensionality just like other stochastic optimization algorithms including GA. This implies that the performance deteriorates as the dimensions of the search space increase. It does not possess the ability to improve upon the quality of the solutions as the number of generations was increased. We give a simple explanation in the following. Consider a basic stochastic global search algorithm that generates a sample from a uniform distribution on the entire search space. The algorithm stops when it generates a solution that falls in the optimality region, a small volume of the search space surrounding the global optimum such that the fitness is under a given threshold. The probability of generating a sample inside the optimality region is simply the volume of the optimality region divided by the volume of the search space. This probability will decrease exponentially as the dimensionality of the search space increases. Given this explanation, it is clear that it is typically significantly harder to find the global optimum of a high-dimensional problem, compared with a low-dimensional problem with similar topology. Beside the curse of dimensionality, another drawback of PSO is associated with the lack of a mechanism responsible for the control of the magnitude of the velocities, which may incur the danger of swarm explosion and divergence. To address the explosion problem a threshold on the absolute value of the velocity that can be assumed by any particle was incorporated. Also, PSO suffers from being trapped in local minima or having slower convergence rate. To overcome the above difficulties, many researches are devoted to improving the performance of the original PSO. For example, the original PSO does not have an inertia weight, i.e., (r ) = 0 in (6), while Shi and Eberhart [23] introduced the improvement by adding an inertia weight, which results in faster convergence and increases the overall performance of PSO. Angeline [1] introduced a form of selection such that the properties making some solutions superior are transferred directly to some less effective particles. Angeline used a tournament selection process based on the particles current fitness, copying the current positions and velocities of the better half of the population onto the worse half. This technique improved the performance of PSO in three of the four test functions. Kennedy [17] proposed the LBEST method by dividing the swarm into multiple neighborhoods, with each neighborhood maintaining its own local best solution. This approach is less prone to becoming trapped in local minima, but typically has slower convergence. Since PSO and Genetic Algorithm (GA) both work with a population of solutions, Bergh and Engelbrecht [2] combined the searching abilities of both methods. Based on the compensatory property of GA and PSO, they propose a new algorithm, a variation on the traditional PSO algorithm, called the cooperative particle swarm optimizer (CPSO) that combines the evolution ideas of both. On the other side Parsopoulos and Vrahatis [22] propose a technique aiming to compute all global optimizers, while avoiding local optimizers, through PSO. Because the existing methods are all simulation-based, in this paper the theoretical part of PSO will be the main concern. A simple explanation of the PSO operation is as follows. Each particle in the whole swarm represents a possible solution to the optimization task at hand. During iterations each particle accelerates in the direction of its own personal (individual) best solution found so far, as well as in the direction of the global best position discovered so far by all the particles in the swarm. This means that if a particle discovers a promising new solution, all the other particles will move closer to it, exploring the region more thoroughly in the process.
Instead of using evolutionary operators to manipulate the individuals, like in evolutionary computational algorithms, each particle in PSO flies in the search space with a velocity that is dynamically adjusted according to its own flying experience and its companions flying experience. Through this paper, we assume a D-dimensional search space S with
S = (a1 , b1 ) (a2 , b2 ) L (a D , bD ),
(1)
where a1 , a2 , L a D and b1 , b2 , L bD are the lower bounds and upper bounds of respective dimensions. We have a swarm consisting of I particles on S. Let i be the index of particle and r be the index of iteration. The position of the ith particle after its rth iteration is a D-dimensional random vector
~ X i ( r ) = ( X i ,1 ( r ), X i , 2 ( r ), L , X i ,D ( r )) S , i = 1,2, L , I ; r = 1,2, L , R.
(2)
The velocity acting on the ith particle at its rth iteration is also a D-dimensional random vector
(3)
The personal best position of the ith particle after its rth iteration is recorded and represented as
~ P P X iP (r ) = ( X iP ,1 ( r ), X i , 2 ( r ), L , X i , D ( r )) S , i = 1,2, L , I ; r = 1,2, L , R.
The global best position after the rth iteration of the ith particle is recorded and represented as
(4)
~ G G X iG (r ) = ( X iG ,1 ( r ), X i , 2 ( r ), L , X i , D ( r )) S , i = 1,2, L , I ; r = 1,2, L , R.
~
(5)
The conditions of the whole swarm before the first iteration are given as follows. X i (0) S is given at random, i.e., X i ,1 (0), X i , 2 (0), L , X i ,1 (0), i = 1,2, L , I are mutually independent and uniformly distributed on ( a1 , b1 ), ( a2 , b2 ), L (a D , bD ) , respectively. The initial velocities are set to be zero vectors, i.e. Vi (0) = 0 . The initial personal and global best positions are set to be that
~ X iP (0) X i (0)
and
first iteration of first particle ( r = 1, i = 1 ), first iteration of second particle ( r = 1, i = 2 ), L , first iteration of Ith particle ( r = 1, i = I ), second iteration of first particle ( r = 2, i = 1 ),L .In (6)-(9), the action of the r th iteration, r = 1, 2,L , R ,on particle i , i = 1, 2,L , I , is given as follows :
~ ~ ~ ~ ~ ~ ~ ~ Vi (r) = (r)Vi (r 1) + (U (0,c1) (r) ( XiP (r 1) Xi (r 1))) + (U (0,c2 ) (r) ( XiG ) Xi (r 1))), (6) 1 (r 1
where ( r ) is a parameter called the inertia weight and is a function of r ; c1 , c2 are positive constants, referred to as cognitive and social parameters, respectively;
( 0 , ck ) ( 0 , ck ) ( 0 ,c k ) ~ U ( 0,ck ) (r ) = U 1 (r ), U 2 (r ),L , U D (r ), k = 1,2,
After the computation of the new velocity, the new position of the particle i after its r th iteration is computed by
(7)
With the new position in (7) and its new fitness function value f ( X i (r + 1)), the personal best position and global best position are possibly renewed using (8) and (9).
~ ~ X iP ( r 1) and X i ( r ), ~ ~P X iG 1 ( r ) and X i ( r ),
(8) (9)
Here X iG 1 ( r ) denotes the global best after the r th iteration of the particle ( i -1). Note that, to be consistent in (9). As for i =1, we define
~ ~ ~ f ( X 0 ) f ( X ), X S .
Furthermore, let X 1 , X 2 , L , X n be i.i.d. uniform random vector on S in (1) and
~ ~
~ ~ f ( X 1:n ) = f ( X 0 ) in probability.
Theorem 2.1 proposes a convergence behavior of a non-cooperative i.i.d. searching algorithm if the fitness function is continuous. Furthermore, the real distribution can be approximated by empirical distribution for large data set and the estimation of probability percentile is of great importance in practical application. In Lemma 2.1, the minimum of n realized function values can be used to estimate the 1/n-percentile of the distribution of the function value and, with the help of Theorem 2.1, the estimate will converge to the true minimum. This can help to estimate the percentiles for the distribution and can be used in our adaptive parameter selection. Lemma 2.1. (Estimation of Percentile) As in Theorem 1, let Y , X 1 , X 2 , L , X n be i.i.d. uniform random vector and the random functions f ( X 1:n ) f ( X 2:n ) L f ( X n:n ) be the order statistics
~ ~ ~
evaluation of the second particle (i = 2) , and so on. For every t only one particle moves and recompute its fitness function value. The index i of the moving particle and the time t satisfy the following relation:
t (mod I ), i= i = I ,
if t (mod I ) 0 . if t (mod I ) = 0
(10)
~ ~ ~ ~ (t )Vi (t 1) + (U ( 0,c1 ) (t ) ( X iP (t 1) X i (t 1))) ~ ~ ~ ~ Vi (t ) = + (U ( 0,c2 ) (t ) ( X G (t 1) X i (t 1))), for i satisfies (10), (11) 0, other i. ~ ~ ~ X i (t 1) + Vi (t ) for i satisfies (10), X i (t ) = ~ other i. X i (t 1), ~P ~ ~P the better one of ( X i (t 1) , X i (t)), X i (t ) = ~ P X i (t 1) ~G ~ ~ X (t ) = the better one of ( X G (t 1) , X iP (t)), for i satisfies (10), other i.
(12)
(13) (14)
Here X i (t ), i = 1,2, L , I , denote the position (vector) at time t ; X iP (t ), i = 1,2, L , I , the personal best position (vector) at time t ; Vi (t ), for i satisfies (10), the moving particles velocity at time t and X G (t ) denotes the global best position at time t . We use t instead of r in (11) to (14) because the state changes with t increases. Every time only one particle moves according to (11) and (12). Then function evaluation is made to renew the position of the moving particle, the fitness function value is re-computed and the personal best position (of moving particle) and global best position (of whole swarm) are possibly changed ((13) and (14)). Note that every time only one particle is moved and the other particles inherit their previous values of positions and personal best positions ((12) and (13)). Also note that f ( X iP (t )), i = 1,2, L , I , and f ( X G (t )) are all nonincreasing functions of t . Combine the formulas (11)-(14) and define the state at time t , as W (t ),
Theorem 3.1. The states of the PSO, as defined in (15), are memory-less, i.e. the conditional distribution function of the next state W (t + 1) , given the present and historical states, W (t ), W (t 1), L , W (1), is the same as the conditional distribution function of W (t + 1) , given only the present state W (t ) , i.e.
(16)
Proof. Without loss of generality we assume that at present time t only particle i moves and recompute its fitness function, while at future time (t + 1) only particle (i + 1) moves and re-compute its fitness function. We discuss the change of the components from that in W (t ) to that in W (t + 1) in the following. Firstly, the future velocity of particle will be set to be zero, i.e.
~ ~ Vi (t + 1) = 0 since particle will not move at time (t + 1) . The other components of W (t + 1) inherit ~ except only four components: their previous values in W (t ) ~ ~ ~P ~G ~ Vi+1 (t + 1) = 0, X i+1 (t + 1), X i+1 (t + 1) ,and X (t + 1) . Among them Vi +1 (t + 1) = 0 depends only ~ ~ on the present state W (t ) by (11).The other three also depend only on the present state W (t ) by ~ ~ (12), (13) and (14) respectively. Thus W (t + 1) depends only on W (t ) . This completes the proof.
3.3. The Concept of _Dominating
With the memory-less property, the fate of PSOs depends only on the present state while not on the past history. Therefore what is the most important is what the state is now instead of what the state used to be. Therefore we can investigate the influence of the present state on the future state regardless of the historical state. Also, the future states of the PSOs do not depend directly on the time t . This paper will give the result that the state of PSOs constitutes a stationary Markov chain and therefore depend on t at most through the present state. We explain this phenomenon in details in the following. As t increases, the state may gain considerably better achievement, but it is not definite. If for a specific sample process, as t increases while the state may gain no significant achievement, then the future state distribution will not be favored even t is very large. Hence it seems better to consider different parameters portfolio during different state stage instead of time stage as the previous research did. To evaluate the achievement (goodness) of a state, we define an index for the state to represent the present achievement of PSO. This index is based on the concept of probability. Since the achievement is position-dependent, firstly we define an index relevant to the positions. Definition3.1.(
_Dominating, 0 1 , with respect to f and its domain S , if the probability that a uniform distributed random vector is better than Y0 is , i.e. Pr( f (Y1 ) < f (Y0 )) = .
We discuss two extreme cases in the following: (i) When = 0 , Y0 is a best position. (ii) When = 1 , Y0 is a worst position.
With smaller, the position (fitness function) is better achieved, but it is not to be assured that the position is geographically nearer the global optimum. Maybe it is only nearer to one local optimum. In the following, we discuss a concept similar to that in Definition 3.1 with respect to the state space. Definition3.2. ( _Dominating of states) A state W is said to be _Dominating, or with value (in dominating), with respect to f and its domain S, if its global best position is _Dominating. According to Definition 3.2, the immediate result in Lemma 3.1 follows. Lemma 3.1. The value (in dominating) of a state depends only on its global best position. Furthermore, we define one relation between two states in Definition 3.3. Definition 3.3. (Equal_Dominating relation) Two state W1 and W2 are said to be Equal_Dominating to each other, if they are both _Dominating for the same . Note that a state is better if its value is smaller and two states are of little difference in view of achievement if they are Equal_Dominating. Lemma 3.2 gives the result that its an equivalent relation. Lemma 3.2. The relation of Equal_Dominating is an equivalent relation. Proof. Its rather intuitive that the Equal_Dominating relation is reflexive, symmetric and transitive. We omit the proof. According to Lemma 3.2, we can divide the whole state space of PSO into uncountable number of equivalent classes. Each class corresponds to one value. Within one class every pair of the ~ ~ states, W1 and W2 , are Equal_Dominating to each other. Whereas for any pair of states that are not in the same class, they are not Equal_Dominating. But the total number of values, or equivalent classes, is uncountable and thus very difficult for analysis. In the following, the equivalent classes will be merged into countable ones and the analysis tasks will thus be simplified.
(17)
Definition 4.1. Given a countable sequence {n , n = 0,1,L } as in (17), a state W is said to be Level_ n _Dominating, , (or simply in Level n ), denote by L (W ) = n, with respect to f , S ,
With Definition 4.1 the states are merged into countable number of classes, called levels, by taking into account the countable intervals of -value instead of uncountable single -value. Now a similar relation can be defined as that in Definition 4.2.
9
Definition 4.2. (Equal_Level relation) Two states and are said to be Equal_Level to each other, ~ ~ if L (W1 ) = L (W2 ) = n for some n {0,1, 2,L} . Just as in Lemma 3.2, a similar result of equivalent equation for Equal_Level is stated in the following without proof. Lemma 4.1. The relation of Equal_Level is an equivalent relation. Note that if two states are Equal_Dominating, then they must be Equal_Level, but not vice versa. To help the proof in the following theorem, define n , n = 0,1, 2,L , to be the classes of state space for W such that n = {W | L(W ) = n} Give an example by
then
we
have
the
sequence { n , n = 0,1, 2,L} be a partition of the whole state space. In this example a state is in Level-n means that its global best position attains the 10 n -percentile but not yet the 10 ( n +1) percentile. With the aid of Definition 4.2, a Markov chain can be established. Theorem 4.1. In PSO, the stochastic process of the levels of the state {L (W (t )), t = 1,2, L}, or simply denote by {L (t ), t = 1, 2,L} as defined in Definition 4.2, constitutes a Markov chain. Proof. Given the history of the previous levels L (t ) = m, L (t 1) = mt 1 , L (t 2) = mt 2 ,L , L (1) = m1 , m, mt 1 , mt 2 ,L , m1 {0,1,L} , the conditional probability of L (W (t + 1)) = n {0,1,2, L} is
~ ~ ~ ~ ~ Pr[ L(W (t + 1)) = n | L(W (t )) = m, L(W (t 1)) = mt 1 , L (W (t 2)) = mt 2 , L , L (W (1)) = m1 ] ~ ~ ~ ~ ~ = DF (W (t + 1) | L(W (t )) = m, L(W (t 1)) = mt 1 , L(W (t 2)) = mt 2 , L , L(W (1)) = m1 )d
~ W ( t +1) n
~ W ( t +1) n
DF (W (t + 1) | L(W (t )) = m)d
Pr[ L (t + 1) = n | L (t ) = m] pm ,n (t ), m, n, t .
(18)
Observe the last equality in the proof of Theorem 4.1. Since the conditional probability in the formula is independent of t , i.e. the resulting conditional probability for different in (18) are all the same. Hence, we have the following further result. Theorem 4.2. The Markov chain defined in Theorem 4.1 is a stationary (time homogeneous) one, i.e. the formula in (18) is reduced to
10
Pr[ L (t + 1) = n | L (t ) = m] pm,n , m, n, t
(19)
In (19), the state (level) transition probability pm , n is independent of t . This justifies our argument that parameter selection should be a function of state (achievement) instead of a function of t (or iteration). The transition probability matrix can be denoted by
(20)
Note that pm , n = 0, m > n, i.e. P is an upper triangular matrix since the levels will at least stay the same comparing with previous state, if not increase. In (20) pm , n is one step transition probability and P is one step transition probability matrix. Besides the one step probability, l -step probability can also be established for general positive integer . Define the l -step transition probability and l -step transition probability matrix as
l Pr[L(t + l) = n | L(t) = m] pm ,n , l 1, m, n
(21)
and
l l l p0,0 p0,1 K p0, n K l l l p1,0 p1,1 K p1,n K l P l [ pm M L M L . (22) ,n ] M p l p l K p l K m, n m ,0 m ,1 M M L M L l l For the same reason as in (20), pm,n = 0, m > n, i.e. P is an upper triangular matrix. By the
theory of Markov chain, P l can be easily computed by the matrix product of matrix P , i.e. Pl = 1 PP L P . 4 24 3
l
Take a closer look at Theorem 4.2. The PSO faces the common fate of a stationary Markov chain: the iteration number r (or equivalently, t ) may be very large, while the present achievement is poor (for example, just like a new beginner), then the efforts done before will be almost in vain. To be more precise, the future fate of the PSO depends only on the present achievement (state) but not directly on the iteration number or t . The theoretical results convince us that: different parameters may render different transition probability in different levels of achievements rather than in different number of iterations. As a result, unlike most of the methods in other papers, we propose a heuristic idea. When dynamically choosing (tuning) parameters, the present achievement (level) will be taken into consideration instead of the number of iterations as traditional papers did. For example, we can compare the transition probability between different parameter set (transition probability matrix is parameter dependent). During different level of
11
achievements, choose the parameter set which is most favorable and help to promote the levels to higher ones. The further application can be the investigation in the relationship between parameter set and transition probability for different optimizers. One reliable method is by using of computer simulation. For example, given a fixed fitness function, we can estimate the transition probability pn,n+1 in different parameter set and choose the optimum parameter set in each level to maximize the value of pn ,n +1 ( , Vmax , ck ) (a function of , Vmax , ck ) or minimize the value of pn ,n ( , Vmax , ck ) . Here for practical purpose we assume that the probability that PSO jumps more than or equals to two levels at one time is very small with respect to jumps one level, i.e.
pn ,n +l pn ,n +1
0, l 2.
The method to find the percentile of the function value of each fitness function can be done by simulation according to the result of Lemma 1. For the example given before, % = (100 ,101,102 ,L), the percentile of the fitness function value f n , n = 0,1, 2,L , satisfying the relation that Pr[ f ( X ) < f n ] = 10 n , n = 0,1, 2,L , can be estimated by use of the minimum of 10 n i.i.d. random function evaluations.
5. CONCLUSION
In this section we firstly propose some indexes to measure and compare the performance of different PSO algorithms, and then some steps of adaptive parameters selection are given based on the theoretical results in this work.
B3. Best Level_ n _Dominating: the largest n of the state the algorithm can attain by using the same fixed number of function evaluations. C1. Improving Ratio based on errors: to achieve the same accuracy (error), the ratio between the numbers of function evaluations needed by the non-cooperative i.i.d. random selection and that by PSOs, or between two PSOs, it measures the ratio of the times of function evaluations needed by two methods. C2. Improving Ratio based on value: to achieve the same value, the ratio between the numbers of function evaluations needed by the non-cooperative i.i.d. random selection and that by PSOs, or between two PSOs. C3. Improving Ratio based on level_ n : to achieve the same level, the ratio between the numbers of function evaluations needed by the non-cooperative i.i.d. random selection and that by PSOs, or between two PSOs.
% = ( 0 , 1 , 2 ,L , n ) , here for practical purpose we use finite dimensions. 1. Set the values
2. Find by simulation the function values
f1 , f 2 ,L , f n
such
that Pr[ f ( X ) f i ] = i , i = 1, 2,L , n , the respective function values are the threshold of each level. 3. Find by simulation the best portfolio of parameters (such that pm,m+1 is maximum) at different levels (dynamically parameters selection). 4. The best portfolio of parameters at each level found in step 3 constitute the best leveldependent parameter sets, {PS(m), m =1,2,L, n}.
REFERENCES
[1] [2] [3] [4] [5]
P. Angeline, Using selection to improve particle swarm optimization, in Proc. IJCNN99, Washington, DC, July 1999, pp. 8489. F. V. D. Bergh and A. P. Engelbrecht, A Cooperative Approach to Particle Swarm Optimization, IEEE Trans. Evol. Comput., vol. 8, pp. 225239, June 2004. D. W. Boeringer and D. H. Werner, Particle Swarm Optimization Versus Genetic Algorithms for Phased Array Synthesis, IEEE Trans .Ant. Prop., vol. 52, no. 3, Mar. 2004. X. Cai , Z. Cui , J. Zeng , and Y. Tan, Dispersed particle swarm optimization, Information Processing Letters, vol.105, no.6, pp.294-301, MAR. 16 2008. C.W. Chou, H.H. Lin and J.H. Lin, "A note on Particle Swarm Optimization: An integrated and theoretical approach," in 4th IEEE International Conference on Computational Cybernetics, Tallinn, Estonia, AUG. 2006.
[6]
L. D. S. Coelho and B. M. Herrera, Fuzzy Identification Based on a Chaotic Particle Swarm Optimization Approach Applied to a Nonlinear Yo-yo Motion System, IEEE Transactions on Industrial Electronics, vol.54, no.6, pp. 3234-3245, DEC. 2007. 13
[7]
S. Das, A. Abraham and A. Konar, Automatic Clustering Using an Improved Differential Evolution Algorithm, IEEE Transactions on Systems, Man and Cybernetics Part A-Systems and Humans, vol.38, no.1, pp.218-237, JAN. 2008. S. Das and A. Konar, A swarm intelligence approach to the synthesis of two-dimensional IIR filters, Engineering Applications of Artificial Intelligence, vol.20, no.8, pp.1086-1096, DEC. 2007. G. G. Dimopoulos and C. A. Frangopoulos, Optimization of energy systems based on Evolutionary and Social metaphors, Energy, vol.33, no.2, pp.171-179, FEB. 2008.
[8] [9]
[10] R. C. Eberhart and J. Kennedy, A new optimizer using particle swarm theory, in Proc. 6th Int.
Symp. Micro Machine and Human Science, Nagoya, Japan, 1995, pp. 3943.
[11] R. C. Eberhart, P. Simpson, and R. Dobbins, Computational Intelligence PC Tools: Academic, 1996,
ch. 6, pp. 212226.
[12] M. El-Telbany and F. El-Karmi, Short-term forecasting of Jordanian electricity demand using
particle swarm optimization, Electric Power System Research, vol.78, no.3, pp.425-433, MAR. 2008.
[13] H.W. Ge, L.Sun, Y.C. Liang, and F. Qian, An Effective PSO and AIS-Based Hybrid Intelligent
Algorithm for Job-Shop Scheduling, IEEE Transactions on Systems, Man and Cybernetics Part ASystems and Humans, vol.38, no.2, pp.358-368, FEB. 2008. [14] F.A. Guerra and L.D.S. Coelho, Multi-step ahead nonlinear identification of Lorenzs chaotic system using radial basis neural network with learning by clustering and particle swarm optimization, Chaos Solitons & Fractals, vol.35, no.5, pp.967-979, MAR. 2008.
[15] B. Jiao, Z. Lian, and X. Gu, A dynamic inertia weight particle swarm optimization algorithm,
Chaos Solitons & Fractals, vol.37, no.3, pp.698-705, AUG. 2008.
[16] Y. T. Kao and E. Zahara, A hybrid genetic algorithm and particle swarm optimization for
multimodal functions, Applied Soft Computing, vol.8, no.2, pp.849-857, MAR. 2008.
[17] J. Kennedy, Stereotyping: Improving particle swarm performance with cluster analysis, in Proc.
2000 Congr. Evolutionary Computing, 2000, pp. 15071512.
[18] M. Khodier, N. Dib, and J. Ababneh, Design of multi-band multi-section transmission line
transformer using particle swarm optimization, Electrical Engineering, vol.90, no.4, pp.293-300, APR. 2008. [19] Z. J. Lee, An integrated algorithm for gene selection and classification applied to microarray data of ovarian cancer, Artificial Intelligence in Medicine, vol.42, no.1, pp.81-93, JAN. 2008. [20] Y. Liu and X. Gu, Skeleton-Network Reconfiguration Based on Topological Characteristics of Scale-Free Networks and Discrete Particle Swarm Optimization, IEEE Transactions on Power Systems, vol.22, no.3, pp.1267-1274, AUG. 2007.
[22] K. E. Parsopoulos and M. N. Vrahatis, On the Computation of All Global Minimizers Through
Particle Swarm Optimization, IEEE Trans. Evol. Comput., vol. 8, pp. 211224, June 2004.
[23] Y. Shi and R. C. Eberhart, A modified particle swarm optimizer, in Proc. IEEE Int. Conf.
Evolutionary Computation, Anchorage, AK, May 1998.
[24] H. S. Yazdi and S. Effati, Eigenvalue spread criteria in the particle swarm optimization algorithm for
solving of constraint parametric problems, Applied Mathematics and Computation, vol.192, no.1, pp.40-50, SEP. 1 2007. 14
Authors
Chao-Wei Chou received his B. S. degree in Electrical Engineering from National Sun-Yat Sen University, Taiwan in 1988. He then received his M.S. degree in Business Administration from National Taiwan University, Taiwan in 1994. He received his Ph. D. in Statistics from National Sun-Yat Sen University, Taiwan in 2004. He has subsequently joined the Department of Information Management of I-Shou University as an assistant Professor. His research interests include artificial intelligence, computer network, information security and information economics. Jiann-Horng Lin received his B. S. and M. S. both in Computer Science and Information Engineering from Feng-Chia University, Taiwan in 1987 and 1989, respectively. He then received his Ph.D. in Electrical Engineering and Computer Science from Syracuse University, New York in 1999. He is currently an assistant professor at the Department of Information Management at I-Shou University, Taiwan. He is also the department chair from 2004 to 2007. His research interests include artificial intelligence, data mining, chaos and information theory. He is also interested in the area of evolutionary computation and bioinformatics. Dr. Lin is a member of the IEEE, the Association for Computing Machinery and the Institute of Information and Computing Machinery. Rong Jeng received his B. S. degree in Control Engineering from National ChiaoTung University, Taiwan in 1981. He then received his M.S. degree in Electrical Engineering from Memphis State University, Memphis, TN in 1986. He received his Ph. D. in Electrical and Computer Engineering from University of Cincinnati, Cincinnati OH in 1991. He is currently an associate professor at the Department of Information Management at I-Shou University, Taiwan. He was also the department chair from 1991 to 1994. His research interests include Operations Research, Simulation, information theory.
15