Sunteți pe pagina 1din 15

Particle Swarm Optimization

A new evolutionary computation technique, called particle


swarm optimization (PSO), has been proposed and
introduced recently. This technique combines social
psychology principles in socio-cognition human agents
and evolutionary computations. PSO has been motivated
by the behavior of organisms such as fish schooling and
bird flocking. Generally, PSO is characterized as simple in
concept, easy to implement, and computationally efficient.
Unlike the other heuristic techniques, PSO has a flexible
and well-balanced mechanism to enhance the global and
local exploration abilities.
Like evolutionary algorithms, PSO technique conducts
search using a population of particles, corresponding
to individuals.
Each particle represents a candidate solution to the

problem at hand.
In a PSO system, particles change their positions by
flying around in a multi-dimensional search space until
a relatively unchanging position has been encountered,
or until computational limitations are exceeded.
PSO system combines a social-only model and a
cognition-only model. The social-only component
suggests that individuals ignore their own experience
and adjust their behavior according to the successful
beliefs of individuals in the neighborhood. On the
other hand, the cognition-only component treats
individuals as isolated beings. A particle changes its
position using these models.

The advantages of PSO over other traditional optimization


techniques can be summarized as follows: -

(a)

PSO is a population-based search algorithm i.e.,


PSO has implicit parallelism. This property ensures

PSO to be less susceptible to getting trapped on local


minima.
(b)

PSO uses performance index or objective function


information to guide the search in the problem
space. Therefore, PSO can easily deal with nondifferentiable objective functions. Additionally, this
property

relieves

PSO

of

assumptions

and

approximations, which are often required by


traditional optimization methods.
(c)

PSO

uses

probabilistic

transition

rules,

not

deterministic rules. Hence, PSO is a kind of


stochastic optimization algorithm that can search a
complicated and uncertain area. This makes PSO
more flexible and robust than conventional methods.
(d)

Unlike GA and other heuristic algorithms, PSO has


the flexibility to control the balance between the
global and local exploration of the search space.
This unique feature of PSO overcomes the
premature convergence problem and enhances the

search capability.
(e)

Unlike the traditional methods, the solution quality


of the proposed approach does not rely on the initial
population. Starting anywhere in the search space,
the algorithm ensures the convergence to the optimal
solution.

PSO Algorithm

The basic elements of PSO technique are briefly stated


and defined as follows: Particle, X(t),: It is a candidate solution represented by
an m-dimensional real-valued vector, where m is
the number of optimized parameters. At time t, the
jth particle Xj(t) can be described as Xj(t)=[xj,1(t),
xj,2(t), , xj,m(t)], where xs are the optimized
parameters and xj,k(t) is the position of the jth

particle with respect to the kth dimension, i.e., the


value of the kth optimized parameter in the jth
candidate solution.
Population, pop(t),: It is a set of n particles at time t,
i.e., pop(t)=[X1(t), X2(t), , Xn(t)]T.
Swarm: it is a population of moving particles that tend
to cluster together while each particle seems to be
moving in a random direction.
Particle velocity, V(t),: It is the velocity of the moving
particles represented by an m-dimensional realvalued vector. At time t, the jth particle velocity
Vj(t) can be described as Vj(t)=[vj,1(t), vj,2(t), ,
vj,m(t)], where vj,k(t) is the velocity component of
the jth particle with respect to the kth dimension.
Inertia weight, w(t),: It is a control parameter that is
used to control the impact of the previous velocities
on the current velocity. Hence, it influences the
trade-off between the global and local exploration

abilities of the particles. For initial stages of the


search process, large inertia weight to enhance the
global exploration is recommended while, for last
stages, the inertia weight is reduced for better local
exploration. The decrement function for decreasing
the inertia weight given as w(t)= w(t-1), is a
decrement constant smaller than but close to 1, is
proposed in this study.
Individual best, X*(t),: As a particle moves through the
search space, it compares its fitness value at the
current position to the best fitness value it has ever
attained at any time up to the current time. The best
position that is associated with the best fitness
encountered so far is called the individual best,
X*(t). For each particle in the swarm, X*(t) can be
determined and updated during the search. In a
minimization problem with objective function J,
the individual best of the jth particle Xj*(t) is

determined such that J(Xj*(t)) J(Xj()), t. For


simplicity, assume that Jj*=J(Xj*(t)). For the jth
particle, individual best can be expressed as
Xj*(t)=[xj*,1(t), xj*,2(t), , xj*,m(t)].
Global best, X**(t),: It is the best position among all
individual best positions achieved so far. Hence,
the global best can be determined such that
J(X**(t)) J(Xj*(t)), j=1, 2, , n. For simplicity,
assume that J**=J(X**(t)).
Stopping criteria: These are the conditions under
which the search process will terminate. In this
study, the search will terminate if one of the
following criteria is satisfied: (a) the number of
iterations since the last change of the best solution
is greater than a prespecified number; or (b) the
number of iterations reaches the maximum
allowable number.

The particle velocity in the kth dimension is limited by


some maximum value, vkmax. This limit enhances the local
exploration of the problem space and it realistically
simulates the incremental changes of human learning. The
maximum velocity in the kth dimension is characterized by
the range of the kth optimized parameter and given by

v kmax ( x kmax x kmin ) / N

(1)

where N is a chosen number of intervals in the kth


dimension.

In PSO algorithm, the population has n particles that


represent candidate solutions. Each particle is an mdimensional real-valued vector, where m is the number of
optimized

parameters.

Therefore,

each

optimized

parameter represents a dimension of the problem space.


The PSO technique can be described in the following

steps.

Step 1 (Initialization): Set the time counter t=0 and


generate randomly n particles, {Xj(0), j=1, 2, , n},
where Xj(0)=[xj,1(0), xj,2(0), , xj,m(0)]. xj,k(0) is
generated by randomly selecting a value with
uniform probability over the kth optimized parameter
search space [xkmin , xkmax]. Similarly, generate
randomly initial velocities of all particles, {Vj(0),
j=1, 2, , n}, where Vj(0)=[vj,1(0), vj,2(0), ,
vj,m(0)]. vj,k(0) is generated by randomly selecting a
value with uniform probability over the kth
dimension [-vkmax , vkmax]. Each particle in the initial
population is evaluated using the objective function,
J. For each particle, set Xj*(0)=Xj(0) and Jj*=Jj, j=1,
2, , n. Search for the best value of the objective
function Jbest. Set the particle associated with Jbest as
the global best, X**(0), with an objective function of
J**. Set the initial value of the inertia weight w(0).

Step 2 (Time updating): Update the time counter t=t+1.

Step 3 (Weight updating): Update the inertia weight


w(t)= w(t-1).

Step 4 (Velocity updating): Using the global best and


individual best, the jth particle velocity in the kth
dimension is updated according to the following
equation:
v j , k (t ) w(t ) v j , k (t 1) c1r1 ( x*j , k (t 1) x j , k (t 1))
c2 r2 ( x*j *, k (t 1) x j , k (t 1))

(2)

where c1 and c2 are positive constants and r1 and r2


are uniformly distributed random numbers in [0,1].
Check the velocity limits. If the velocity violated its
limit set it at its proper limit. It is worth mentioning

that the second term represents the cognitive part of


PSO where the particle changes its velocity based on
its own thinking and memory. The third term
represents the social part of PSO where the particle
changes

its

velocity

based

on

the

social-

psychological adaptation of knowledge.

Step 5 (Position updating): Based on the updated


velocities, each particle changes its position
according to the following equation

x j , k (t ) v j , k (t ) x j , k (t 1)

(3)

Step 6 (Individual best updating): Each particle is


evaluated according to the updated position. If Jj<Jj*,
j=1, 2, , n, then update individual best as Xj*(t)=
Xj(t) and Jj*=Jj and go to step 7; else go to step 7.

Step 7 (Global best updating): Search for the minimum


value Jmin among Jj*, where min is the index of the
particle with minimum objective function value, i.e.,
min{j; j=1, 2, , n}. If Jmin<J** then update global
best as X**= Xmin(t) and J**=Jmin and go to step 8; else
go to step 8.

Step 8 (Stopping criteria): If one of the stopping criteria


is satisfied then stop, else go to step 2.

Initialization: population, velocity,


weight, iteration counter
Objective function evaluation
Search for the best solution
Set global best = best solution

For each particle, set local best = particle current position

Initialization
Loop
Update the particle velocity
Update the particle position
Objective function evaluation
Update local best for each particle
Update the global best

Global best better


than best solution?
Update the weight

Yes

Update the best solution


No

Iteration = iteration + 1
Stopping
Criteria met?
No
Yes
Stop

References
[1] J. Kennedy, "The Particle Swarm: Social Adaptation
of Knowledge," Proceedings of the 1997 IEEE
international

Conference

on

Evolutionary

Computation ICEC'97, Indianapolis, Indiana, USA,


1997, pp. 303-308.
[2] P. Angeline, "Evolutionary Optimization versus
Particle

Swarm

Optimization:

Philosophy

and

Performance Differences, " Proc. of the 7th Annual


Conf. on Evolutionary Programming, 1998, pp. 601610.
[3] Y. Shi and R. Eberhart, "Parameter Selection in
Particle Swarm Optimization," Proceedings of the 7th
Annual Conference on Evolutionary Programming,
March 1998, pp. 591-600.
[4] E. Ozcan and C. Mohan, "Analysis of a Simple
Particle Swarm Optimization System," Intelligent
Engineering Systems Through Artificial Neural
Networks, Vol. 8, 1998, pp. 253-258.

[5] M. A. Abido, Optimal Design of Power System


Stabilizers Using Particle Swarm Optimization,
IEEE Trans. on Energy Conversion, Vol. 17, No. 3,
September 2002, pp. 406-413.
[6] M. A. Abido, Optimal Power Flow Using Particle
Swarm Optimization International Journal of
Electrical Power and Energy Systems, Vol. 24, No. 7,
October 2002, pp. 563-571.

S-ar putea să vă placă și