A Review of Particle Swarm Optimization (PSO) Algorithms for Optimal Distributed Generation Placement
Musa H., Ibrahim S. B.
Department of Electrical Engineering, Bayero University, Kano, Nigeria
Musa H., Ibrahim S. B.. A Review of Particle Swarm Optimization (PSO) Algorithms for Optimal Distributed Generation Placement. International Journal of Energy and Power Engineering. Vol. 4, No. 4, 2015, pp. 232-239. doi: 10.11648/j.ijepe.20150404.16
Abstract: Particle Swarm Optimization (PSO) has became one of the most popular optimization methods in the domain of Swarm Intelligence. Many PSO algorithms have been proposed for distributed generations (DGs) deployed into grids for quality power delivery and reliability to consumers. These can only be achieved by placing the DG units at optimal locations. This made DG planning problem solution to be of two steps namely, finding the optimal placement bus in the distribution system as well as optimal sizing of the DG. This paper reviews some of the PSO and hybrids of PSO Algorithms formulated for DG placement being one of the meta-heuristic optimization methods that fits stochastic optimization problems. The review has shown that PSO Algorithms are very efficient in handling the DG placement and sizing problems.
Distributed generation placement and sizing is an important issue which requires special attention of both planners and system operators. DG installation at non-optimal places can lead to increase in system losses which imply increase in costs and hence having a negative impact opposite to the desired. The selection of the best location and size in large and complex system is a combinatorial optimization problem . Researchers have employed various methods in addressing the placement problems. It has been observed that among all the methods reviewed so far analytical method is found to be the most accurate and more practical technique for placement. However, obtaining a truly optimal solution has presented a challenge as some computational methods do not yield global solution as many local solutions exists.
Due to this problem, deterministic algorithms such as Dynamic programming, NLP, LP, QP and SQP are considered to be the elegant options . However, meta heusutic algorithms such as GA, PSO, EP, Tabu search (TS), simulated annealing (SA) seems to have shared the same dominance as deterministic methods . This is due to the fact that meta heuristic are derivative free problems unlike the deterministic methods and can be solved without need for convexcity. Apart from that the meta heuristic algorithms are independent of initial solution and can avoid local optima . The techniques are robust and can provide near optimal solution for large and complex systems. The only drawback is the high computational efforts required for good solution. On the other hand, meta heuristic methods have their own drawbacks, such as use of trial and error process during parameter tunning and lack of guarantee of global solution attainment atimes. This is one of the reasons that made researchers to direct a lot of effort towards eliminating such problems by combining more than one algorithms to form a hydrid algorithm. The hybrid algorithms currently utilised are; such as LP – QP, EP – SQP GA – SA and PSO – SQP  e.t.c. These efforts have yielded results with a lot of enhancement over the individual algorithms when utilized alone.
Particle swarm optimization (PSO) algorithm is a population based optimization method that has gained much popularity among researchers since after its introduction. The algorithm mimics birds’ behavior during flight in space. Each of the bird in the aggregation of the birds called swarm is represented as a particle. These particles that form the swarm searches for food based on their own experience and that of the other particles within the same swarm. The PSO have been studied by many researchers and several newer versions have been developed for applications in different real-world problems and are found to be robust and fast in solving nonlinear non-differentiable multi-modal problems.
Many survey papers with applications have been presented in literature reviews regarding these studies but it is still in its infancy stage requiring a lot of research work. Authors in  presented a method for optimal siting and sizing of multiple distributed generators (DGs) using PSO based approach. Similar multiple DGs placement was also presented in  with PSO as an optimization tool for variable power load with non-unity power factor. In another development an improved PSO algorithm was also proposed for optimal placement with an in built mechanism for better search that is capable of escaping local optima in . As part of improvement on initial PSO algorithm, the concept of hybridized PSO was introduced by authors in  in which Genetic Algorithm (GA) was combined with PSO. The GA is made to search for the DG site while the PSO optimizes the DG sizes which resulted in drastic reduction in system losses and improvement in voltage profile.
In this paper an introduction of PSO concepts and its algorithm is presented and survey of existing work follows based on objectives, methods and contributions of the work towards finding of optimal solutions to placement and sizing problems.
2. DG Placement and Sizing
The DG placement optimization problem has not been assessed thoroughly as done in many optimization problems. The review in this paper differs from previous reviews in the sense that all work done is going to be categorized based on optimization algorithms employed. Figure 1.0 shows the published research work done on placement and sizing based on IEEE Explore Digital library data base. The analysis shows significant improvement in papers published.
The gradual increase in published papers is a clear indication of growing interest of researchers willing to find solution to DG placement problems.
Many researchers have used various methods to tackle the placement and sizing problems. Methods of optimal placements and sizing of the DGs within networks always depend on the objectives and solution techniques employed. There are three basic models for DG optimization problems that are currently in use which are:
i) Objective function model
ii) Constraints model
iii) Optimization algorithms model
Objective functions are optimized subject to operating constraints by using different techniques. The objective function can be single or multiple for the purpose of achieving maximum benefits of the DG without violating the equality and inequality constraints of the entire power system. The most popular objective is the minimization of power losses . Among other objectives considered by many researchers includes minimization of real and reactive power loss, maximization of DG capacity, maximization of profit and social welfare to mention a few [11-14]. Other objectives handled by many authors are the technical issues; environmental issues and voltage limit [15-16].
The constraints are state variable limits that are placed on operating conditions. Constraints are basically categorized into equality and inequality which must be satisfied by the objective functions either single or multi-objectives. The common constraints generally in DG placement includes but not limited to the following; line thermal limit, short circuit ratio limit, voltage step limit, phase angle limit, power generation limit, DG power generation limit, number of DG limit, power flow equality constraint, voltage profile limits, short circuit level limit, total line loss limit, substation transformer capacity limit, tap position limit, and power factor limit [12,14,17,23].
Many researchers proposed different methods such as analytic as well as deterministic and heuristic methods to solve placements problems. Out of these methods metaheuristic is the only method that has proven their effectiveness in solving optimization problems with appreciable feasible search space . They can also be modified easily to become a hybrid of more than one algorithm to cope with different elements commonly used in most studies.
3. The Concept of PSO Algorithm
The search process is similar to the social behaviour of flying birds when searching for food. The individual bird called particles or swarm flies in the optimization problem hyperspace to search for optimal food location. The position and velocity of the particles is always changing and adjusted according to the cooperative communication among the particles and each individual’s own experience simultaneously. Therefore the particle changes position by balancing its social and individual experience. Each particle is assigned a velocity as well as position vector
For a swarm of m-particle in hyperspace, the position and velocity vectors are ;
where i is the particle index, is the swarm velocity and is the optimization problem dimension. The particle’s new position is;
is particle new position at iteration
is particle old position at iteration
is particle new velocity at iteration
Equation (3) is the updated position equation. The updated velocity vector for particle is
is the previous velocity of particle
is the inertia weight
c1, c2 are the individual and social acceleration positive constants.
r1, r2 are the random values in the range [0, 1], sampled from a uniform distribution
is the personal best position associated with particle own experience
is the global best position associated with the whole neighborhood experience
3.1. Updated Velocity
The updated velocity as given by equation (4) has three major components consisting of the following;
1. The first component is related to particle’s immediate previous velocity, and it consists of two variables. The variables are particle last velocity and inertia weight .
2. The second component is the cognitive component, which shows the individual’s own experience.
3. The last term is the third components that represent the intelligent exchange of information between particleand the swarm.
In the absence of the second and third terms of the velocity formula the particle will continue to fly in the same direction with a speed proportional to its inertia weight until it hits one of the solution space boundaries.
Therefore solution can never be obtained unless the solution lies in the same path of the previous velocity in such a case. The change in direction towards the solution is achievable with the help of second and third term of the equation. Those three components of the velocity update are responsible for the optimization process.
Versions of PSO algorithms have been proposed since after Kennedy and Eberhart which are the local best PSO and the global best PSO. The two models differ in the social component of the velocity update formula. In the case of the local best PSO the swarm is divided into several neighborhoods and the of particle is its neighborhood’s global value. Whereas the global best model considers the swarm as one entity, and therefore the PSO particle’s is the best value for the whole swarm. Generally, the global model is more popular version since it needs less work to achieve result.
3.2. Previous Velocity Components
The component is given by the product of previous velocity of particle and the inertia weight . This term connects the particle in the existing iteration with immediate past iteration which serves as the particle’s memory. It is important as it prevented the particle from sudden change in its direction, and also allows the particle’s own knowledge of its previous flight information to influence its newer course.
The first version of PSO has no inertia weight as was assumed to be unity. It was in subsequent versions that inertia weight was introduced in order to control the contribution of the particle’s previous velocity in the current velocity decision making and this lead to a significant improvement in the PSO algorithm . The implication of making its value too large is the broadening of the exploration mission of the particle, and if the value is small the exploration will be localized. Several dynamic inertia weights were proposed in literatures. The formulations of the literatures have expressed inertia weight as follows;
Where w(k) is the inertia weight value at iteration k
nk is maximum number of iterations
w(nk) is the inertia weight value at the last iteration nk
Some authors in  have proposed using 0.9 and 0.4 as the initial and final weight values respectively.
Another factor introduced that is similar to inertia weight function is the constriction factor (l) which is used for the balancing of the search mechanism between global and local exploration. Use of this factor improves convergence and the particles velocity is therefore constricted by a factor l as expressed in the following equation;
Hence, the constriction factor is a factor of individual and social acceleration positive constants and respectively. This factor is normally considered as a special case of inertia weight PSO algorithm because of the constraints imposed above. The factor l controls the particle’s velocity vector, whiles the inertia weightcontrols the contribution of the particle’s previous velocity towards calculating the new one velocity. Use of constriction factor eliminates velocity clamping and can safe guards the algorithm against explosion .
3.3. Cognitive Component
The cognitive component of the velocity update equation uses which is referred to as the particle’s best personal position that it has visited so far since the beginning of the PSO iterative process. Each particle in the swarm tries to evaluate its own performance by comparing its own fitness in the current PSO iteration with that evaluated in the proceeding one. The given that its is the best personal position so far, is defined as;
Based on equation (5) each particle is suppose to remember its optimal personal best position achieved for use in the update of velocity for future iteration. This component of velocity update equation diversifies searching process at the same time helps in avoiding possible stagnation.
3.4. Social Component
This represents the social behavior of the PSO particles. The term in this component is referred to as the best position achieved among all the swarm particles. Whenever the best solution among the whole population of the swarm is achieved, all the particles are informed. The fitness value is the optimal among all the particles during the current PSO iteration as;
where is particlefitness value at iteration, and is the swarm size.
3.5. Cognitive and Social Parameters
These parameters and are cognitive and social parameters respectively. They are factors that scaled the and in the updated velocity equation. The trust of the particle in itself is measured by , while reflects the confidence it has on its neighbours. If is 0, the particle’s own experience is eliminated during search process for a new solution, while if is 0 the search is localized and exchange of information between the particles is eliminate. The highest value recommended for the two in most literatures is 2.
and are two random numbers in the range of [0, 1] that are sampled from a uniform distribution. The stochastic exploration nature of PSO is due to these random numbers. A typical illustration for velocity and position update for a single PSO particle during iteration is shown in Fig. 2.0 during iteration.
3.6. Pretty Features of PSO
The advantages associated with PSO are many just like other optimization algorithm. The main advantages are clearly distinct when compared to deterministic methods that are gradient based techniques and have no flexibility of dealing with objective functions that are not continuous or differentiable naturally. The search process for PSO does not involve use of derivative function; instead it uses the fitness function value as a guide for finding optimal solution in problem space. This concept of fitness function employed in PSO helps in eliminating the approximations and assumptions usually adapted on objective functions and constraints as in conventional optimization methods. For this reason PSO is considered as a stochastic optimization method and found to be very efficient in handling problems that their objective functions are time varying or stochastic in nature. Above all the solutions from PSO are not dependant on the initial solutions unlike the deterministic method.
4. Studies on PSO Algorithms for DG Placements and Sizing
This section deals with all the studies done on PSO and hybridized PSO Algorithms for DG placements and sizing. All the literature surveyed are summarized in table 4.1
|1.||Haruna M. ||Minimization of power losses and voltage stability index (VSI)||PSO||Optimal DG Placement model for better search and improved (VSI)|
|2.||El-Zonkoly ||Multi-objective for short circuit level and other technical parameters||PSO||Determination of voltage collapse point for variable load|
|3||Nabavi, S.M.H||Reduce network congestion and minimize locational marginal price (LMP)||PSO||congestion management and social welfare in placement/sizing|
|4||Amanifar, O. ||Minimization of investment cost of DGs and power losses||A PSO algorithm integrated in harmonic power flow algorithm||Investiment cost justifies DG placement for increase in power transfer capability|
|5||Dias, B.H ||Minimization of power losses||PSO in Nonlinear Optimal Power Flow (OPF)||Reduction in search space and improvement convergence|
|6||Gomez-Gonzadez ||Minimization of cost||PSO in Optimal power flow||ODGP model using discrete PSO and OPF|
|7||Wong, L.Y. ||Minimization of power losses||Pso in Newton –Raphson power flow||Effective solutions and improved voltage profile|
|8||Nguyen Cong Hien ||Maximizes reactive power flow||PSO||Enhancement of loadability of the primary distribution feeder|
The PSO algorithms allow the system planner to find not only a single optimum point, but a family of near-optimum planning alternatives. This feature of PSO has become very useful in DG allocation because distribution network operators usually have little or no control on the DG integration and different planning alternatives can be necessary to face uncertainties and minimize risks.
Numerous publications and wide spread implementation of PSO algorithms has been conducted, a lot of barriers to proper implementation of these researchers by network operators is still lacking. Although the algorithm has many advantages over other meta-heuristic methods, the main challenge associated PSO is that of lack of solid mathematical background like many heuristic methods. It’s solution method is problem dependent and for every solution parameters have to be tuned and adjusted for better solution. The conventional PSO introduced by Kennedy and Eberhart in 1995 and even those that had under gone modifications, are still dependent on a number of parameters that are externally set or arbitrary selected by the user . This setting or selection by the user is a very delicate operation involving a lot of trials and errors before a reasonable tuning can be achieved especially in practical problems that require defining of inertia weight at each iteration step. Apart from that the usual initial assumption that inertia term is eliminated at an early stage of the optimization process, can cause the algorithm to be trapped at some local minimum. As a solution to this problem some authors have proposed some procedures of "re-seeding" the search by generating new particles at distinct places of the search space . Other shortcoming of PSO is the random operation involved that a particle is initialized randomly and its location in the search space is updated during each iteration in the PSO algorithm. The problem is that when the particle initialization is not well done the issue of local minimum trapping can also arise or convergence can be prolonged as sharing of information between particles is not properly coordinated.
These challenges made researchers to direct a lot of effort towards eliminating such problems by combining more than one algorithms to form a hybrid algorithms. The efforts have yielded results with a lot of enhancement over the individual algorithms when utilized alone as indicated in table 4.2.
|1||Moradi ||Multi-objective with weights||Hybrid (GA &PSO)||Optimal DG Planning for placement using GA and sizing using PSO|
|2.||Sedighizadeh, M ||Minimization of power losses and voltage profile improvement||Hybrid (PSO and Clonal Selection )||Better quality of solutions and less number of iterations|
|3.||Musa, H. ||Multi-objective for power loss reduction and voltage stability index improvement||Hybrid (PSO & Evolutionary Programming)||Well-distributed Pareto optimal non-dominated solutions of DG sizes obtained|
|4.||Afzalan, M ||Minimization of power losses||Hybrid ( PSO & HBMO)||Voltage profile improvement and branch current reduction|
|5.||Ziari, I ||Minimizes loss and improves system reliability.||Hybrid (Discrete PSO & GA)||Increase in the diversity optimization variables in DPSO not to be trapped in a local minimum|
|6||Soroudi, A. ||Multi-objective for technical constraint dissatisfaction, costs and environmental emissions||Hybrid (Binary PSO-based & Fuzzy)||Better timing of investment for both distributed generation (DG) units and network components obtained|
|7||Wen S. T.||Multi-objective index-based approach for total real power losses, voltage profile, MVA intake by the grid, number of DG and greenhouse gases emission||Hybrid (PSO & Gravitational Search Algorithm )||Algorithm can provide efficient and robust solution to mixed integer nonlinear optimization problem|
|8||Haruna Musa||power loss reduction (PLR) value||Hybrid (PSO & Ranked Evolutionary programming)||A simple and effective algorithm for power loss reduction and voltage profile improvement|
Although there have been numerous publications in the area of the siting and sizing of DG, it is evident that, widespread implementation of the PSO optimizations has not been much. Development of the DG integration strategies and PSO optimization methods requires proper distribution system planning especially with the proliferation of electric vehicle in existing distribution networks. Further challenges that are yet to be tackled for better optimization are the modelling of the distribution networks with all the necessary details needed by network operators.
The authors acknowledged with gratitude the financial support offered by Bayero University Kano Nigeria and the provision of suitable research facilities.