Slideshare uses cookies to improve functionality and performance, and to provide you with relevant advertising. If you continue browsing the site, you agree to the use of cookies on this website. See our User Agreement and Privacy Policy.

Slideshare uses cookies to improve functionality and performance, and to provide you with relevant advertising. If you continue browsing the site, you agree to the use of cookies on this website. See our Privacy Policy and User Agreement for details.

Like this presentation? Why not share!

- Particle Swarm Optimization by Stelios Petrakis 22332 views
- Particle Swarm optimization by midhulavijayan 6074 views
- Particle Swarm Optimization: The Al... by adil raja 2863 views
- Particle Swarm Optimization by Rajo... by Rajorshi Mukherjee 2171 views
- Particle swarm optimization by anurag singh 1342 views
- Particle Swarm Optimization (Pso) by rkmohammadi 5226 views

12,953 views

Published on

Practical Swarm Optimization (PSO)

Published in:
Technology

No Downloads

Total views

12,953

On SlideShare

0

From Embeds

0

Number of Embeds

14

Shares

0

Downloads

1,364

Comments

0

Likes

12

No embeds

No notes for slide

- 1. Department of Mechanical and Material Engineering Practical Swarm Optimization (PSO)
- 2. Goal of Optimization Find values of the variables that minimize or maximize the objective function while satisfying the constraints.
- 3. . Need for optimization Choose design variables Formulate constraints Formulate objective function Set up variable bounds Select an optimization algorithm Obtain solution(s) Flowchart of Optimal Design Procedure
- 4. Particle Swarm Optimization Swarm Intelligence (SI) • SI is artificial intelligence, based on the collective behavior of decentralized, self- organized systems. • The expression was introduced by Gerardo Beni and Jing Wang in 1989, in the context of cellular robotic systems. • SI systems are typically made up of a population of simple agents interacting locally with one another and with their environment. • Natural examples of SI include ant colonies, bird flocking, animal herding, bacterial growth, and fish schooling.
- 5. Particle Swarm Optimization Some SI Application • U.S. military is investigating swarm techniques for controlling unmanned vehicles. • NASA is investigating the use of swarm technology for planetary mapping.
- 6. Particle Swarm Optimization • The PSO algorithm was first described in 1995 by James Kennedy and Russell C. Eberhart inspired by social behavior of bird flocking or fish schooling. • PSO is an artificial intelligence (AI) technique that can be used to find approximate solutions to extremely difficult or impossible numeric maximization and minimization problems. • Hypotheses are plotted in this space and seeded with an initial velocity, as well as a communication channel between the particles. • Simple algorithm, easy to implement and few parameters to adjust mainly the velocity.
- 7. Particle Swarm Optimization How it works: • PSO is initialized with a group of random particles (solutions) and then searches for optimal by updating generations. • Particles move through the solution space, and are evaluated according to some fitness criterion after each time step. In every iteration, each particle is updated by following two "best" values. • The first one is the best solution (fitness) it has achieved so far (the fitness value is also stored). This value is called pbest.
- 8. Particle Swarm Optimization How it works: • Another "best" value that is tracked by the particle swarm optimizer is the best value obtained so far by any particle in the population. This second best value is a global best and called gbest. • When a particle takes part of the population as its topological neighbors, the second best value is a local best and is called lbest. Neighborhood bests allow parallel exploration of the search space and reduce the susceptibility of PSO to falling into local minima, but slow down convergence speed.
- 9. Particle Swarm Optimization Each particle tries to modify its current position and velocity according to the distance between its current position and pbest, and the distance between its current position and gbest. 2 1( ) * ( CurrentPosition ) 2( ) * ( CurrentPositionn n1 1 best,n best,n )randv v c rand p c gnn Current Position[n+1] = Current Position [n] + v[n+1] current position[n+1]: position of particle at n+1th iteration current position[n]: position of particle at nth iteration v[n+1]: particle velocity at n+1th iteration vn+1: Velocity of particle at n+1 th iteration Vn : Velocity of particle at nth iteration c1 : acceleration factor related to gbest c2 : acceleration factor related to lbest rand1( ): random number between 0 and 1 rand2( ): random number between 0 and 1 gbest: gbest position of swarm pbest: pbest position of particle
- 10. Particle Swarm Optimization Algorithm For each particle Initialize particle with feasible random number End Do For each particle Calculate the fitness value If the fitness value is better than the best fitness value (pbest) in history Set current value as the new pbest End Choose the particle with the best fitness value of all the particles as the gbest For each particle Calculate particle velocity according to velocity update equation Update particle position according to position update equation End While maximum iterations or minimum error criteria is not attained
- 11. Particle Swarm Optimization Swarm Topology • In PSO, there have been two basic topologies used in the literature I4 I0 I1 I2I3 I4 I0 I1 I2I3 Star Topology (global neighborhood)Ring Topology (neighborhood of 3)
- 12. Particle Swarm Optimization Variant of PSO .
- 13. The Basic Variant of PSO PSO Basic Variant Function Advantages Disadvantages Velocity Clamping (VC) Control the global exploration of the particle. Reduces the size of the step velocity, so that the particles remain in the search area, but it cannot change the search direction of the particle VC reduces the size of the step velocity so it will control the movement of the particle If all the velocity becomes equal to the particle will continue to conduct searches within a hypercube and will probably remain in the optima but will not converge in the local area. Inertia Weight Controls the momentum of the particle by weighing the contribution of the previous velocity, A larger inertia weight in the end of search will foster the convergence ability. Achieve optimality convergence strongly influenced by the inertia weight Constriction Coefficient To ensure the stable convergence of the PSO algorithm [21] Similar with inertia weight when the algorithm converges, the fixed values of the parameters might cause the unnecessary fluctuation of particles Synchronous and Asynchronous Updates Optimization in parallel processing Improved convergence rate Higher throughput: More sophisticated finite element formulations Higher accuracy (mesh densities)
- 14. Particle Swarm Optimization in Summary The process of PSO algorithm in finding optimal values follows the work of an animal society which has no leader. Particle swarm optimization consists of a swarm of particles, where particle represent a potential solution (better condition). Particle will move through a multidimensional search space to find the best position in that space (the best position may possible to the maximum or minimum values).
- 15. Question and Answer

No public clipboards found for this slide

×
### Save the most important slides with Clipping

Clipping is a handy way to collect and organize the most important slides from a presentation. You can keep your great finds in clipboards organized around topics.

Be the first to comment