Professional Documents
Culture Documents
AIM:
Implement particle swarm optimization.
INTRODUCTION:
PSO shares many similarities with evolutionary computation techniques such as Genetic
Algorithms (GA). The system is initialized with a population of random solutions and
searches for optima by updating generations. However, unlike GA, PSO has no
evolution operators such as crossover and mutation. In PSO, the potential solutions,
called particles, fly through the problem space by following the current optimum
particles. The detailed information will be given in following sections.
Compared to GA, the advantages of PSO are that PSO is easy to implement and there
are few parameters to adjust. PSO has been successfully applied in many areas:
function optimization, artificial neural network training, fuzzy system control, and other
areas where GA can be applied.
CODE:
iter_max = 10000
pop_size = 100
dimensions = 2
c1 = 2
c2 = 2
err_crit = 0.00001
def f6(param):
'''Schaffer's F6 function'''
para = param*10
para = param[0:2]
num = (sin(sqrt((para[0] * para[0]) + (para[1] *
para[1]))))*\ (sin(sqrt((para[0] * para[0]) +
(para[1] * para[1])))) - 0.5 denom = (1.0 +
0.001 * ((para[0] * para[0]) + (para[1] *
para[1]))) * \ (1.0 + 0.001 * ((para[0] *
para[0]) + (para[1] * para[1]))) f6 = 0.5 -
(num/denom)
errorf6 = 1 - f6
return f6, errorf6;
i += 1
if err < err_crit:
break
#progress bar. '.' = 10%
if i % (iter_max/10) == 0:
print ('.')
for p in particles:
print ('params: %s, fitness: %s, best: %s' % (p.params, p.fitness,
p.best))
OUTPUT: