Swarm intelligence/Algorithms

From Wikiversity
Jump to navigation Jump to search

Learning Tasks[edit]

List of Algorithms[edit]

Altruism algorithm[edit]

Researchers in Switzerland have developed an algorithm based on Hamilton's rule of kin selection. This algorithm shows how altruism in a swarm of entities can, over time, evolve and result in more effective swarm behaviour.[2][3]

Ant colony optimization[edit]

Ant colony optimization (ACO), introduced by Dorigo in his doctoral dissertation, is a class of optimization algorithms modeled on the actions of an ant colony. ACO is a probabilistic technique useful in problems that deal with finding better paths through graphs. Artificial 'ants'—simulation agents—locate optimal solutions by moving through a parameter space representing all possible solutions. Natural ants lay down pheromones directing each other to resources while exploring their environment. The simulated 'ants' similarly record their positions and the quality of their solutions, so that in later simulation iterations more ants locate better solutions.[4]

Artificial bee colony algorithm[edit]

Artificial bee colony algorithm (ABC) is a meta-heuristic algorithm introduced by Karaboga in 2005,[5] and simulates the foraging behaviour of honey bees. The ABC algorithm has three phases: employed bee, onlooker bee and scout bee. In the employed bee and the onlooker bee phases, bees exploit the sources by local searches in the neighbourhood of the solutions selected based on deterministic selection in the employed bee phase and the probabilistic selection in the onlooker bee phase. In the scout bee phase which is an analogy of abandoning exhausted food sources in the foraging process, solutions that are not beneficial anymore for search progress are abandoned, and new solutions are inserted instead of them to explore new regions in the search space. The algorithm has a well-balanced exploration and exploitation ability.

Artificial immune systems[edit]

Artificial immune systems (AIS) concerns the usage of abstract structure and function of the immune system to computational systems, and investigating the application of these systems towards solving computational problems from mathematics, engineering, and information technology. AIS is a sub-field of Biologically inspired computing, and natural computation, with interests in Machine Learning and belonging to the broader field of Artificial Intelligence.

Bat algorithm[edit]

Bat algorithm (BA) is a swarm-intelligence-based algorithm, inspired by the echolocation behavior of microbats. BA uses a frequency-tuning and automatic balance of exploration and exploitation by controlling loudness and pulse emission rates.[6]

Gravitational search algorithm[edit]

Gravitational search algorithm (GSA) based on the law of gravity and the notion of mass interactions. The GSA algorithm uses the theory of Newtonian physics and its searcher agents are the collection of masses. In GSA, there is an isolated system of masses. Using the gravitational force, every mass in the system can see the situation of other masses. The gravitational force is therefore a way of transferring information between different masses (Rashedi, Nezamabadi-pour and Saryazdi 2009).[7] In GSA, agents are considered as objects and their performance is measured by their masses. All these objects attract each other by a gravity force, and this force causes a movement of all objects globally towards the objects with heavier masses. The heavy masses correspond to good solutions of the problem. The position of the agent corresponds to a solution of the problem, and its mass is determined using a fitness function. By lapse of time, masses are attracted by the heaviest mass. We hope that this mass would present an optimum solution in the search space. The GSA could be considered as an isolated system of masses. It is like a small artificial world of masses obeying the Newtonian laws of gravitation and motion (Rashedi, Nezamabadi-pour and Saryazdi 2009). A multi-objective variant of GSA, called Non-dominated Sorting Gravitational Search Algorithm (NSGSA), was proposed by Nobahari and Nikusokhan in 2011.[8]

Glowworm swarm optimization[edit]

Glowworm swarm optimization (GSO), introduced by Krishnanand and Ghose in 2005 for simultaneous computation of multiple optima of multimodal functions.[9][10][11][12] The algorithm shares a few features with some better known algorithms, such as ant colony optimization and particle swarm optimization, but with several significant differences. The agents in GSO are thought of as glowworms that carry a luminescence quantity called luciferin along with them. The glowworms encode the fitness of their current locations, evaluated using the objective function, into a luciferin value that they broadcast to their neighbors. The glowworm identifies its neighbors and computes its movements by exploiting an adaptive neighborhood, which is bounded above by its sensor range. Each glowworm selects, using a probabilistic mechanism, a neighbor that has a luciferin value higher than its own and moves toward it. These movements—based only on local information and selective neighbor interactions—enable the swarm of glowworms to partition into disjoint subgroups that converge on multiple optima of a given multimodal function.

Intelligent water drops[edit]

Intelligent water drops algorithm or the IWD algorithm is a nature-inspired swarm-based optimization algorithm which was first introduced in 2007. The algorithm attempts to mimic the behavior of natural water drops in rivers. Here, the soil is the quantity that is carried by each artificial water drop in the algorithm. Several versions of the algorithm have been suggested for different applications.[13]

Particle swarm optimization[edit]

Particle swarm optimization (PSO) is a global optimization algorithm for dealing with problems in which a best solution can be represented as a point or surface in an n-dimensional space. Hypotheses are plotted in this space and seeded with an initial velocity, as well as a communication channel between the particles.[14][15] Particles then move through the solution space, and are evaluated according to some fitness criterion after each timestep. Over time, particles are accelerated towards those particles within their communication grouping which have better fitness values. The main advantage of such an approach over other global minimization strategies such as simulated annealing is that the large number of members that make up the particle swarm make the technique impressively resilient to the problem of local minima.

River Formation Dynamics[edit]

River Formation Dynamics (RFD) is based on imitating how water forms rivers by eroding the ground and depositing sediments. After drops transform the landscape by increasing/decreasing the altitude of places, solutions are given in the form of paths of decreasing altitudes. Decreasing gradients are constructed, and these gradients are followed by subsequent drops to compose new gradients and reinforce the best ones. This heuristic optimization method was first presented in 2007 by Rabanal et al.[16] The applicability of RFD to other NP-complete problems was studied in.[17][18][19]

Self-propelled particles[edit]

Self-propelled particles (SPP), also referred to as the Vicsek model, was introduced in 1995 by Vicsek et al.[20] as a special case of the boids model introduced in 1986 by Reynolds.[21] A swarm is modelled in SPP by a collection of particles that move with a constant speed but respond to a random perturbation by adopting at each time increment the average direction of motion of the other particles in their local neighbourhood.[22] SPP models predict that swarming animals share certain properties at the group level, regardless of the type of animals in the swarm.[23] Swarming systems give rise to emergent behaviours which occur at many different scales, some of which are turning out to be both universal and robust. It has become a challenge in theoretical physics to find minimal statistical models that capture these behaviours.[24][25][26]

Stochastic diffusion search[edit]

Stochastic diffusion search (SDS)[27][28] is an agent-based probabilistic global search and optimization technique best suited to problems where the objective function can be decomposed into multiple independent partial-functions. Each agent maintains a hypothesis which is iteratively tested by evaluating a randomly selected partial objective function parameterised by the agent's current hypothesis. In the standard version of SDS such partial function evaluations are binary, resulting in each agent becoming active or inactive. Information on hypotheses is diffused across the population via inter-agent communication. Unlike the stigmergic communication used in ACO, in SDS agents communicate hypotheses via a one-to-one communication strategy analogous to the tandem running procedure observed in Leptothorax acervorum.[29] A positive feedback mechanism ensures that, over time, a population of agents stabilise around the global-best solution. SDS is both an efficient and robust global search and optimisation algorithm, which has been extensively mathematically described.[30][31][32] Recent work has involved merging the global search properties of SDS with other swarm intelligence algorithms.[33][34]

Multi-swarm optimization[edit]

Multi-swarm optimization is a variant of particle swarm optimization (PSO) based on the use of multiple sub-swarms instead of one (standard) swarm. The general approach in multi-swarm optimization is that each sub-swarm focuses on a specific region while a specific diversification method decides where and when to launch the sub-swarms. The multi-swarm framework is especially fitted for the optimization on multi-modal problems, where multiple (local) optima exist.

References[edit]

  1. Machine Learning for Engineers - on GitHub - by trenkert forked from Nam Vu (accessed 2017/09/05) - https://github.com/trenkert/machine-learning-for-software-engineers
  2. Altruism helps swarming robots fly better genevalunch.com, 4 May 2011.
  3. Waibel, M; Floreano, D; Keller, L (2011). "A quantitative test of Hamilton's rule for the evolution of altruism". PLoS Biology 9 (5): e1000615. doi:10.1371/journal.pbio.1000615. http://www.plosbiology.org/article/fetchObjectAttachment.action?uri=info%3Adoi%2F10.1371%2Fjournal.pbio.1000615&representation=PDF. 
  4. Ant Colony Optimization by Marco Dorigo and Thomas Stützle, MIT Press, 2004. ISBN 0-262-04219-3
  5. Karaboga, Dervis (2010). "Artificial bee colony algorithm". Scholarpedia 5 (3): 6915. http://www.scholarpedia.org/article/Artificial_bee_colony_algorithm. 
  6. X. S. Yang, A New Metaheuristic Bat-Inspired Algorithm, in: Nature Inspired Cooperative Strategies for Optimization (NISCO 2010) (Eds. J. R. Gonzalez et al.), Studies in Computational Intelligence, Springer Berlin, 284, Springer, 65-74 (2010).
  7. Rashedi, E.; Nezamabadi-pour, H.; Saryazdi, S. (2009). "GSA: a gravitational search algorithm". Information Science 179 (13): 2232–2248. 
  8. Nobahari, H.; Nikusokhan, M.. "Non-dominated Sorting Gravitational Search Algorithm". International Conference on Swarm Intelligence. 
  9. Krishnanand K.N. and D. Ghose (2005) "Detection of multiple source locations using a glowworm metaphor with applications to collective robotics". IEEE Swarm Intelligence Symposium, Pasadena, California, USA, pp. 84–91.
  10. Krishnanand, K.N.; Ghose, D. (2009). "Glowworm swarm optimization for simultaneous capture of multiple local optima of multimodal functions". Swarm Intelligence 3 (2): 87–124. 
  11. Krishnanand, K.N.; Ghose, D. (2008). "Theoretical foundations for rendezvous of glowworm-inspired agent swarms at multiple locations". Robotics and Autonomous Systems 56 (7): 549–569. 
  12. Krishnanand, K.N.; Ghose, D. (2006). "Glowworm swarm based optimization algorithm for multimodal functions with collective robotics applications". Multi-agent and Grid Systems 2 (3): 209–222. 
  13. Shah-Hosseini, H. (2009). "The intelligent water drops algorithm: a nature-inspired swarm-based optimization algorithm". International Journal of Bio-inspired Computation (IJBIC) 1 (1/2): 71–79. doi:10.1504/IJBIC.2009.022775. 
  14. Parsopoulos, K. E.; Vrahatis, M. N. (2002). "Recent Approaches to Global Optimization Problems Through Particle Swarm Optimization". Natural Computing 1 (2-3): 235–306. doi:10.1023/A:1016568309421. 
  15. Particle Swarm Optimization by Maurice Clerc, ISTE, ISBN 1-905209-04-5, 2006.
  16. P. Rabanal, I. Rodríguez, and F. Rubio (2007) "Using River Formation Dynamics to Design Heuristic Algorithms". Unconventional Computation, Springer, LNCS 4616, pp. 163–177.
  17. P. Rabanal, I. Rodríguez, and F. Rubio (2008) "Finding minimum spanning/distances trees by using river formation dynamics". Ant Colony Optimization and Swarm Intelligence, Springer, LNCS 5217, pp. 60–71.
  18. P. Rabanal, I. Rodríguez, and F. Rubio (2009) "Applying River Formation Dynamics to Solve NP-Complete Problems". Nature-Inspired Algorithms for Optimisation, Springer, SCI 193, pp. 333–368.
  19. P. Rabanal, I. Rodríguez, and F. Rubio (2013) "Testing restorable systems: formal definition and heuristic solution based on river formation dynamics". Formal Aspects of Computing, Springer, Volume 25, Number 5, pp. 743–768.
  20. Vicsek, T.; Czirok, A.; Ben-Jacob, E.;; Cohen, I.; Shochet, O. (1995). "Novel type of phase transition in a system of self-driven particles". Physical Review Letters 75: 1226–1229. doi:10.1103/PhysRevLett.75.1226. PMID 10060237. 
  21. Reynolds, C. W. (1987). "Flocks, herds and schools: A distributed behavioral model". Computer Graphics 21 (4): 25–34. doi:10.1145/37401.37406. citeseerx-10.1.1.103.7187. 
  22. Czirók, A.; Vicsek, T. (2006). "Collective behavior of interacting self-propelled particles". Physica A 281: 17–29. doi:10.1016/S0378-4371(00)00013-3. 
  23. Buhl, J.; Sumpter, D.J.T.; Couzin, D.; Hale, J.J.; Despland, E.; Miller, E.R.; Simpson, S.J. et al. (2006). "From disorder to order in marching locusts" (PDF). Science 312 (5778): 1402–1406. doi:10.1126/science.1125142. PMID 16741126. http://webscript.princeton.edu/~icouzin/website/wp-content/plugins/bib2html/data/papers/buhl06.pdf. 
  24. Toner, J.; Tu, Y.; Ramaswamy, S. (2005). "Hydrodynamics and phases of flocks" (PDF). Annals of Physics 318: 170. doi:10.1016/j.aop.2005.04.011. http://eprints.iisc.ernet.in/3397/1/A89.pdf. 
  25. Bertin, E.; Droz, M.; Grégoire, G. (2009). "Hydrodynamic equations for self-propelled particles: microscopic derivation and stability analysis". J. Phys. A 42 (44): 445001. doi:10.1088/1751-8113/42/44/445001. 
  26. Li, Y.X.; Lukeman, R.; Edelstein-Keshet, L. et al. (2007). "Minimal mechanisms for school formation in self-propelled particles" (PDF). Physica D: Nonlinear Phenomena 237 (5): 699–720. doi:10.1016/j.physd.2007.10.009. http://www.iam.ubc.ca/~lukeman/fish_school_f.pdf. 
  27. Bishop, J.M., Stochastic Searching Networks, Proc. 1st IEE Int. Conf. on Artificial Neural Networks, pp. 329-331, London, UK, (1989).
  28. Nasuto, S.J. & Bishop, J.M., (2008), Stabilizing swarm intelligence search via positive feedback resource allocation, In: Krasnogor, N., Nicosia, G, Pavone, M., & Pelta, D. (eds), Nature Inspired Cooperative Strategies for Optimization, Studies in Computational Intelligence, vol 129, Springer, Berlin, Heidelberg, New York, pp. 115-123.
  29. Moglich, M.; Maschwitz, U.; Holldobler, B., Tandem Calling: A New Kind of Signal in Ant Communication, Science, Volume 186, Issue 4168, pp. 1046-1047
  30. Nasuto, S.J., Bishop, J.M. & Lauria, S., Time complexity analysis of the Stochastic Diffusion Search, Proc. Neural Computation '98, pp. 260-266, Vienna, Austria, (1998).
  31. Nasuto, S.J., & Bishop, J.M., (1999), Convergence of the Stochastic Diffusion Search, Parallel Algorithms, 14:2, pp: 89-107.
  32. Myatt, D.M., Bishop, J.M., Nasuto, S.J., (2004), Minimum stable convergence criteria for Stochastic Diffusion Search, Electronics Letters, 22:40, pp. 112-113.
  33. al-Rifaie, M.M., Bishop, J.M. & Blackwell, T., An investigation into the merger of stochastic diffusion search and particle swarm optimisation, Proc. 13th Conf. Genetic and Evolutionary Computation, (GECCO), pp.37-44, (2012).
  34. al-Rifaie, Mohammad Majid, John Mark Bishop, and Tim Blackwell. "Information sharing impact of stochastic diffusion search on differential evolution algorithm." Memetic Computing 4.4 (2012): 327-338.