A basic variant of the PSO algorithm works by having a population (called a swarm) of candidate solutions (called particles). These particles are moved around in the search-space according to a few simple formulae. The movements of the particles are guided by their own best-known position in the search-space as well as the entire swarm's best-known position. When improved positions are being discovered these will then come to guide the movements of the swarm. The process is repeated and by doing so it is hoped, but not guaranteed, that a satisfactory solution will eventually be discovered.
The values blo and bup represent the lower and upper boundaries of the search-space respectively. The w parameter is the inertia weight. The parameters φp and φg are often called cognitive coefficient and social coefficient.
The termination criterion can be the number of iterations performed, or a solution where the adequate objective function value is found. The parameters w, φp, and φg are selected by the practitioner and control the behaviour and efficacy of the PSO method (below).
The choice of PSO parameters can have a large impact on optimization performance. Selecting PSO parameters that yield good performance has therefore been the subject of much research.
To prevent divergence ("explosion") the inertia weight must be smaller than 1. The two other parameters can be then derived thanks to the constriction approach, or freely selected, but the analyses suggest convergence domains to constrain them. Typical values are in
[
1
,
3
]
{\displaystyle [1,3]}
.
The PSO parameters can also be tuned by using another overlaying optimizer, a concept known as meta-optimization, or even fine-tuned during the optimization, e.g., by means of fuzzy logic.
Parameters have also been tuned for various optimization scenarios.
The topology of the swarm defines the subset of particles with which each particle can exchange information. The basic version of the algorithm uses the global topology as the swarm communication structure. This topology allows all particles to communicate with all the other particles, thus the whole swarm share the same best position g from a single particle. However, this approach might lead the swarm to be trapped into a local minimum, thus different topologies have been used to control the flow of information among particles. For instance, in local topologies, particles only share information with a subset of particles. This subset can be a geometrical one – for example "the m nearest particles" – or, more often, a social one, i.e. a set of particles that is not depending on any distance. In such cases, the PSO variant is said to be local best (vs global best for the basic PSO).
A commonly used swarm topology is the ring, in which each particle has just two neighbours, but there are many others. The topology is not necessarily static. In fact, since the topology is related to the diversity of communication of the particles, some efforts have been done to create adaptive topologies (SPSO, APSO, stochastic star, TRIBES, Cyber Swarm, and C-PSO)
By using the ring topology, PSO can attain generation-level parallelism, significantly enhancing the evolutionary speed.
A common belief amongst researchers is that the swarm behaviour varies between exploratory behaviour, that is, searching a broader region of the search-space, and exploitative behaviour, that is, a locally oriented search so as to get closer to a (possibly local) optimum. This school of thought has been prevalent since the inception of PSO. This school of thought contends that the PSO algorithm and its parameters must be chosen so as to properly balance between exploration and exploitation to avoid premature convergence to a local optimum yet still ensure a good rate of convergence to the optimum. This belief is the precursor of many PSO variants, see below.
Another school of thought is that the behaviour of a PSO swarm is not well understood in terms of how it affects actual optimization performance, especially for higher-dimensional search-spaces and optimization problems that may be discontinuous, noisy, and time-varying. This school of thought merely tries to find PSO algorithms and parameters that cause good performance regardless of how the swarm behaviour can be interpreted in relation to e.g. exploration and exploitation. Such studies have led to the simplification of the PSO algorithm, see below.
Convergence of the sequence of solutions has been investigated for PSO. These analyses have resulted in guidelines for selecting PSO parameters that are believed to cause convergence to a point and prevent divergence of the swarm's particles (particles do not move unboundedly and will converge to somewhere). However, the analyses were criticized by Pedersen for being oversimplified as they assume the swarm has only one particle, that it does not use stochastic variables and that the points of attraction, that is, the particle's best known position p and the swarm's best known position g, remain constant throughout the optimization process. However, it was shown that these simplifications do not affect the boundaries found by these studies for parameter where the swarm is convergent. Considerable effort has been made in recent years to weaken the modeling assumption utilized during the stability analysis of PSO, with the most recent generalized result applying to numerous PSO variants and utilized what was shown to be the minimal necessary modeling assumptions.
Convergence to a local optimum has been analyzed for PSO in and. It has been proven that PSO needs some modification to guarantee finding a local optimum.
This means that determining the convergence capabilities of different PSO algorithms and parameters still depends on empirical results. One attempt at addressing this issue is the development of an "orthogonal learning" strategy for an improved use of the information already existing in the relationship between p and g, so as to form a leading converging exemplar and to be effective with any PSO topology. The aims are to improve the performance of PSO overall, including faster global convergence, higher solution quality, and stronger robustness. However, such studies do not provide theoretical evidence to actually prove their claims.
Without the need for a trade-off between convergence ('exploitation') and divergence ('exploration'), an adaptive mechanism can be introduced. Adaptive particle swarm optimization (APSO) features better search efficiency than standard PSO. APSO can perform global search over the entire search space with a higher convergence speed. It enables automatic control of the inertia weight, acceleration coefficients, and other algorithmic parameters at the run time, thereby improving the search effectiveness and efficiency at the same time. Also, APSO can act on the globally best particle to jump out of the likely local optima. However, APSO will introduce new algorithm parameters, it does not introduce additional design or implementation complexity nonetheless.
Besides, through the utilization of a scale-adaptive fitness evaluation mechanism, PSO can efficiently address computationally expensive optimization problems.
Numerous variants of even a basic PSO algorithm are possible. For example, there are different ways to initialize the particles and velocities (e.g. start with zero velocities instead), how to dampen the velocity, only update pi and g after the entire swarm has been updated, etc. Some of these choices and their possible performance impact have been discussed in the literature.
A series of standard implementations have been created by leading researchers, "intended for use both as a baseline for performance testing of improvements to the technique, as well as to represent PSO to the wider optimization community. Having a well-known, strictly-defined standard algorithm provides a valuable point of comparison which can be used throughout the field of research to better test new advances." The latest is Standard PSO 2011 (SPSO-2011).
In addition, some PSO variants have been developed to solve large-scale global optimization (LSGO) problems with more than 1000 dimensions. Representative variants include competitive swarm optimizer (CSO) and level-based learning swarm optimizer (LLSO). Recently, PSO has also been extended to solve multi-agent consensus-based distributed optimization problems, e.g., multi-agent consensus-based PSO with adaptive internal and external learning (MASOIE), etc.
New and more sophisticated PSO variants are also continually being introduced in an attempt to improve optimization performance. There are certain trends in that research; one is to make a hybrid optimization method using PSO combined with other optimizers, e.g., combined PSO with biogeography-based optimization, and the incorporation of an effective learning method.
Another research trend is to try to alleviate premature convergence (that is, optimization stagnation), e.g. by reversing or perturbing the movement of the PSO particles, another approach to deal with premature convergence is the use of multiple swarms (multi-swarm optimization). The multi-swarm approach can also be used to implement multi-objective optimization. Finally, there are developments in adapting the behavioural parameters of PSO during optimization.
Another school of thought is that PSO should be simplified as much as possible without impairing its performance; a general concept often referred to as Occam's razor. Simplifying PSO was originally suggested by Kennedy and has been studied more extensively, where it appeared that optimization performance was improved, and the parameters were easier to tune and they performed more consistently across different optimization problems.
Initialization of velocities may require extra inputs. The Bare Bones PSO variant has been proposed in 2003 by James Kennedy, and does not need to use velocity at all.
In this variant of PSO one dispences with the velocity of the particles and instead updates the positions of the particles using the following simple rule,
x
→
i
=
G
(
p
→
i
+
g
→
2
,
|
|
p
→
i
−
g
→
|
|
)
,
{\displaystyle {\vec {x}}_{i}=G\left({\frac {{\vec {p}}_{i}+{\vec {g}}}{2}},||{\vec {p}}_{i}-{\vec {g}}||\right)\,,}
where
x
→
i
{\displaystyle {\vec {x}}_{i}}
,
p
→
i
{\displaystyle {\vec {p}}_{i}}
are the position and the best position of the particle
i
{\displaystyle i}
;
g
→
{\displaystyle {\vec {g}}}
is the global best position;
G
(
x
→
,
σ
)
{\displaystyle G({\vec {x}},\sigma )}
is the normal distribution with the mean
x
→
{\displaystyle {\vec {x}}}
and standard deviation
σ
{\displaystyle \sigma }
; and where
|
|
…
|
|
{\displaystyle ||\dots ||}
signifies the norm of a vector.
Another simpler variant is the accelerated particle swarm optimization (APSO), which also does not need to use velocity and can speed up the convergence in many applications. A simple demo code of APSO is available.
In this variant of PSO one dispences with both the particle's velocity and the particle's best position. The particle position is updated according to the following rule,
x
→
i
←
(
1
−
β
)
x
→
i
+
β
g
→
+
α
L
u
→
,
{\displaystyle {\vec {x}}_{i}\leftarrow (1-\beta ){\vec {x}}_{i}+\beta {\vec {g}}+\alpha L{\vec {u}}\,,}
where
u
→
{\displaystyle {\vec {u}}}
is a random uniformly distributed vector,
L
{\displaystyle L}
is the typical length of the problem at hand, and
β
∼
0.1
−
0.7
{\displaystyle \beta \sim 0.1-0.7}
and
α
∼
0.1
−
0.5
{\displaystyle \alpha \sim 0.1-0.5}
are the parameters of the method. As a refinement of the method one can decrease
α
{\displaystyle \alpha }
with each iteration,
α
n
=
α
0
γ
n
{\displaystyle \alpha _{n}=\alpha _{0}\gamma ^{n}}
, where
n
{\displaystyle n}
is the number of the iteration and
0
<
γ
<
1
{\displaystyle 0<\gamma <1}
is the decrease control parameter.
As the PSO equations given above work on real numbers, a commonly used method to solve discrete problems is to map the discrete search space to a continuous domain, to apply a classical PSO, and then to demap the result. Such a mapping can be very simple (for example by just using rounded values) or more sophisticated.
However, it can be noted that the equations of movement make use of operators that perform four actions:
Bonyadi, M. R.; Michalewicz, Z. (2017). "Particle swarm optimization for single objective continuous space problems: a review". Evolutionary Computation. 25 (1): 1–54. doi:10.1162/EVCO_r_00180. PMID 26953883. S2CID 8783143. /wiki/Doi_(identifier)
Kennedy, J.; Eberhart, R. (1995). "Particle Swarm Optimization". Proceedings of IEEE International Conference on Neural Networks. Vol. IV. pp. 1942–1948. doi:10.1109/ICNN.1995.488968. /wiki/Doi_(identifier)
Shi, Y.; Eberhart, R.C. (1998). "A modified particle swarm optimizer". Proceedings of IEEE International Conference on Evolutionary Computation. pp. 69–73. doi:10.1109/ICEC.1998.699146. /wiki/Doi_(identifier)
Kennedy, J. (1997). "The particle swarm: social adaptation of knowledge". Proceedings of IEEE International Conference on Evolutionary Computation. pp. 303–308. doi:10.1109/ICEC.1997.592326. /wiki/Doi_(identifier)
Kennedy, J.; Eberhart, R.C. (2001). Swarm Intelligence. Morgan Kaufmann. ISBN 978-1-55860-595-4. 978-1-55860-595-4
Poli, R. (2007). "An analysis of publications on particle swarm optimisation applications" (PDF). Technical Report CSM-469. Archived from the original (PDF) on 2011-07-16. Retrieved 2010-05-03. https://web.archive.org/web/20110716231935/http://cswww.essex.ac.uk/technical-reports/2007/tr-csm469.pdf
Poli, R. (2008). "Analysis of the publications on the applications of particle swarm optimisation" (PDF). Journal of Artificial Evolution and Applications. 2008: 1–10. doi:10.1155/2008/685175. http://downloads.hindawi.com/archive/2008/685175.pdf
Bonyadi, M. R.; Michalewicz, Z. (2017). "Particle swarm optimization for single objective continuous space problems: a review". Evolutionary Computation. 25 (1): 1–54. doi:10.1162/EVCO_r_00180. PMID 26953883. S2CID 8783143. /wiki/Doi_(identifier)
Zhang, Y. (2015). "A Comprehensive Survey on Particle Swarm Optimization Algorithm and Its Applications". Mathematical Problems in Engineering. 2015: 931256. http://www.hindawi.com/journals/mpe/2015/931256
Clerc, M. (2012). "Standard Particle Swarm Optimisation" (PDF). HAL Open Access Archive. http://hal.archives-ouvertes.fr/docs/00/76/49/96/PDF/SPSO_descriptions.pdf
Bratton, Daniel; Kennedy, James (2007). "Defining a Standard for Particle Swarm Optimization". 2007 IEEE Swarm Intelligence Symposium (PDF). pp. 120–127. doi:10.1109/SIS.2007.368035. ISBN 978-1-4244-0708-8. S2CID 6217309. 978-1-4244-0708-8
Taherkhani, M.; Safabakhsh, R. (2016). "A novel stability-based adaptive inertia weight for particle swarm optimization". Applied Soft Computing. 38: 281–295. doi:10.1016/j.asoc.2015.10.004. /wiki/Doi_(identifier)
Shi, Y.; Eberhart, R.C. (1998). "Parameter selection in particle swarm optimization". Proceedings of Evolutionary Programming VII (EP98). pp. 591–600.
Eberhart, R.C.; Shi, Y. (2000). "Comparing inertia weights and constriction factors in particle swarm optimization". Proceedings of the Congress on Evolutionary Computation. Vol. 1. pp. 84–88. https://www.researchgate.net/publication/3865142
Carlisle, A.; Dozier, G. (2001). "An Off-The-Shelf PSO" (PDF). Proceedings of the Particle Swarm Optimization Workshop. pp. 1–6. Archived from the original (PDF) on 2003-05-03. https://web.archive.org/web/20030503203304/http://antho.huntingdon.edu/publications/Off-The-Shelf_PSO.pdf
van den Bergh, F. (2001). An Analysis of Particle Swarm Optimizers (PDF) (PhD thesis). University of Pretoria, Faculty of Natural and Agricultural Science. http://bee22.com/resources/Bergh%202006.pdf
Clerc, M.; Kennedy, J. (2002). "The particle swarm - explosion, stability, and convergence in a multidimensional complex space". IEEE Transactions on Evolutionary Computation. 6 (1): 58–73. CiteSeerX 10.1.1.460.6608. doi:10.1109/4235.985692. /wiki/CiteSeerX_(identifier)
Trelea, I.C. (2003). "The Particle Swarm Optimization Algorithm: convergence analysis and parameter selection". Information Processing Letters. 85 (6): 317–325. doi:10.1016/S0020-0190(02)00447-7. /wiki/Doi_(identifier)
Bratton, D.; Blackwell, T. (2008). "A Simplified Recombinant PSO" (PDF). Journal of Artificial Evolution and Applications. 2008: 1–10. doi:10.1155/2008/654184. http://downloads.hindawi.com/archive/2008/654184.pdf
Evers, G. (2009). An Automatic Regrouping Mechanism to Deal with Stagnation in Particle Swarm Optimization. The University of Texas - Pan American, Department of Electrical Engineering. Archived from the original (Master's thesis) on 2011-05-18. Retrieved 2010-05-05. https://web.archive.org/web/20110518164430/http://www.georgeevers.org/publications.htm
Clerc, M.; Kennedy, J. (2002). "The particle swarm - explosion, stability, and convergence in a multidimensional complex space". IEEE Transactions on Evolutionary Computation. 6 (1): 58–73. CiteSeerX 10.1.1.460.6608. doi:10.1109/4235.985692. /wiki/CiteSeerX_(identifier)
Meissner, M.; Schmuker, M.; Schneider, G. (2006). "Optimized Particle Swarm Optimization (OPSO) and its application to artificial neural network training". BMC Bioinformatics. 7 (1): 125. doi:10.1186/1471-2105-7-125. PMC 1464136. PMID 16529661. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC1464136
Pedersen, M.E.H. (2010). Tuning & Simplifying Heuristical Optimization (PDF). University of Southampton, School of Engineering Sciences, Computational Engineering and Design Group. S2CID 107805461. Archived from the original (PhD thesis) on 2020-02-13. https://web.archive.org/web/20200213130101/https://pdfs.semanticscholar.org/a5d2/8c26a2e2824170d67b69f14120cf67cabe26.pdf
Pedersen, M.E.H.; Chipperfield, A.J. (2010). "Simplifying particle swarm optimization". Applied Soft Computing. 10 (2): 618–628. CiteSeerX 10.1.1.149.8300. doi:10.1016/j.asoc.2009.08.029. /wiki/CiteSeerX_(identifier)
Mason, Karl; Duggan, Jim; Howley, Enda (2018). "A Meta Optimisation Analysis of Particle Swarm Optimisation Velocity Update Equations for Watershed Management Learning". Applied Soft Computing. 62: 148–161. doi:10.1016/j.asoc.2017.10.018. /wiki/Doi_(identifier)
Nobile, M.S; Cazzaniga, P.; Besozzi, D.; Colombo, R.; Mauri, G.; Pasi, G. (2018). "Fuzzy Self-Tuning PSO: a settings-free algorithm for global optimization". Swarm and Evolutionary Computation. 39: 70–85. doi:10.1016/j.swevo.2017.09.001. hdl:10446/106467. /wiki/Doi_(identifier)
Nobile, M.S; Pasi, G.; Cazzaniga, P.; Besozzi, D.; Colombo, R.; Mauri, G. (2015). "Proactive particles in swarm optimization: a self-tuning algorithm based on fuzzy logic". Proceedings of the 2015 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE 2015), Istanbul (Turkey). pp. 1–8. doi:10.1109/FUZZ-IEEE.2015.7337957. /wiki/Doi_(identifier)
Cazzaniga, P.; Nobile, M.S.; Besozzi, D. (2015). "The impact of particles initialization in PSO: parameter estimation as a case in point, (Canada)". Proceedings of IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology. doi:10.1109/CIBCB.2015.7300288. /wiki/Doi_(identifier)
Pedersen, M.E.H. (2010). "Good parameters for particle swarm optimization". Technical Report HL1001. CiteSeerX 10.1.1.298.4359. /wiki/CiteSeerX_(identifier)
Kennedy, J.; Mendes, R. (2002). "Population structure and particle swarm performance". Proceedings of the 2002 Congress on Evolutionary Computation. CEC'02 (Cat. No.02TH8600). Vol. 2. pp. 1671–1676 vol.2. CiteSeerX 10.1.1.114.7988. doi:10.1109/CEC.2002.1004493. ISBN 978-0-7803-7282-5. S2CID 14364974. 978-0-7803-7282-5
Bratton, Daniel; Kennedy, James (2007). "Defining a Standard for Particle Swarm Optimization". 2007 IEEE Swarm Intelligence Symposium (PDF). pp. 120–127. doi:10.1109/SIS.2007.368035. ISBN 978-1-4244-0708-8. S2CID 6217309. 978-1-4244-0708-8
Mendes, R. (2004). Population Topologies and Their Influence in Particle Swarm Performance (PhD thesis). Universidade do Minho. https://pdfs.semanticscholar.org/d224/80b09d1f0759fb20e0fb0bd2de205457c8bc.pdf
Bratton, Daniel; Kennedy, James (2007). "Defining a Standard for Particle Swarm Optimization". 2007 IEEE Swarm Intelligence Symposium (PDF). pp. 120–127. doi:10.1109/SIS.2007.368035. ISBN 978-1-4244-0708-8. S2CID 6217309. 978-1-4244-0708-8
Suganthan, Ponnuthurai N. "Particle swarm optimiser with neighbourhood operator." Evolutionary Computation, 1999. CEC 99. Proceedings of the 1999 Congress on. Vol. 3. IEEE, 1999. https://ieeexplore.ieee.org/abstract/document/785514/
Bratton, Daniel; Kennedy, James (2007). "Defining a Standard for Particle Swarm Optimization". 2007 IEEE Swarm Intelligence Symposium (PDF). pp. 120–127. doi:10.1109/SIS.2007.368035. ISBN 978-1-4244-0708-8. S2CID 6217309. 978-1-4244-0708-8
Oliveira, M.; Pinheiro, D.; Andrade, B.; Bastos-Filho, C.; Menezes, R. (2016). "Communication Diversity in Particle Swarm Optimizers". Swarm Intelligence. Lecture Notes in Computer Science. Vol. 9882. pp. 77–88. doi:10.1007/978-3-319-44427-7_7. ISBN 978-3-319-44426-0. S2CID 37588745. 978-3-319-44426-0
SPSO Particle Swarm Central http://www.particleswarm.info
Almasi, O. N. and Khooban, M. H. (2017). A parsimonious SVM model selection criterion for classification of real-world data sets via an adaptive population-based algorithm. Neural Computing and Applications, 1-9. https://doi.org/10.1007/s00521-017-2930-y https://link.springer.com/article/10.1007/s00521-017-2930-y
Miranda, V., Keko, H. and Duque, Á. J. (2008). Stochastic Star Communication Topology in Evolutionary Particle Swarms (EPSO). International Journal of Computational Intelligence Research (IJCIR), Volume 4, Number 2, pp. 105-116 https://repositorio.inesctec.pt/bitstream/123456789/1561/1/PS-05818.pdf
Clerc, M. (2006). Particle Swarm Optimization. ISTE (International Scientific and Technical Encyclopedia), 2006
Yin, P., Glover, F., Laguna, M., & Zhu, J. (2011). A Complementary Cyber Swarm Algorithm. International Journal of Swarm Intelligence Research (IJSIR), 2(2), 22-41 http://leeds-faculty.colorado.edu/glover/fred%20pubs/428%20-%20A_complementary_cyber_swarm_algorithm_pub%20version%20w%20pen%20et%20al.pdf
Elshamy, W.; Rashad, H.; Bahgat, A. (2007). "Clubs-based Particle Swarm Optimization" (PDF). IEEE Swarm Intelligence Symposium 2007 (SIS2007). Honolulu, HI. pp. 289–296. Archived from the original (PDF) on 2013-10-23. Retrieved 2012-04-27. https://web.archive.org/web/20131023025232/http://people.cis.ksu.edu/~welshamy/pubs/ieee_sis07.pdf
Jian-Yu, Li (2021). "Generation-Level Parallelism for Evolutionary Computation: A Pipeline-Based Parallel Particle Swarm Optimization". IEEE Transactions on Cybernetics. 51 (10): 4848–4859. doi:10.1109/TCYB.2020.3028070. PMID 33147159. https://ieeexplore.ieee.org/document/9248594
Shi, Y.; Eberhart, R.C. (1998). "A modified particle swarm optimizer". Proceedings of IEEE International Conference on Evolutionary Computation. pp. 69–73. doi:10.1109/ICEC.1998.699146. /wiki/Doi_(identifier)
Kennedy, J. (1997). "The particle swarm: social adaptation of knowledge". Proceedings of IEEE International Conference on Evolutionary Computation. pp. 303–308. doi:10.1109/ICEC.1997.592326. /wiki/Doi_(identifier)
Shi, Y.; Eberhart, R.C. (1998). "Parameter selection in particle swarm optimization". Proceedings of Evolutionary Programming VII (EP98). pp. 591–600.
Clerc, M.; Kennedy, J. (2002). "The particle swarm - explosion, stability, and convergence in a multidimensional complex space". IEEE Transactions on Evolutionary Computation. 6 (1): 58–73. CiteSeerX 10.1.1.460.6608. doi:10.1109/4235.985692. /wiki/CiteSeerX_(identifier)
van den Bergh, F. (2001). An Analysis of Particle Swarm Optimizers (PDF) (PhD thesis). University of Pretoria, Faculty of Natural and Agricultural Science. http://bee22.com/resources/Bergh%202006.pdf
Clerc, M.; Kennedy, J. (2002). "The particle swarm - explosion, stability, and convergence in a multidimensional complex space". IEEE Transactions on Evolutionary Computation. 6 (1): 58–73. CiteSeerX 10.1.1.460.6608. doi:10.1109/4235.985692. /wiki/CiteSeerX_(identifier)
Trelea, I.C. (2003). "The Particle Swarm Optimization Algorithm: convergence analysis and parameter selection". Information Processing Letters. 85 (6): 317–325. doi:10.1016/S0020-0190(02)00447-7. /wiki/Doi_(identifier)
Pedersen, M.E.H.; Chipperfield, A.J. (2010). "Simplifying particle swarm optimization". Applied Soft Computing. 10 (2): 618–628. CiteSeerX 10.1.1.149.8300. doi:10.1016/j.asoc.2009.08.029. /wiki/CiteSeerX_(identifier)
Cleghorn, Christopher W (2014). "Particle Swarm Convergence: Standardized Analysis and Topological Influence". Swarm Intelligence. Lecture Notes in Computer Science. Vol. 8667. pp. 134–145. doi:10.1007/978-3-319-09952-1_12. ISBN 978-3-319-09951-4. 978-3-319-09951-4
Liu, Q (2015). "Order-2 stability analysis of particle swarm optimization". Evolutionary Computation. 23 (2): 187–216. doi:10.1162/EVCO_a_00129. PMID 24738856. S2CID 25471827. /wiki/Doi_(identifier)
Cleghorn, Christopher W.; Engelbrecht, Andries. (2018). "Particle Swarm Stability: A Theoretical Extension using the Non-Stagnate Distribution Assumption". Swarm Intelligence. 12 (1): 1–22. doi:10.1007/s11721-017-0141-x. hdl:2263/62934. S2CID 9778346. /wiki/Doi_(identifier)
Van den Bergh, F. "A convergence proof for the particle swarm optimizer" (PDF). Fundamenta Informaticae. https://repository.up.ac.za/bitstream/handle/2263/17262/VanDenBergh_Convergence(2010).pdf?sequence=1
Bonyadi, Mohammad reza.; Michalewicz, Z. (2014). "A locally convergent rotationally invariant particle swarm optimization algorithm" (PDF). Swarm Intelligence. 8 (3): 159–198. doi:10.1007/s11721-014-0095-1. S2CID 2261683. https://espace.library.uq.edu.au/view/UQ:396054/ERAUQ396054.pdf
Zhan, Z-H.; Zhang, J.; Li, Y; Shi, Y-H. (2011). "Orthogonal Learning Particle Swarm Optimization" (PDF). IEEE Transactions on Evolutionary Computation. 15 (6): 832–847. doi:10.1109/TEVC.2010.2052054. http://eprints.gla.ac.uk/44801/1/44801.pdf
Zhan, Z-H.; Zhang, J.; Li, Y; Chung, H.S-H. (2009). "Adaptive Particle Swarm Optimization" (PDF). IEEE Transactions on Systems, Man, and Cybernetics. 39 (6): 1362–1381. doi:10.1109/TSMCB.2009.2015956. PMID 19362911. S2CID 11191625. http://eprints.gla.ac.uk/7645/1/7645.pdf
Wang, Ye-Qun; Li, Jian-Yu; Chen, Chun-Hua; Zhang, Jun; Zhan, Zhi-Hui (September 2023). "Scale adaptive fitness evaluation-based particle swarm optimisation for hyperparameter and architecture optimisation in neural networks and deep learning". CAAI Transactions on Intelligence Technology. 8 (3): 849-862. doi:10.1049/cit2.12106. https://doi.org/10.1049%2Fcit2.12106
Carlisle, A.; Dozier, G. (2001). "An Off-The-Shelf PSO" (PDF). Proceedings of the Particle Swarm Optimization Workshop. pp. 1–6. Archived from the original (PDF) on 2003-05-03. https://web.archive.org/web/20030503203304/http://antho.huntingdon.edu/publications/Off-The-Shelf_PSO.pdf
Bratton, Daniel; Kennedy, James (2007). "Defining a Standard for Particle Swarm Optimization". 2007 IEEE Swarm Intelligence Symposium (PDF). pp. 120–127. doi:10.1109/SIS.2007.368035. ISBN 978-1-4244-0708-8. S2CID 6217309. 978-1-4244-0708-8
Zambrano-Bigiarini, M.; Clerc, M.; Rojas, R. (2013). "Standard Particle Swarm Optimisation 2011 at CEC-2013: A baseline for future PSO improvements". 2013 IEEE Congress on Evolutionary Computation. Evolutionary Computation (CEC), 2013 IEEE Congress on. pp. 2337–2344. doi:10.1109/CEC.2013.6557848. ISBN 978-1-4799-0454-9. S2CID 206553432. 978-1-4799-0454-9
Yang, Q.; CHEN, W-N.; Deng, J-D.; Li, Y.; Gu, T.; Zhang, J. (2018). "A Level-based Learning Swarm Optimizer for Large Scale Optimization". IEEE Transactions on Evolutionary Computation. 22 (4): 578–594. doi:10.1109/TEVC.2017.2743016. https://ieeexplore.ieee.org/document/8025724
Chen, T-Y.; Chen, W-N.; Wei, F-F.; Hu, X-M.; Zhang, J. (2024). "Multi-Agent Swarm Optimization With Adaptive Internal and External Learning for Complex Consensus-Based Distributed Optimization". IEEE Transactions on Evolutionary Computation: 1. doi:10.1109/TEVC.2024.3380436. https://ieeexplore.ieee.org/document/10477458
Lovbjerg, M.; Krink, T. (2002). "The LifeCycle Model: combining particle swarm optimisation, genetic algorithms and hillclimbers" (PDF). Proceedings of Parallel Problem Solving from Nature VII (PPSN). pp. 621–630. http://www.lovbjerghome.dk/Morten/EvaLife/TK_PPSN2002_LifeCycle_PSO_HC_GA.pdf
Niknam, T.; Amiri, B. (2010). "An efficient hybrid approach based on PSO, ACO and k-means for cluster analysis". Applied Soft Computing. 10 (1): 183–197. doi:10.1016/j.asoc.2009.07.001. /wiki/Doi_(identifier)
Zhang, Wen-Jun; Xie, Xiao-Feng (2003). DEPSO: hybrid particle swarm with differential evolution operator. IEEE International Conference on Systems, Man, and Cybernetics (SMCC), Washington, DC, USA: 3816-3821. http://www.wiomax.com/team/xie/paper/SMCC03.pdf
Zhang, Y.; Wang, S. (2015). "Pathological Brain Detection in Magnetic Resonance Imaging Scanning by Wavelet Entropy and Hybridization of Biogeography-based Optimization and Particle Swarm Optimization". Progress in Electromagnetics Research. 152: 41–58. doi:10.2528/pier15040602. https://doi.org/10.2528%2Fpier15040602
Zhan, Z-H.; Zhang, J.; Li, Y; Shi, Y-H. (2011). "Orthogonal Learning Particle Swarm Optimization" (PDF). IEEE Transactions on Evolutionary Computation. 15 (6): 832–847. doi:10.1109/TEVC.2010.2052054. http://eprints.gla.ac.uk/44801/1/44801.pdf
Evers, G. (2009). An Automatic Regrouping Mechanism to Deal with Stagnation in Particle Swarm Optimization. The University of Texas - Pan American, Department of Electrical Engineering. Archived from the original (Master's thesis) on 2011-05-18. Retrieved 2010-05-05. https://web.archive.org/web/20110518164430/http://www.georgeevers.org/publications.htm
Lovbjerg, M.; Krink, T. (2002). "Extending Particle Swarm Optimisers with Self-Organized Criticality" (PDF). Proceedings of the Fourth Congress on Evolutionary Computation (CEC). Vol. 2. pp. 1588–1593. http://www.lovbjerghome.dk/Morten/EvaLife/ML_CEC2002_SOCPSO.pdf
Xinchao, Z. (2010). "A perturbed particle swarm algorithm for numerical optimization". Applied Soft Computing. 10 (1): 119–124. doi:10.1016/j.asoc.2009.06.010. /wiki/Doi_(identifier)
Xie, Xiao-Feng; Zhang, Wen-Jun; Yang, Zhi-Lian (2002). A dissipative particle swarm optimization. Congress on Evolutionary Computation (CEC), Honolulu, HI, USA: 1456-1461. http://www.wiomax.com/team/xie/paper/CEC02.pdf
Cheung, N. J.; Ding, X.-M.; Shen, H.-B. (2013). "OptiFel: A Convergent Heterogeneous Particle Sarm Optimization Algorithm for Takagi-Sugeno Fuzzy Modeling". IEEE Transactions on Fuzzy Systems. 22 (4): 919–933. doi:10.1109/TFUZZ.2013.2278972. S2CID 27974467. /wiki/Doi_(identifier)
Nobile, M.; Besozzi, D.; Cazzaniga, P.; Mauri, G.; Pescini, D. (2012). "A GPU-Based Multi-Swarm PSO Method for Parameter Estimation in Stochastic Biological Systems Exploiting Discrete-Time Target Series". Evolutionary Computation, Machine Learning and Data Mining in Bioinformatics. Lecture Notes in Computer Science. Vol. 7264. pp. 74–85. doi:10.1007/978-3-642-29066-4_7. /wiki/Doi_(identifier)
Zhan, Z-H.; Zhang, J.; Li, Y; Chung, H.S-H. (2009). "Adaptive Particle Swarm Optimization" (PDF). IEEE Transactions on Systems, Man, and Cybernetics. 39 (6): 1362–1381. doi:10.1109/TSMCB.2009.2015956. PMID 19362911. S2CID 11191625. http://eprints.gla.ac.uk/7645/1/7645.pdf
Nobile, M.S; Cazzaniga, P.; Besozzi, D.; Colombo, R.; Mauri, G.; Pasi, G. (2018). "Fuzzy Self-Tuning PSO: a settings-free algorithm for global optimization". Swarm and Evolutionary Computation. 39: 70–85. doi:10.1016/j.swevo.2017.09.001. hdl:10446/106467. /wiki/Doi_(identifier)
Kennedy, J. (1997). "The particle swarm: social adaptation of knowledge". Proceedings of IEEE International Conference on Evolutionary Computation. pp. 303–308. doi:10.1109/ICEC.1997.592326. /wiki/Doi_(identifier)
Bratton, D.; Blackwell, T. (2008). "A Simplified Recombinant PSO" (PDF). Journal of Artificial Evolution and Applications. 2008: 1–10. doi:10.1155/2008/654184. http://downloads.hindawi.com/archive/2008/654184.pdf
Pedersen, M.E.H. (2010). Tuning & Simplifying Heuristical Optimization (PDF). University of Southampton, School of Engineering Sciences, Computational Engineering and Design Group. S2CID 107805461. Archived from the original (PhD thesis) on 2020-02-13. https://web.archive.org/web/20200213130101/https://pdfs.semanticscholar.org/a5d2/8c26a2e2824170d67b69f14120cf67cabe26.pdf
Pedersen, M.E.H.; Chipperfield, A.J. (2010). "Simplifying particle swarm optimization". Applied Soft Computing. 10 (2): 618–628. CiteSeerX 10.1.1.149.8300. doi:10.1016/j.asoc.2009.08.029. /wiki/CiteSeerX_(identifier)
Yang, X.S. (2008). Nature-Inspired Metaheuristic Algorithms. Luniver Press. ISBN 978-1-905986-10-1. 978-1-905986-10-1
Tu, Z.; Lu, Y. (2004). "A robust stochastic genetic algorithm (StGA) for global numerical optimization". IEEE Transactions on Evolutionary Computation. 8 (5): 456–470. doi:10.1109/TEVC.2004.831258. S2CID 22382958. /wiki/Doi_(identifier)
Tu, Z.; Lu, Y. (2008). "Corrections to "A Robust Stochastic Genetic Algorithm (StGA) for Global Numerical Optimization"". IEEE Transactions on Evolutionary Computation. 12 (6): 781. doi:10.1109/TEVC.2008.926734. S2CID 2864886. /wiki/Doi_(identifier)
Kennedy, James (2003). "Bare bones particle swarms". Proceedings of the 2003 IEEE Swarm Intelligence Symposium. SIS'03 (Cat. No.03EX706). pp. 80–87. doi:10.1109/SIS.2003.1202251. ISBN 0-7803-7914-4. S2CID 37185749. 0-7803-7914-4
X. S. Yang, S. Deb and S. Fong, Accelerated particle swarm optimization and support vector machine for business optimization and applications, NDT 2011, Springer CCIS 136, pp. 53-66 (2011). https://arxiv.org/abs/1203.6577
"Search Results: APSO - File Exchange - MATLAB Central". http://www.mathworks.com/matlabcentral/fileexchange/?term=APSO
Parsopoulos, K.; Vrahatis, M. (2002). "Particle swarm optimization method in multiobjective problems". Proceedings of the ACM Symposium on Applied Computing (SAC). pp. 603–607. doi:10.1145/508791.508907. /wiki/Doi_(identifier)
Coello Coello, C.; Salazar Lechuga, M. (2002). "MOPSO: A Proposal for Multiple Objective Particle Swarm Optimization". Congress on Evolutionary Computation (CEC'2002). pp. 1051–1056. http://portal.acm.org/citation.cfm?id=1252327
Mason, Karl; Duggan, Jim; Howley, Enda (2017). "Multi-objective dynamic economic emission dispatch using particle swarm optimisation variants". Neurocomputing. 270: 188–197. doi:10.1016/j.neucom.2017.03.086. /wiki/Doi_(identifier)
Roy, R., Dehuri, S., & Cho, S. B. (2012). A Novel Particle Swarm Optimization Algorithm for Multi-Objective Combinatorial Optimization Problem. 'International Journal of Applied Metaheuristic Computing (IJAMC)', 2(4), 41-57 http://sclab.yonsei.ac.kr/publications/Papers/IJ/A%20Novel%20Particle%20Swarm%20Optimization%20Algorithm%20for%20Multi-Objective%20Combinatorial%20Optimization%20Problem.pdf
Kennedy, J. & Eberhart, R. C. (1997). A discrete binary version of the particle swarm algorithm, Conference on Systems, Man, and Cybernetics, Piscataway, NJ: IEEE Service Center, pp. 4104-4109 http://ahmetcevahircinar.com.tr/wp-content/uploads/2017/02/A_discrete_binary_version_of_the_particle_swarm_algorithm.pdf
Clerc, M. (2004). Discrete Particle Swarm Optimization, illustrated by the Traveling Salesman Problem, New Optimization Techniques in Engineering, Springer, pp. 219-239 https://link.springer.com/chapter/10.1007/978-3-540-39930-8_8
Clerc, M. (2005). Binary Particle Swarm Optimisers: toolbox, derivations, and mathematical insights, Open Archive HAL http://hal.archives-ouvertes.fr/hal-00122809/en/
Jarboui, B.; Damak, N.; Siarry, P.; Rebai, A. (2008). "A combinatorial particle swarm optimization for solving multi-mode resource-constrained project scheduling problems". Applied Mathematics and Computation. 195: 299–308. doi:10.1016/j.amc.2007.04.096. /wiki/Doi_(identifier)
Chen, Wei-neng; Zhang, Jun (2010). "A novel set-based particle swarm optimization method for discrete optimization problem". IEEE Transactions on Evolutionary Computation. 14 (2): 278–300. CiteSeerX 10.1.1.224.5378. doi:10.1109/tevc.2009.2030331. S2CID 17984726. /wiki/CiteSeerX_(identifier)