Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
2007, Studies in Computational Intelligence
AI
Particle Swarm Optimization (PSO) is a robust optimization technique, traditionally effective for static problems, but its effectiveness in dynamic environments has gained attention. This research explores the modifications needed for PSO to effectively find and track changing optima, focusing on challenges like diversity loss and outdated memory. Proposed mechanisms for overcoming these challenges include randomization, repulsion, dynamic networks, and multi-populations, supported by empirical results demonstrating improved performance in dynamic optimization scenarios.
2009
In the real world, many applications are nonstationary optimization problems. This requires that optimization algorithms need to not only find the global optimal solution but also track the trajectory of the changing global best solution in a dynamic environment. To achieve this, this paper proposes a clustering particle swarm optimizer (CPSO) for dynamic optimization problems. The algorithm employs hierarchical clustering method to track multiple peaks based on a nearest neighbor search strategy. A fast local search method is also proposed to find the near optimal solutions in a local promising region in the search space. Six test problems generated from a generalized dynamic benchmark generator (GDBG) are used to test the performance of the proposed algorithm. The numerical experimental results show the efficiency of the proposed algorithm for locating and tracking multiple optima in dynamic environments.
2007
In recent years, there has been an increasing concern from the evolutionary computation community on dynamic optimization problems since many real-world optimization problems are time-varying. In this paper, a triggered memory scheme is introduced into the particle swarm optimization to deal with dynamic environments. The triggered memory scheme enhances traditional memory scheme with a triggered memory generator. Experimental study over a benchmark dynamic problem shows that the triggered memory-based particle swarm optimization algorithm has stronger robustness and adaptability than traditional particle swarm optimization algorithms, both with and without traditional memory scheme, for dynamic optimization problems.
2013 IEEE Congress on Evolutionary Computation, CEC 2013, 2013
This paper investigates whether the optimal parameter configurations for particle swarm optimizers (PSO) change when changes in the search landscape occur. To test this, specific environmental changes that may occur during dynamic function optimization are deliberately constructed, using the moving peaks function generator. The parameters of the charged-and quantum PSO algorithms are then optimized for the initial environment, as well as for each of the constructed problems. It is shown that the optimal parameter configurations for the various environments differ not only with respect to the initial optimal configurations, but also with respect to each other. The results lead to the conclusion that PSO parameters need to be re-optimized or selfadapted whenever environmental changes are detected.
Particle swarm optimization is a heuristic global optimization method put forward originally by Doctor Kennedy and Eberhart in 1995. Various efforts have been made for solving unimodal and multimodal problems as well as two dimensional to multidimensional problems. Efforts were put towards topology of communication, parameter adjustment, initial distribution of particles and efficient problem solving capabilities. Here we presented detail study of PSO and limitation in present work. Based on the limitation we proposed future direction. I. INTRODUCTION Swarm Intelligence (SI) is an innovative distributed intelligent paradigm for solving optimization problems that originally took its inspiration from the biological examples by swarming, flocking and herding phenomena in vertebrates. Particle Swarm Optimization (PSO) incorporates swarming behaviors observed in flocks of birds, schools of fish, or swarms of bees, and even human social behavior, from which the idea is emerged. PSO is a population-based optimization tool, which could be implemented and applied easily to solve various function optimization problems, or the problems that can be transformed to function optimization problems. As an algorithm, the main strength of PSO is its fast convergence, which compares favorably with many global optimization algorithms like Genetic Algorithms (GA), Simulated Annealing (SA) and other global optimization algorithms. While population-based heuristics are more costly because of their dependency directly upon function values rather than derivative information, they are however susceptible to premature convergence, which is especially the case when there are many decision variables or dimensions to be optimized. Particle swarm optimization is a heuristic global optimization method put forward originally by Doctor Kennedy and Eberhart in 1995. While searching for food, the birds are either scattered or go together before they locate the place where they can find the food. While the birds are searching for food from one place to another, there is always a bird that can smell the food very well, that is, the bird is perceptible of the place where the food can be found, having the better food resource information. Because they are transmitting the information, especially the good information at any time while searching the food from one place to another, conduced by the good information, the birds will eventually flock to the place where food can be found. As far as particle swam optimization algorithm is concerned, solution swam is compared to the bird swarm, the birds' moving from one place to another is equal to the development of the solution swarm, good information is equal to the most optimist solution, and the food resource is equal to the most optimist solution during the whole course. The most optimist solution can be worked out in particle swarm optimization algorithm by the cooperation of each individual. The particle without quality and volume serves as each individual, and the simple behavioral pattern is regulated for each particle to show the complexity of the whole particle swarm. In PSO, the potential solution called particles fly through the problem space by following the current optimum particles. Each particles keeps tracks of its coordinates in the problem space which are associated with the best solution (fitness) achieved so far. This value is called as pbest. Another best value that is tracked by the particle swarm optimizer is the best value, obtained so far by any particle in the neighbors of the particle. This value is called lbest. When a particle takes all the population as its topological neighbors, the best value is a global best and is called gbest. The particle swarm optimization concept consists of, at each time step, changing the velocity of (accelerating) each particle toward its pbest and lbest (for lbest version). Acceleration is weighted by random term, with separate random numbers being generated for acceleration towards pbest and lbest locations. After finding the best values, the particle updates its velocity and positions with following equations.
IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 2000
In recent years, there has been a growing interest in the study of particle swarm optimization (PSO) in dynamic environments. This paper presents a new PSO model, called PSO with composite particles (PSO-CP), to address dynamic optimization problems. PSO-CP partitions the swarm into a set of composite particles based on their similarity using a "worst first" principle. Inspired by the composite particle phenomenon in physics, the elementary members in each composite particle interact via a velocity-anisotropic reflection scheme to integrate valuable information for effectively and rapidly finding the promising optima in the search space. Each composite particle maintains the diversity by a scattering operator. In addition, an integral movement strategy is introduced to promote the swarm diversity. Experiments on a typical dynamic test benchmark problem provide a guideline for setting the involved parameters and show that PSO-CP is efficient in comparison with several state-of-the-art PSO algorithms for dynamic optimization problems.
Evolutionary Computation, 2014
The multi-population method has been widely used to solve dynamic optimization problems (DOPs) with the aim of maintaining multiple populations on different peaks to locate and track multiple changing optima simultaneously. However, to make this approach effective for solving DOPs, two challenging issues need to be addressed. They are how to adapt the number of populations to changes and how to adaptively maintain the population diversity in a situation where changes are complicated or hard to detect or predict. Tracking the changing global optimum in dynamic environments is difficult because we cannot know when and where changes occur and what the characteristics of changes would be. Therefore, we need to design algorithms that are able to adapt to changes by taking the challenging issues into account. To address the issues when multi-population methods are applied for solving DOPs, this paper proposes an adaptive multi-swarm algorithm, where the populations are enabled to be adaptive in dynamic environments without change detection. An experimental study is conducted based on the moving peaks problem to investigate the behavior of the proposed method. The performance of the proposed algorithm is also compared with a set of algorithms that are based on multi-population methods from different research areas in the literature of evolutionary computation.
2007
Evolutionary Algorithms (EAs) can be used for designing Particle Swarm Optimization (PSO) algorithms that work, in some cases, considerably better than the human-designed ones. By analyzing the evolutionary process of design PSO algorithm we can identify different swarm phenomena (such as patterns or rules) that can give us deep insights about the swarm's behaviours. The observed rules can help us to design better PSO algorithms for optimization. In this paper we investigate and analyze swarm phenomena by looking to process of evolving PSO algorithms. Several interesting facts are inferred from the strategy evolution process (the particle quality could influence the update order, some particles are updated more frequently than others are, the initial swarm size is not always optimal).
Swarm Intelligence, 2012
Optimization in dynamic optimization problems (DOPs) requires the optimization algorithms not only to locate, but also to continuously track the moving optima. Particle swarm optimization (PSO) is a population-based optimization algorithm, originally developed for static problems. Recently, several researchers have proposed variants of PSO for optimization in DOPs. This paper presents a novel multi-swarm PSO algorithm, namely competitive clustering PSO (CCPSO), designed specially for DOPs. Employing a multistage clustering procedure, CCPSO splits the particles of the main swarm over a number of sub-swarms based on the particles positions and on their objective function values. The algorithm automatically adjusts the number of sub-swarms and the corresponding region of each sub-swarm. In addition to the sub-swarms, there is also a group of free particles that explore the environment to locate new emerging optima or exploit the current optima which are not followed by any sub-swarm. The adaptive search strategy adopted by the sub-swarms improves both the exploitation and tracking characteristics of CCPSO. A set of experiments is conducted to study the behavior of the proposed algorithm in different DOPs and to provide guidelines for setting the algorithm's parameters in different problems. The results of CCPSO on a variety of moving peaks benchmark (MPB) functions are compared with those of several state-of-the-art PSO algorithms, indicating the efficiency of the proposed model.
Computing Research Repository, 2005
In order to overcome difficult dynamic optimization and environment extrema tracking problems, we propose a Self-Regulated Swarm (SRS) algorithm which hybridizes the advantageous characteristics of Swarm Intelligence as the emergence of a societal environmental memory or cognitive map via collective pheromone laying in the landscape (properly balancing the exploration/exploitation nature of the search strategy), with a simple Evolutionary mechanism that trough a direct reproduction procedure linked to local environmental features is able to self-regulate the above exploratory swarm population, speeding it up globally. In order to test his adaptive response and robustness, we have recurred to different dynamic multimodal complex functions as well as to Dynamic Optimization Control (DOC) problems. Measures were made for different dynamic settings and parameters such as, environmental upgrade frequencies, landscape changing speed severity, type of dynamic (linear or circular), and to dramatic changes on the algorithmic search purpose over each test environment (e.g. shifting the extrema). Finally, comparisons were made with traditional Genetic Algorithms (GA) as well as with more recently proposed Co-Evolutionary approaches. SRS, were able to demonstrate quick adaptive responses, while outperforming the results obtained by the other approaches. Additionally, some successful behaviors were found: SRS was able not only to achieve quick adaptive responses, as to maintaining a number of different solutions, while adapting to new unforeseen extrema; the possibility to spontaneously create and maintain different subpopulations on different peaks, emerging different exploratory corridors with intelligent path planning capabilities; the ability to request for new agents over dramatic changing periods, and economizing those foraging resources over periods of stabilization. Finally, results prove that the present SRS collective swarm of bio-inspired agents is able to track about 65% of moving peaks traveling up to ten times faster than the velocity of a single ant composing that precise swarm tracking system. This emerged behavior is probably one of the most interesting ones achieved by the present work. M Ramos, V., Fernandes, C., Rosa, A.C., Tracking Extrema over Dynamic Environments using Self-Regulatory Swarms.
2020
In real life, mostly problems are dynamic. Many algorithms have been proposed to handle the static problems, but these algorithms do not handle or poorly handle the dynamic environment problems. Although, many algorithms have been proposed to handle dynamic problems but still, there are some limitations or drawbacks in every algorithm regarding diversity of particles and tracking of already found optima. To overcome these limitations/drawbacks, we have proposed a new efficient algorithm to handle the dynamic environment effectively by tracking and locating multiple optima and by improving the diversity and convergence speed of algorithm. In this algorithm, a new method has been proposed which explore the undiscovered areas of search space to increase the diversity of algorithm. This algorithm also uses a method to effectively handle the overlapped and overcrowded particles. Branke has proposed moving peak benchmark which is commonly used MBP in literature. We also have performed dif...
Swarm and Evolutionary Computation, 2017
Swarm intelligence (SI) algorithms, including ant colony optimization, particle swarm optimization, bee-inspired algorithms, bacterial foraging optimization, firefly algorithms, fish swarm optimization and many more, have been proven to be good methods to address difficult optimization problems under stationary environments. Most SI algorithms have been developed to address stationary optimization problems and hence, they can converge on the (near-) optimum solution efficiently. However, many real-world problems have a dynamic environment that changes over time. For such dynamic optimization problems (DOPs), it is difficult for a conventional SI algorithm to track the changing optimum once the algorithm has converged on a solution. In the last two decades, there has been a growing interest of addressing DOPs using SI algorithms due to their adaptation capabilities. This paper presents a broad review on SI dynamic optimization (SIDO) focused on several classes of problems, such as discrete, continuous, constrained, multi-objective and classification problems, and real-world applications. In addition, this paper focuses on the enhancement strategies integrated in SI algorithms to address dynamic changes, the performance measurements and benchmark generators used in SIDO. Finally, some considerations about future directions in the subject are given.
Each particle in a swarm maintains its current position and its personal best position. It is useful to think of these personal best positions as a population of attractors – updates to current positions are based on attractions to these personal best positions. If the population of attractors has high diversity, it will encourage a broad exploration of the search space with particles being drawn in many different directions. However, the population of attractors can converge quickly – attractors can draw other particles towards them, and these particles can update their own personal bests to be near the first attractor. This convergence of attractors can be reduced by having a particle update the attractor it has approached rather than its own attractor/personal best. This simple change to the update procedure in particle swarm optimization incurs minimal computational cost, and it can lead to large performance improvements in multi-modal search spaces.
Natural Computing, 2010
A hybrid approach called Evolutionary Swarm Cooperative Algorithm (ESCA) based on the collaboration between a particle swarm optimization algorithm and an evolutionary algorithm is presented. ESCA is designed to deal with moving optima of optimization problems in dynamic environments. ESCA uses three populations of individuals: two EA populations and one Particle Swarm Population. The EA populations evolve by the rules of an evolutionary multimodal optimization algorithm being used to maintain the diversity of the search. The particle swarm confers precision to the search process. The efficiency of ESCA is evaluated by means of numerical experiments.
PeerJ Computer Science, 2019
This paper investigates the performance and scalability of a new update strategy for the particle swarm optimization (PSO) algorithm. The strategy is inspired by the Bak-Sneppen model of co-evolution between interacting species, which is basically a network of fitness values (representing species) that change over time according to a simple rule: the least fit species and its neighbors are iteratively replaced with random values. Following these guidelines, a steady state and dynamic update strategy for PSO algorithms is proposed: only the least fit particle and its neighbors are updated and evaluated in each time-step; the remaining particles maintain the same position and fitness, unless they meet the update criterion. The steady state PSO was tested on a set of unimodal, multimodal, noisy and rotated benchmark functions, significantly improving the quality of results and convergence speed of the standard PSOs and more sophisticated PSOs with dynamic parameters and neighborhood. A sensitivity analysis of the parameters confirms the performance enhancement with different parameter settings and scalability tests show that the algorithm behavior is consistent throughout a substantial range of solution vector dimensions.
Proceedings 2005 IEEE Swarm Intelligence Symposium, 2005. SIS 2005., 2005
Genetic programming (GP) is used to create fitness landscapes which highlight strengths and weaknesses of different types of PSO and to contrast population-based swarm approaches with non stochastic gradient followers (i.e. hill climbers). These automatically generated benchmark problems yield insights into the operation of PSOs, illustrate benefits and drawbacks of different population sizes and constriction (friction) coefficients, and reveal new swarm phenomena such as deception and the exploration/exploitation tradeoff. The method could be applied to any type of optimizer.
2008
Adaptation to dynamic optimization problems is currently receiving a growing interest as one of the most important applications of evolutionary algorithms. In this paper, a compound particle swarm optimization (CPSO) is proposed as a new variant of particle swarm optimization to enhance its performance in dynamic environments. Within CPSO, compound particles are constructed as a novel type of particles in the search space and their motions are integrated into the swarm. A special reflection scheme is introduced in order to explore the search space more comprehensively. Furthermore, some information preserving and anti-convergence strategies are also developed to improve the performance of CPSO in a new environment. An experimental study shows the efficiency of CPSO in dynamic environments.
Swarm Intelligence (SI) describes the evolving collective intelligence of population/groups of autonomous agents with a low level of intelligence. Particle Swarm Optimization (PSO) is an evolutionary algorithm inspired by animal social behaviour. PSO achieves performance by iteratively directing its particles toward the optimum using its social and cognitive components. Various modifications have been applied to PSO focusing on addressing a variety of methods for adjusting PSO's parameters (i.e., parameter adjustment), social interaction of the particles (i.e., neighbourhood topology) and ways to address the search objectives (i.e., sub-swarm topology). The PSO approach can easily fit in different search optimization categories such as Self Learning, Unsupervised Learning, Stochastic Search, Population-based, and Behaviour-based Search. This study addresses these principal aspects of PSO. In addition, conventional and Basic PSO are introduced and their shortcomings are discussed. Later on, various suggestions and modifications proposed by literature are introduced and discussed.
Loading Preview
Sorry, preview is currently unavailable. You can download the paper by clicking the button above.