A novel heuristic search algorithm called seeker op- timization algorithm (SOA) is proposed for the real-parameter optimization. The proposed SOA is based on simulating the act of human searching. In the SOA, search...A novel heuristic search algorithm called seeker op- timization algorithm (SOA) is proposed for the real-parameter optimization. The proposed SOA is based on simulating the act of human searching. In the SOA, search direction is based on empir- ical gradients by evaluating the response to the position changes, while step length is based on uncertainty reasoning by using a simple fuzzy rule. The effectiveness of the SOA is evaluated by using a challenging set of typically complex functions in compari- son to differential evolution (DE) and three modified particle swarm optimization (PSO) algorithms. The simulation results show that the performance of the SOA is superior or comparable to that of the other algorithms.展开更多
Evolutionary computation is a kind of adaptive non--numerical computation method which is designed tosimulate evolution of nature. In this paper, evolutionary algorithm behavior is described in terms of theconstructio...Evolutionary computation is a kind of adaptive non--numerical computation method which is designed tosimulate evolution of nature. In this paper, evolutionary algorithm behavior is described in terms of theconstruction and evolution of the sampling distributions over the space of candidate solutions. Iterativeconstruction of the sampling distributions is based on the idea of the global random search of generationalmethods. Under this frame, propontional selection is characterized as a gobal search operator, and recombination is characerized as the search process that exploits similarities. It is shown-that by properly constraining the search breadth of recombination operators, weak convergence of evolutionary algorithms to aglobal optimum can be ensured.展开更多
A modification of evolutionary programming or evolution strategies for ndimensional global optimization is proposed. Based on the ergodicity and inherentrandomness of chaos, the main characteristic of the new algorith...A modification of evolutionary programming or evolution strategies for ndimensional global optimization is proposed. Based on the ergodicity and inherentrandomness of chaos, the main characteristic of the new algorithm which includes two phases is that chaotic behavior is exploited to conduct a rough search of the problem space in order to find the promising individuals in Phase I. Adjustment strategy of steplength and intensive searches in Phase II are employed. The population sequences generated by the algorithm asymptotically converge to global optimal solutions with probability one. The proposed algorithm is applied to several typical test problems. Numerical results illustrate that this algorithm can more efficiently solve complex global optimization problems than evolutionary programming and evolution strategies in most cases.展开更多
In this paper, we develop a new theoretical framework by means of the absorbing Markov process theory for analyzing some stochastic global optimization algorithms. Applying the framework to the pure random search, we ...In this paper, we develop a new theoretical framework by means of the absorbing Markov process theory for analyzing some stochastic global optimization algorithms. Applying the framework to the pure random search, we prove that the pure random search converges to the global minimum in probability and its time has geometry distribution. We also analyze the pure adaptive search by this framework and turn out that the pure adaptive search converges to the global minimum in probability and its time has Poisson distribution.展开更多
This paper presents an improved gravitational search algorithm (IGSA) as a hybridization of a relatively recent evolutionary algorithm called gravitational search algorithm (GSA), with the free search differential...This paper presents an improved gravitational search algorithm (IGSA) as a hybridization of a relatively recent evolutionary algorithm called gravitational search algorithm (GSA), with the free search differential evolution (FSDE). This combination incorporates FSDE into the optimization process of GSA with an attempt to avoid the premature convergence in GSA. This strategy makes full use of the exploration ability of GSA and the exploitation ability of FSDE. IGSA is tested on a suite of benchmark functions. The experimental results demonstrate the good performance of IGSA.展开更多
In this paper, the improvement of pure random search is studied. By taking some information of the function to be minimized into consideration, the authors propose two stochastic global optimization algorithms. Some n...In this paper, the improvement of pure random search is studied. By taking some information of the function to be minimized into consideration, the authors propose two stochastic global optimization algorithms. Some numerical experiments for the new stochastic global optimization algorithms are presented for a class of test problems.展开更多
In this paper, the principle of Cuckoo algorithm is introduced, and the traditional Cuckoo algorithm is improved to establish a mathematical model of multi-objective optimization scheduling. Based on the improved algo...In this paper, the principle of Cuckoo algorithm is introduced, and the traditional Cuckoo algorithm is improved to establish a mathematical model of multi-objective optimization scheduling. Based on the improved algorithm, the model is optimized to a certain extent. Through analysis, it is proved that the improved algorithm has higher computational accuracy and can effectively improve the global convergence.展开更多
Particle Swarm Optimization (PSO) is a new optimization algorithm, which is applied in many fields widely. But the original PSO is likely to cause the local optimization with premature convergence phenomenon. By using...Particle Swarm Optimization (PSO) is a new optimization algorithm, which is applied in many fields widely. But the original PSO is likely to cause the local optimization with premature convergence phenomenon. By using the idea of simulated annealing algo-rithm, we propose a modified algorithm which makes the most optimal particle of every time of iteration evolving continu-ously, and assign the worst particle with a new value to increase its disturbance. By the testing of three classic testing functions, we conclude the modified PSO algorithm has the better performance of convergence and global searching than the original PSO.展开更多
A clonal selection based memetic algorithm is proposed for solving job shop scheduling problems in this paper. In the proposed algorithm, the clonal selection and the local search mechanism are designed to enhance exp...A clonal selection based memetic algorithm is proposed for solving job shop scheduling problems in this paper. In the proposed algorithm, the clonal selection and the local search mechanism are designed to enhance exploration and exploitation. In the clonal selection mechanism, clonal selection, hypermutation and receptor edit theories are presented to construct an evolutionary searching mechanism which is used for exploration. In the local search mechanism, a simulated annealing local search algorithm based on Nowicki and Smutnicki's neighborhood is presented to exploit local optima. The proposed algorithm is examined using some well-known benchmark problems. Numerical results validate the effectiveness of the proposed algorithm.展开更多
Coordinate descent method is a unconstrained optimization technique. When it is applied to support vector machine (SVM), at each step the method updates one component of w by solving a one-variable sub-problem while...Coordinate descent method is a unconstrained optimization technique. When it is applied to support vector machine (SVM), at each step the method updates one component of w by solving a one-variable sub-problem while fixing other components. All components of w update after one iteration. Then go to next iteration. Though the method converges and converges fast in the beginning, it converges slow for final convergence. To improve the speed of final convergence of coordinate descent method, Hooke and Jeeves algorithm which adds pattern search after every iteration in coordinate descent method was applied to SVM and a global Newton algorithm was used to solve one-variable subproblems. We proved the convergence of the algorithm. Experimental results show Hooke and Jeeves' method does accelerate convergence specially for final convergence and achieves higher testing accuracy more quickly in classification.展开更多
基金supported by the National Natural Science Foundation of China(60870004)
文摘A novel heuristic search algorithm called seeker op- timization algorithm (SOA) is proposed for the real-parameter optimization. The proposed SOA is based on simulating the act of human searching. In the SOA, search direction is based on empir- ical gradients by evaluating the response to the position changes, while step length is based on uncertainty reasoning by using a simple fuzzy rule. The effectiveness of the SOA is evaluated by using a challenging set of typically complex functions in compari- son to differential evolution (DE) and three modified particle swarm optimization (PSO) algorithms. The simulation results show that the performance of the SOA is superior or comparable to that of the other algorithms.
文摘Evolutionary computation is a kind of adaptive non--numerical computation method which is designed tosimulate evolution of nature. In this paper, evolutionary algorithm behavior is described in terms of theconstruction and evolution of the sampling distributions over the space of candidate solutions. Iterativeconstruction of the sampling distributions is based on the idea of the global random search of generationalmethods. Under this frame, propontional selection is characterized as a gobal search operator, and recombination is characerized as the search process that exploits similarities. It is shown-that by properly constraining the search breadth of recombination operators, weak convergence of evolutionary algorithms to aglobal optimum can be ensured.
文摘A modification of evolutionary programming or evolution strategies for ndimensional global optimization is proposed. Based on the ergodicity and inherentrandomness of chaos, the main characteristic of the new algorithm which includes two phases is that chaotic behavior is exploited to conduct a rough search of the problem space in order to find the promising individuals in Phase I. Adjustment strategy of steplength and intensive searches in Phase II are employed. The population sequences generated by the algorithm asymptotically converge to global optimal solutions with probability one. The proposed algorithm is applied to several typical test problems. Numerical results illustrate that this algorithm can more efficiently solve complex global optimization problems than evolutionary programming and evolution strategies in most cases.
文摘In this paper, we develop a new theoretical framework by means of the absorbing Markov process theory for analyzing some stochastic global optimization algorithms. Applying the framework to the pure random search, we prove that the pure random search converges to the global minimum in probability and its time has geometry distribution. We also analyze the pure adaptive search by this framework and turn out that the pure adaptive search converges to the global minimum in probability and its time has Poisson distribution.
基金supported by the National Natural Science Foundation of China (70871081)the Shanghai Leading Academic Discipline Project of China (S1205YLXK)
文摘This paper presents an improved gravitational search algorithm (IGSA) as a hybridization of a relatively recent evolutionary algorithm called gravitational search algorithm (GSA), with the free search differential evolution (FSDE). This combination incorporates FSDE into the optimization process of GSA with an attempt to avoid the premature convergence in GSA. This strategy makes full use of the exploration ability of GSA and the exploitation ability of FSDE. IGSA is tested on a suite of benchmark functions. The experimental results demonstrate the good performance of IGSA.
文摘In this paper, the improvement of pure random search is studied. By taking some information of the function to be minimized into consideration, the authors propose two stochastic global optimization algorithms. Some numerical experiments for the new stochastic global optimization algorithms are presented for a class of test problems.
文摘In this paper, the principle of Cuckoo algorithm is introduced, and the traditional Cuckoo algorithm is improved to establish a mathematical model of multi-objective optimization scheduling. Based on the improved algorithm, the model is optimized to a certain extent. Through analysis, it is proved that the improved algorithm has higher computational accuracy and can effectively improve the global convergence.
文摘Particle Swarm Optimization (PSO) is a new optimization algorithm, which is applied in many fields widely. But the original PSO is likely to cause the local optimization with premature convergence phenomenon. By using the idea of simulated annealing algo-rithm, we propose a modified algorithm which makes the most optimal particle of every time of iteration evolving continu-ously, and assign the worst particle with a new value to increase its disturbance. By the testing of three classic testing functions, we conclude the modified PSO algorithm has the better performance of convergence and global searching than the original PSO.
文摘A clonal selection based memetic algorithm is proposed for solving job shop scheduling problems in this paper. In the proposed algorithm, the clonal selection and the local search mechanism are designed to enhance exploration and exploitation. In the clonal selection mechanism, clonal selection, hypermutation and receptor edit theories are presented to construct an evolutionary searching mechanism which is used for exploration. In the local search mechanism, a simulated annealing local search algorithm based on Nowicki and Smutnicki's neighborhood is presented to exploit local optima. The proposed algorithm is examined using some well-known benchmark problems. Numerical results validate the effectiveness of the proposed algorithm.
基金supported by the National Natural Science Foundation of China (6057407560705004)
文摘Coordinate descent method is a unconstrained optimization technique. When it is applied to support vector machine (SVM), at each step the method updates one component of w by solving a one-variable sub-problem while fixing other components. All components of w update after one iteration. Then go to next iteration. Though the method converges and converges fast in the beginning, it converges slow for final convergence. To improve the speed of final convergence of coordinate descent method, Hooke and Jeeves algorithm which adds pattern search after every iteration in coordinate descent method was applied to SVM and a global Newton algorithm was used to solve one-variable subproblems. We proved the convergence of the algorithm. Experimental results show Hooke and Jeeves' method does accelerate convergence specially for final convergence and achieves higher testing accuracy more quickly in classification.