• Title/Summary/Keyword: Premature convergence phenomenon

Search Result 7, Processing Time 0.018 seconds

Selective Mutation for Performance Improvement of Genetic Algorithms (유전자알고리즘의 성능향상을 위한 선택적 돌연변이)

  • Jung, Sung-Hoon
    • The KIPS Transactions:PartB
    • /
    • v.17B no.2
    • /
    • pp.149-156
    • /
    • 2010
  • Since the premature convergence phenomenon of genetic algorithms (GAs) degrades the performances of GAs significantly, solving this problem provides a lot of effects to the performances of GAs. In this paper, we propose a selective mutation method in order to improve the performances of GAs by alleviating this phenomenon. In the selective mutation, individuals are additionally mutated at the specific region according to their ranks. From this selective mutation, individuals with low ranks are changed a lot and those with high ranks are changed small in the phenotype. Finally, some good individuals search around them in detail and the other individuals have more chances to search new areas. This results in enhancing the performances of GAs through alleviating of the premature convergence phenomenon. We measured the performances of our method with four typical function optimization problems. It was found from experiments that our proposed method considerably improved the performances of GAs.

Speeding-up for error back-propagation algorithm using micro-genetic algorithms (미소-유전 알고리듬을 이용한 오류 역전파 알고리듬의 학습 속도 개선 방법)

  • 강경운;최영길;심귀보;전홍태
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1993.10a
    • /
    • pp.853-858
    • /
    • 1993
  • The error back-propagation(BP) algorithm is widely used for finding optimum weights of multi-layer neural networks. However, the critical drawback of the BP algorithm is its slow convergence of error. The major reason for this slow convergence is the premature saturation which is a phenomenon that the error of a neural network stays almost constant for some period time during learning. An inappropriate selections of initial weights cause each neuron to be trapped in the premature saturation state, which brings in slow convergence speed of the multi-layer neural network. In this paper, to overcome the above problem, Micro-Genetic algorithms(.mu.-GAs) which can allow to find the near-optimal values, are used to select the proper weights and slopes of activation function of neurons. The effectiveness of the proposed algorithms will be demonstrated by some computer simulations of two d.o.f planar robot manipulator.

  • PDF

Preventing Premature Convergence in Genetic Algorithms with Adaptive Population Size (유전자 집단의 크기 조절을 통한 Genetic Algorithm의 조기 포화 방지)

  • 박래정;박철훈
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.32B no.12
    • /
    • pp.1680-1686
    • /
    • 1995
  • GAs, effective stochastic search algorithms based on the model of natural evolution and genetics, have been successfully applied to various optimization problems. When population size is not large, GAs often suffer from the phenomenon of premature convergence in which all chromosomes in the population lose the diversity of genes before they find the optimal solution. In this paper, we propose that a new heuristic that maintains the diversity of genes by adding some chromosomes with random mutation and selective mutation into population during evolution. And population size changes dynamically with supplement of new chromosomes. Experimental results for several test functions show that when population size is rather small and the length of chromosome is not long, this method is effective.

  • PDF

Fast Optimization by Queen-bee Evolution and Derivative Evaluation in Genetic Algorithms

  • Jung, Sung-Hoon
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.5 no.4
    • /
    • pp.310-315
    • /
    • 2005
  • This paper proposes a fast optimization method by combining queen-bee evolution and derivative evaluation in genetic algorithms. These two operations make it possible for genetic algorithms to focus on highly fitted individuals and rapidly evolved individuals, respectively. Even though the two operations can also increase the probability that genetic algorithms fall into premature convergence phenomenon, that can be controlled by strong mutation rates. That is, the two operations and the strong mutation strengthen exploitation and exploration of the genetic algorithms, respectively. As a result, the genetic algorithm employing queen-bee evolution and derivative evaluation finds optimum solutions more quickly than those employing one of them. This was proved by experiments with one pattern matching problem and two function optimization problems.

Proportional-Integral-Derivative Evaluation for Enhancing Performance of Genetic Algorithms (유전자 알고리즘의 성능향상을 위한 비례-적분-미분 평가방법)

  • Jung, Sung-Hoon
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.13 no.4
    • /
    • pp.439-447
    • /
    • 2003
  • This paper proposes a proportional-integral-derivative (PID) evaluation method for enhancing performance of genetic algorithms. In PID evaluation, the fitness of individuals is evaluated by not only the fitness derived from an evaluation function, but also the parents fitness of each individual and the minimum and maximum fitness from initial generation to previous generation. This evaluation decreases the probability that the genetic algorithms fall into a premature convergence phenomenon and results in enhancing the performance of genetic algorithms. We experimented our evaluation method with typical numerical function optimization problems. It was found from extensive experiments that out evaluation method can increase the performance of genetic algorithms greatly. This evaluation method can be easily applied to the other types of genetic algorithms for improving their performance.

A Novel Hybrid Intelligence Algorithm for Solving Combinatorial Optimization Problems

  • Deng, Wu;Chen, Han;Li, He
    • Journal of Computing Science and Engineering
    • /
    • v.8 no.4
    • /
    • pp.199-206
    • /
    • 2014
  • The ant colony optimization (ACO) algorithm is a new heuristic algorithm that offers good robustness and searching ability. With in-depth exploration, the ACO algorithm exhibits slow convergence speed, and yields local optimization solutions. Based on analysis of the ACO algorithm and the genetic algorithm, we propose a novel hybrid genetic ant colony optimization (NHGAO) algorithm that integrates multi-population strategy, collaborative strategy, genetic strategy, and ant colony strategy, to avoid the premature phenomenon, dynamically balance the global search ability and local search ability, and accelerate the convergence speed. We select the traveling salesman problem to demonstrate the validity and feasibility of the NHGAO algorithm for solving complex optimization problems. The simulation experiment results show that the proposed NHGAO algorithm can obtain the global optimal solution, achieve self-adaptive control parameters, and avoid the phenomena of stagnation and prematurity.

Performance Improvement of Queen-bee Genetic Algorithms through Multiple Queen-bee Evolution (다중 여왕벌 진화를 통한 여왕벌 유전자알고리즘의 성능향상)

  • Jung, Sung-Hoon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.17 no.4
    • /
    • pp.129-137
    • /
    • 2012
  • The queen-bee genetic algorithm that we made by mimicking of the reproduction of queen-bee has considerably improved the performances of genetic algorithm. However, since we used only one queen-bee in the queen-bee genetic algorithm, a problem that individuals of genetic algorithm were driven to one place where the queen-bee existed occurred. This made the performances of the queen-bee genetic algorithm degrade. In order to solve this problem, we introduce a multiple queen-bee evolution method by employing another queen-bee whose fitness is the most significantly increased than its parents as well as the original queen-bee that is the best individual in a generation. This multiple queen-bee evolution makes the probability of falling into local optimum areas decrease and allows the individuals to easily get out of the local optimum areas even if the individuals fall into a local optimum area. This results in increasing the performances of the genetic algorithm. Experimental results with four function optimization problems showed that the performances of the proposed method were better than those of the existing method in the most cases.