• Title/Summary/Keyword: cosine-modulated gaussian function

Search Result 2, Processing Time 0.017 seconds

An Improvement of Performance for Cascade Correlation Learning Algorithm using a Cosine Modulated Gaussian Activation Function (코사인 모듈화 된 가우스 활성화 함수를 사용한 캐스케이드 코릴레이션 학습 알고리즘의 성능 향상)

  • Lee, Sang-Wha;Song, Hae-Sang
    • Journal of the Korea Society of Computer and Information
    • /
    • v.11 no.3
    • /
    • pp.107-115
    • /
    • 2006
  • This paper presents a new class of activation functions for Cascade Correlation learning algorithm, which herein will be called CosGauss function. This function is a cosine modulated gaussian function. In contrast to the sigmoidal, hyperbolic tangent and gaussian functions, more ridges can be obtained by the CosGauss function. Because of the ridges, it is quickly convergent and improves a pattern recognition speed. Consequently it will be able to improve a learning capability. This function was tested with a Cascade Correlation Network on the two spirals problem and results are compared with those obtained with other activation functions.

  • PDF

Approximation of Polynomials and Step function for cosine modulated Gaussian Function in Neural Network Architecture (뉴로 네트워크에서 코사인 모듈화 된 가우스함수의 다항식과 계단함수의 근사)

  • Lee, Sang-Wha
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.49 no.2
    • /
    • pp.115-122
    • /
    • 2012
  • We present here a new class of activation functions for neural networks, which herein will be called CosGauss function. This function is a cosine-modulated gaussian function. In contrast to the sigmoidal-, hyperbolic tangent- and gaussian activation functions, more ridges can be obtained by the CosGauss function. It will be proven that this function can be used to aproximate polynomials and step functions. The CosGauss function was tested with a Cascade-Correlation-Network of the multilayer structure on the Tic-Tac-Toe game and iris plants problems, and results are compared with those obtained with other activation functions.