• Title/Summary/Keyword: back-propagation learning algorithm

Search Result 386, Processing Time 0.028 seconds

A Modified Error Back Propagation Algorithm Adding Neurons to Hidden Layer (은닉층 뉴우런 추가에 의한 역전파 학습 알고리즘)

  • 백준호;김유신;손경식
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.29B no.4
    • /
    • pp.58-65
    • /
    • 1992
  • In this paper new back propagation algorithm which adds neurons to hidden layer is proposed. this proposed algorithm is applied to the pattern recognition of written number coupled with back propagation algorithm through omitting redundant learning. Learning rate and recognition rate of the proposed algorithm are compared with those of the conventional back propagation algorithm and the back propagation through omitting redundant learning. The learning rate of proposed algorithm is 4 times as fast as the conventional back propagation algorithm and 2 times as fast as the back propagation through omitting redundant learning. The recognition rate is 96.2% in case of the conventional back propagation algorithm, 96.5% in case of the back propagation through omitting redundant learning and 97.4% in the proposed algorithm.

  • PDF

Estimating Regression Function with $\varepsilon-Insensitive$ Supervised Learning Algorithm

  • Hwang, Chang-Ha
    • Journal of the Korean Data and Information Science Society
    • /
    • v.15 no.2
    • /
    • pp.477-483
    • /
    • 2004
  • One of the major paradigms for supervised learning in neural network community is back-propagation learning. The standard implementations of back-propagation learning are optimal under the assumptions of identical and independent Gaussian noise. In this paper, for regression function estimation, we introduce $\varepsilon-insensitive$ back-propagation learning algorithm, which corresponds to minimizing the least absolute error. We compare this algorithm with support vector machine(SVM), which is another $\varepsilon-insensitive$ supervised learning algorithm and has been very successful in pattern recognition and function estimation problems. For comparison, we consider a more realistic model would allow the noise variance itself to depend on the input variables.

  • PDF

Edge detection method using unbalanced mutation operator in noise image (잡음 영상에서 불균등 돌연변이 연산자를 이용한 효율적 에지 검출)

  • Kim, Su-Jung;Lim, Hee-Kyoung;Seo, Yo-Han;Jung, Chai-Yeoung
    • The KIPS Transactions:PartB
    • /
    • v.9B no.5
    • /
    • pp.673-680
    • /
    • 2002
  • This paper proposes a method for detecting edge using an evolutionary programming and a momentum back-propagation algorithm. The evolutionary programming does not perform crossover operation as to consider reduction of capability of algorithm and calculation cost, but uses selection operator and mutation operator. The momentum back-propagation algorithm uses assistant to weight of learning step when weight is changed at learning step. Because learning rate o is settled as less in last back-propagation algorithm the momentum back-propagation algorithm discard the problem that learning is slow as relative reduction because change rate of weight at each learning step. The method using EP-MBP is batter than GA-BP method in both learning time and detection rate and showed the decreasing learning time and effective edge detection, in consequence.

On the set up to the Number of Hidden Node of Adaptive Back Propagation Neural Network (적응 역전파 신경회로망의 은닉 층 노드 수 설정에 관한 연구)

  • Hong, Bong-Wha
    • The Journal of Information Technology
    • /
    • v.5 no.2
    • /
    • pp.55-67
    • /
    • 2002
  • This paper presents an adaptive back propagation algorithm that update the learning parameter by the generated error, adaptively and varies the number of hidden layer node. This algorithm is expected to escaping from the local minimum and make the best environment for convergence to be change the number of hidden layer node. On the simulation tested this algorithm on two learning pattern. One was exclusive-OR learning and the other was $7{\times}5$ dot alphabetic font learning. In both examples, the probability of becoming trapped in local minimum was reduce. Furthermore, in alphabetic font learning, the neural network enhanced to learning efficient about 41.56%~58.28% for the conventional back propagation. and HNAD(Hidden Node Adding and Deleting) algorithm.

  • PDF

On the enhancement of the learning efficiency of the adaptive back propagation neural network using the generating and adding the hidden layer node (은닉층 노드의 생성추가를 이용한 적응 역전파 신경회로망의 학습능률 향상에 관한 연구)

  • Kim, Eun-Won;Hong, Bong-Wha
    • Journal of the Institute of Electronics Engineers of Korea TE
    • /
    • v.39 no.2
    • /
    • pp.66-75
    • /
    • 2002
  • This paper presents an adaptive back propagation algorithm that its able to enhancement for the learning efficiency with updating the learning parameter and varies the number of hidden layer node by the generated error, adaptively. This algorithm is expected to escaping from the local minimum and make the best environment for the convergence of the back propagation neural network. On the simulation tested this algorithm on three learning pattern. One was exclusive-OR learning and the another was 3-parity problem and 7${\times}$5 dot alphabetic font learning. In result that the probability of becoming trapped in local minimum was reduce. Furthermore, the neural network enhanced to learning efficient about 17.6%~64.7% for the existed back propagation. 

Implementation of Speed Sensorless Induction Motor drives by Fast Learning Neural Network using RLS Approach

  • Kim, Yoon-Ho;Kook, Yoon-Sang
    • Proceedings of the KIPE Conference
    • /
    • 1998.10a
    • /
    • pp.293-297
    • /
    • 1998
  • This paper presents a newly developed speed sensorless drive using RLS based on Neural Network Training Algorithm. The proposed algorithm has just the time-varying learning rate, while the wellknown back-propagation algorithm based on gradient descent has a constant learning rate. The number of iterations required by the new algorithm to converge is less than that of the back-propagation algorithm. The theoretical analysis and experimental results to verify the effectiveness of the proposed control strategy are described.

  • PDF

Classification of ECG Arrhythmia Signals Using Back-Propagation Network (역전달 신경회로망을 이용한 심전도 파형의 부정맥 분류)

  • 권오철;최진영
    • Journal of Biomedical Engineering Research
    • /
    • v.10 no.3
    • /
    • pp.343-350
    • /
    • 1989
  • A new algorithm classifying ECG Arrhythmia signals using Back-propagation network is proposed. The base-line of ECG signal is detected by high pass filter and probability density function then input data are normalized for learning and classifying. In addition, ECG data are scanned to classify Arrhythmia signal which is hard to find R-wave. A two-layer perceptron with one hidden layer along with error back-propagation learning rule is utilized as an artificial neural network. The proposed algorithm shows outstanding performance under circumstances of amplitude variation, baseline wander and noise contamination.

  • PDF

A study on the realization of color printed material check using Error Back-Propagation rule (오류 역전파법으로구현한 컬러 인쇄물 검사에 관한 연구)

  • 한희석;이규영
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 1998.10a
    • /
    • pp.560-567
    • /
    • 1998
  • This paper concerned about a imputed color printed material image in camera to decrease noise and distortion by processing median filtering with input image to identical condition. Also this paper proposed the way of compares a normal printed material with an abnormal printed material color tone with trained a learning of the error back-propagation to block classification by extracting five place from identical block(3${\times}$3) of color printed material R, G, B value. As a representative algorithm of multi-layer perceptron the error Back-propagation technique used to solve complex problems. However, the Error Back-propagation is algorithm which basically used a gradient descent method which can be converged to local minimum and the Back Propagation train include problems, and that may converge in a local minimum rather than get a global minimum. The network structure appropriate for a given problem. In this paper, a good result is obtained by improve initial condition and adjust th number of hidden layer to solve the problem of real time process, learning and train.

  • PDF

Implementation of back propagation algorithm for wearable devices using FPGA (FPGA를 이용한 웨어러블 디바이스를 위한 역전파 알고리즘 구현)

  • Choi, Hyun-Sik
    • The Journal of Korean Institute of Next Generation Computing
    • /
    • v.15 no.2
    • /
    • pp.7-16
    • /
    • 2019
  • Neural networks can be implemented in variety of ways, and specialized chips is being developed for hardware improvement. In order to apply such neural networks to wearable devices, the compactness and the low power operation are essential. In this point of view, a suitable implementation method is a digital circuit design using field programmable gate array (FPGA). To implement this system, the learning algorithm which takes up a large part in neural networks must be implemented within FPGA for better performance. In this paper, a back propagation algorithm among various learning algorithms is implemented using FPGA, and this neural network is verified by OR gate operation. In addition, it is confirmed that this neural network can be used to analyze various users' bio signal measurement results by learning algorithm.

Back-Propagation Algorithm through Omitting Redundant Learning (중복 학습 방지에 의한 역전파 학습 알고리듬)

  • 백준호;김유신;손경식
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.29B no.9
    • /
    • pp.68-75
    • /
    • 1992
  • In this paper the back-propagation algorithm through omitting redundant learning has been proposed to improve learning speed. The proposed algorithm has been applied to XOR, Parity check and pattern recognition of hand-written numbers. The decrease of the number of patterns to be learned has been confirmed as learning proceeds even in early learning stage. The learning speed in pattern recognition of hand-written numbers is improved more than 2 times in various cases of hidden neuron numbers. It is observed that the improvement of learning speed becomes better as the number of patterns and the number of hidden numbers increase. The recognition rate of the proposed algorithm is nearly the same as that conventional method.

  • PDF