• Title/Summary/Keyword: hebbian learning

Search Result 23, Processing Time 0.026 seconds

Long-term Synaptic Plasticity: Circuit Perturbation and Stabilization

  • Park, Joo Min;Jung, Sung-Cherl;Eun, Su-Yong
    • The Korean Journal of Physiology and Pharmacology
    • /
    • v.18 no.6
    • /
    • pp.457-460
    • /
    • 2014
  • At central synapses, activity-dependent synaptic plasticity has a crucial role in information processing, storage, learning, and memory under both physiological and pathological conditions. One widely accepted model of learning mechanism and information processing in the brain is Hebbian Plasticity: long-term potentiation (LTP) and long-term depression (LTD). LTP and LTD are respectively activity-dependent enhancement and reduction in the efficacy of the synapses, which are rapid and synapse-specific processes. A number of recent studies have a strong focal point on the critical importance of another distinct form of synaptic plasticity, non-Hebbian plasticity. Non-Hebbian plasticity dynamically adjusts synaptic strength to maintain stability. This process may be very slow and occur cell-widely. By putting them all together, this mini review defines an important conceptual difference between Hebbian and non-Hebbian plasticity.

An Alternative Method of Regression: Robust Modified Anti-Hebbian Learning

  • Hwang, Chang-Ha
    • Journal of the Korean Data and Information Science Society
    • /
    • v.7 no.2
    • /
    • pp.203-210
    • /
    • 1996
  • A linear neural unit with a modified anti-Hebbian learning rule has been shown to be able to optimally fit curves, surfaces, and hypersurfaces by adaptively extracting the minor component of the input data set. In this paper, we study how to use the robust version of this neural fitting method for linear regression analysis. Furthermore, we compare this method with other methods when data set is contaminated by outliers.

  • PDF

A Study on the Implementation of Modified Hybrid Learning Rule (변형하이브리드 학습규칙의 구현에 관한 연구)

  • 송도선;김석동;이행세
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.31B no.12
    • /
    • pp.116-123
    • /
    • 1994
  • A modified Hybrid learning rule(MHLR) is proposed, which is derived from combining the Back Propagation algorithm that is known as an excellent classifier with modified Hebbian by changing the orginal Hebbian which is a good feature extractor. The network architecture of MHLR is multi-layered neural network. The weights of MHLR are calculated from sum of the weight of BP and the weight of modified Hebbian between input layer and higgen layer and from the weight of BP between gidden layer and output layer. To evaluate the performance, BP, MHLR and the proposed Hybrid learning rule (HLR) are simulated by Monte Carlo method. As the result, MHLR is the best in recognition rate and HLR is the second. In learning speed, HLR and MHLR are much the same, while BP is relatively slow.

  • PDF

VLSI Implementation of Hopfield Network using Correlation (상관관계를 이용한 홉필드 네트웍의 VLSI 구현)

  • O, Jay-Hyouk;Park, Seong-Beom;Lee, Chong-Ho
    • Proceedings of the KIEE Conference
    • /
    • 1993.07a
    • /
    • pp.254-257
    • /
    • 1993
  • This paper presents a new method to implement Hebbian learning method on artificial neural network. In hebbian learning algorithm, complexity in terms of multiplications is high. To save the chip area, we consider a new learning circuit. By calculating similarity, or correlation between $X_i$ and $O_i$, large portion of circuits commonly used in conventional neural networks is not necessary for this new hebbian learning circuit named COR. The output signals of COR is applied to weight storage capacitors for direct control the voltages of the capacitors. The weighted sum, ${\Sigma}W_{ij}O_j$, is realized by multipliers, whose output currents are summed up in one line which goes to learning circuit or output circuit. The drain current of the multiplier can produce positive or negative synaptic weights. The pass transistor selects eight learning mode or recall mode. The layout of an learnable six-neuron fully connected Hopfield neural network is designed, and is simulated using PSPICE. The network memorizes, and retrieves the patterns correctly under the existence of minor noises.

  • PDF

A Robust Principal Component Neural Network

  • Changha Hwang;Park, Hyejung;A, Eunyoung-N
    • Communications for Statistical Applications and Methods
    • /
    • v.8 no.3
    • /
    • pp.625-632
    • /
    • 2001
  • Principal component analysis(PCA) is a multivariate technique falling under the general title of factor analysis. The purpose of PCA is to Identify the dependence structure behind a multivariate stochastic observation In order to obtain a compact description of it. In engineering field PCA is utilized mainly (or data compression and restoration. In this paper we propose a new robust Hebbian algorithm for robust PCA. This algorithm is based on a hyperbolic tangent function due to Hampel ef al.(1989) which is known to be robust in Statistics. We do two experiments to investigate the performance of the new robust Hebbian learning algorithm for robust PCA.

  • PDF

A study for improvement of Recognition velocity of Korean Character using Neural Oscillator (신경 진동자를 이용한 한글 문자의 인식 속도의 개선에 관한 연구)

  • Kwon, Yong-Bum;Lee, Joon-Tark
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2004.04a
    • /
    • pp.491-494
    • /
    • 2004
  • Neural Oscillator can be applied to oscillatory systems such as the image recognition, the voice recognition, estimate of the weather fluctuation and analysis of geological fluctuation etc in nature and principally, it is used often to pattern recoglition of image information. Conventional BPL(Back-Propagation Learning) and MLNN(Multi Layer Neural Network) are not proper for oscillatory systems because these algorithm complicate Learning structure, have tedious procedures and sluggish convergence problem. However, these problems can be easily solved by using a synchrony characteristic of neural oscillator with PLL(phase-Locked Loop) function and by using a simple Hebbian learning rule. And also, Recognition velocity of Korean Character can be improved by using a Neural Oscillator's learning accelerator factor η$\_$ij/

  • PDF

Multi-layer Neural Network with Hybrid Learning Rules for Improved Robust Capability (Robustness를 형성시키기 위한 Hybrid 학습법칙을 갖는 다층구조 신경회로망)

  • 정동규;이수영
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.31B no.8
    • /
    • pp.211-218
    • /
    • 1994
  • In this paper we develope a hybrid learning rule to improve the robustness of multi-layer Perceptions. In most neural networks the activation of a neuron is deternined by a nonlinear transformation of the weighted sum of inputs to the neurons. Investigating the behaviour of activations of hidden layer neurons a new learning algorithm is developed for improved robustness for multi-layer Perceptrons. Unlike other methods which reduce the network complexity by putting restrictions on synaptic weights our method based on error-backpropagation increases the complexity of the underlying proplem by imposing it saturation requirement on hidden layer neurons. We also found that the additional gradient-descent term for the requirement corresponds to the Hebbian rule and our algorithm incorporates the Hebbian learning rule into the error back-propagation rule. Computer simulation demonstrates fast learning convergence as well as improved robustness for classification and hetero-association of patterns.

  • PDF

A Study on the Implementation of Hybrid Learning Rule for Neural Network (다층신경망에서 하이브리드 학습 규칙의 구현에 관한 연구)

  • Song, Do-Sun;Kim, Suk-Dong;Lee, Haing-Sei
    • The Journal of the Acoustical Society of Korea
    • /
    • v.13 no.4
    • /
    • pp.60-68
    • /
    • 1994
  • In this paper we propose a new Hybrid learning rule applied to multilayer feedforward neural networks, which is constructed by combining Hebbian learning rule that is a good feature extractor and Back-Propagation(BP) learning rule that is an excellent classifier. Unlike the BP rule used in multi-layer perceptron(MLP), the proposed Hybrid learning rule is used for uptate of all connection weights except for output connection weigths becase the Hebbian learning in output layer does not guarantee learning convergence. To evaluate the performance, the proposed hybrid rule is applied to classifier problems in two dimensional space and shows better performance than the one applied only by the BP rule. In terms of learning speed the proposed rule converges faster than the conventional BP. For example, the learning of the proposed Hybrid can be done in 2/10 of the iterations that are required for BP, while the recognition rate of the proposed Hybrid is improved by about $0.778\%$ at the peak.

  • PDF

Time Series Prediction Using a Multi-layer Neural Network with Low Pass Filter Characteristics (저주파 필터 특성을 갖는 다층 구조 신경망을 이용한 시계열 데이터 예측)

  • Min-Ho Lee
    • Journal of Advanced Marine Engineering and Technology
    • /
    • v.21 no.1
    • /
    • pp.66-70
    • /
    • 1997
  • In this paper a new learning algorithm for curvature smoothing and improved generalization for multi-layer neural networks is proposed. To enhance the generalization ability a constraint term of hidden neuron activations is added to the conventional output error, which gives the curvature smoothing characteristics to multi-layer neural networks. When the total cost consisted of the output error and hidden error is minimized by gradient-descent methods, the additional descent term gives not only the Hebbian learning but also the synaptic weight decay. Therefore it incorporates error back-propagation, Hebbian, and weight decay, and additional computational requirements to the standard error back-propagation is negligible. From the computer simulation of the time series prediction with Santafe competition data it is shown that the proposed learning algorithm gives much better generalization performance.

  • PDF

Recognition of the Korean Character Using Phase Synchronization Neural Oscillator

  • Lee, Joon-Tark;Kwon, Yang-Bum
    • Journal of Advanced Marine Engineering and Technology
    • /
    • v.28 no.2
    • /
    • pp.347-353
    • /
    • 2004
  • Neural oscillator can be applied to oscillator systems such as analysis of image information, voice recognition and etc, Conventional learning algorithms(Neural Network or EBPA(Error Back Propagation Algorithm)) are not proper for oscillatory systems with the complicate input patterns because of its too much complex structure. However, these problems can be easily solved by using a synchrony characteristic of neural oscillator with PLL(phase locked loop) function and a simple Hebbian learning rule, Therefore, in this paper, it will introduce an technique for Recognition of the Korean Character using Phase Synchronization Neural Oscillator and will show the result of simulation.