• 제목/요약/키워드: Logic Neurons

검색결과 24건 처리시간 0.024초

Logic-based Fuzzy Neural Networks based on Fuzzy Granulation

  • Kwak, Keun-Chang;Kim, Dong-Hwa
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2005년도 ICCAS
    • /
    • pp.1510-1515
    • /
    • 2005
  • This paper is concerned with a Logic-based Fuzzy Neural Networks (LFNN) with the aid of fuzzy granulation. As the underlying design tool guiding the development of the proposed LFNN, we concentrate on the context-based fuzzy clustering which builds information granules in the form of linguistic contexts as well as OR fuzzy neuron which is logic-driven processing unit realizing the composition operations of T-norm and S-norm. The design process comprises several main phases such as (a) defining context fuzzy sets in the output space, (b) completing context-based fuzzy clustering in each context, (c) aggregating OR fuzzy neuron into linguistic models, and (c) optimizing connections linking information granules and fuzzy neurons in the input and output spaces. The experimental examples are tested through two-dimensional nonlinear function. The obtained results reveal that the proposed model yields better performance in comparison with conventional linguistic model and other approaches.

  • PDF

패턴인식을 위한 다층 신경망의 디지털 구현에 관한 연구 (A Study on the Digital Implementation of Multi-layered Neural Networks for Pattern Recognition)

  • 박영석
    • 융합신호처리학회논문지
    • /
    • 제2권2호
    • /
    • pp.111-118
    • /
    • 2001
  • 본 연구에서는 패턴 인식용 다층 퍼셉트론 신경망을 순수 디지털 논리회로 모델로 구현할 수 있도록 새로운 논리뉴런의 구조, 디지털 정형 다층논리신경망 구조, 그리고 패턴인식의 응용을 위한 다단 다층논리 신경망 구조를 제안하고, 또한 제안된 구조는 매우 단순하면서도 효과적인 증가적인 가법적(Incremental Additive) 학습알고리즘이 존재함을 보였다.

  • PDF

신경회로망을 이용한 조합 논리회로의 테스트 생성 (Test Generation for Combinational Logic Circuits Using Neural Networks)

  • 김영우;임인칠
    • 전자공학회논문지A
    • /
    • 제30A권9호
    • /
    • pp.71-79
    • /
    • 1993
  • This paper proposes a new test pattern generation methodology for combinational logic circuits using neural networks based on a modular structure. The CUT (Circuit Under Test) is described in our gate level hardware description language. By conferring neural database, the CUT is compiled to an ATPG (Automatic Test Pattern Generation) neural network. Each logic gate in CUT is represented as a discrete Hopfield network. Such a neual network is called a gate module in this paper. All the gate modules for a CUT form an ATPG neural network by connecting each module through message passing paths by which the states of modules are transferred to their adjacent modules. A fault is injected by setting the activation values of some neurons at given values and by invalidating connections between some gate modules. A test pattern for an injected fault is obtained when all gate modules in the ATPG neural network are stabilized through evolution and mutual interactions. The proposed methodology is efficient for test generation, known to be NP-complete, through its massive paralelism. Some results on combinational logic circuits confirm the feasibility of the proposed methodology.

  • PDF

Genetically Optimized Hybrid Fuzzy Set-based Polynomial Neural Networks with Polynomial and Fuzzy Polynomial Neurons

  • Oh Sung-Kwun;Roh Seok-Beom;Park Keon-Jun
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • 제5권4호
    • /
    • pp.327-332
    • /
    • 2005
  • We investigatea new fuzzy-neural networks-Hybrid Fuzzy set based polynomial Neural Networks (HFSPNN). These networks consist of genetically optimized multi-layer with two kinds of heterogeneous neurons thatare fuzzy set based polynomial neurons (FSPNs) and polynomial neurons (PNs). We have developed a comprehensive design methodology to determine the optimal structure of networks dynamically. The augmented genetically optimized HFSPNN (namely gHFSPNN) results in a structurally optimized structure and comes with a higher level of flexibility in comparison to the one we encounter in the conventional HFPNN. The GA-based design procedure being applied at each layer of gHFSPNN leads to the selection leads to the selection of preferred nodes (FSPNs or PNs) available within the HFSPNN. In the sequel, the structural optimization is realized via GAs, whereas the ensuing detailed parametric optimization is carried out in the setting of a standard least square method-based learning. The performance of the gHFSPNN is quantified through experimentation where we use a number of modeling benchmarks synthetic and experimental data already experimented with in fuzzy or neurofuzzy modeling.

MLP분류법을 적용한 가스분류기능의 칩 설계 및 응용 (Chip design and application of gas classification function using MLP classification method)

  • 장으뜸;서용수;정완영
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2001년도 하계종합학술대회 논문집(2)
    • /
    • pp.309-312
    • /
    • 2001
  • A primitive gas classification system which can classify limited species of gas was designed and simulated. The 'electronic nose' consists of an array of 4 metal oxide gas sensors with different selectivity patterns, signal collecting unit and a signal pattern recognition and decision Part in PLD(programmable logic device) chip. Sensor array consists of four commercial, tin oxide based, semiconductor type gas sensors. BP(back propagation) neutral networks with MLP(Multilayer Perceptron) structure was designed and implemented on CPLD of fifty thousand gate level chip by VHDL language for processing the input signals from 4 gas sensors and qualification of gases in air. The network contained four input units, one hidden layer with 4 neurons and output with 4 regular neurons. The 'electronic nose' system was successfully classified 4 kinds of industrial gases in computer simulation.

  • PDF

대량의 병렬성을 이용한 고속 자동 테스트 패턴 생성기 (A Fast Automatic Test Pattern Generator Using Massive Parallelism)

  • 김영오;임인칠
    • 전자공학회논문지B
    • /
    • 제32B권5호
    • /
    • pp.661-670
    • /
    • 1995
  • This paper presents a fast massively parallel automatic test pattern generator for digital combinational logic circuits using neural networks. Automatic test pattern generation neural network(ATPGNN) evolves its state to a stable local minima by exchanging messages among neural network modules. In preprocessing phase, we calculate the essential assignments for the stuck-at faults in fault list by adopting dominator concept. It makes more neurons be fixed and the system speed up. Consequently. fast test pattern generation is achieved. Test patterns for stuck-open faults are generated through getting initialization patterns for the obtained stuck-at faults in the corresponding ATPGNN.

  • PDF

합 기반의 전건부를 가지는 뉴로-퍼지 시스템 설계 (Design of a Neuro-Fuzzy System Using Union-Based Rule Antecedent)

  • 한창욱;이돈규
    • 정보처리학회 논문지
    • /
    • 제13권2호
    • /
    • pp.13-17
    • /
    • 2024
  • 본 논문에서는 규칙의 수를 줄여 간결한 지식 기반을 보장할 수 있는 합 기반의 전건부를 가지는 뉴로-퍼지 제어기를 제안하였다. 제안된 뉴로-퍼지 제어기는 모든 입력 변수의 AND 조합을 전건부로 하는 구조의 퍼지 규칙보다 더 큰 입력 영역을 커버하기 위해 전건부에 입력 퍼지 집합의 합집합 연산을 허용하였다. 이러한 뉴로-퍼지 제어기를 구성하기 위해 본 논문에서는 OR 및 AND 퍼지 뉴런으로 구성된 multiple-term unified logic processor (MULP)를 고려하였다. 이러한 OR 및 AND 퍼지 뉴런은 조정 가능한 연결 강도 집합을 가지므로 학습을 통하여 최적의 연결 강도 집합을 찾을 수 있다. 초기 최적화 단계에서 유전 알고리즘은 제안된 뉴로 퍼지 제어기의 최적화된 이진 구조를 구성하고, 이후 확률에 기반한 강화 학습은 성능 지수를 더욱 향상시켜서 유전 알고리즘에 의해 최적화된 제어기의 이진 연결을 개선하였다. 역진자 시스템을 제어하기 위한 모의실험 및 실험을 통해 제안된 방법의 유효성을 검증하였다.

Neural Network Training Using a GMDH Type Algorithm

  • Pandya, Abhijit S.;Gilbar, Thomas;Kim, Kwang-Baek
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • 제5권1호
    • /
    • pp.52-58
    • /
    • 2005
  • We have developed a Group Method of Data Handling (GMDH) type algorithm for designing multi-layered neural networks. The algorithm is general enough that it will accept any number of inputs and any sized training set. Each neuron of the resulting network is a function of two of the inputs to the layer. The equation for each of the neurons is a quadratic polynomial. Several forms of the equation are tested for each neuron to make sure that only the best equation of two inputs is kept. All possible combinations of two inputs to each layer are also tested. By carefully testing each resulting neuron, we have developed an algorithm to keep only the best neurons at each level. The algorithm's goal is to create as accurate a network as possible while minimizing the size of the network. Software was developed to train and simulate networks using our algorithm. Several applications were modeled using our software, and the result was that our algorithm succeeded in developing small, accurate, multi-layer networks.

A neuron computer model embedded Lukasiewicz' implication

  • Kobata, Kenji;Zhu, Hanxi;Aoyama, Tomoo;Yoshihara, Ikuo
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2000년도 제15차 학술회의논문집
    • /
    • pp.449-449
    • /
    • 2000
  • Many researchers have studied architectures for non-Neumann's computers because of escaping its bottleneck. To avoid the bottleneck, a neuron-based computer has been developed. The computer has only neurons and their connections, which are constructed of the learning. But still it has information processing facilities, and at the same time, it is like as a simplified brain to make inference; it is called "neuron-computer". No instructions are considered in any neural network usually; however, to complete complex processing on restricted computing resources, the processing must be reduced to primitive actions. Therefore, we introduce the instructions to the neuron-computer, in which the most important function is implications. There is an implication represented by binary-operators, but general implications for multi-value or fuzzy logics can't be done. Therefore, we need to use Lukasiewicz' operator at least. We investigated a neuron-computer having instructions for general implications. If we use the computer, the effective inferences base on multi-value logic is executed rapidly in a small logical unit.

  • PDF

Pseudoinverse Matrix Decomposition Based Incremental Extreme Learning Machine with Growth of Hidden Nodes

  • Kassani, Peyman Hosseinzadeh;Kim, Euntai
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • 제16권2호
    • /
    • pp.125-130
    • /
    • 2016
  • The proposal of this study is a fast version of the conventional extreme learning machine (ELM), called pseudoinverse matrix decomposition based incremental ELM (PDI-ELM). One of the main problems in ELM is to determine the number of hidden nodes. In this study, the number of hidden nodes is automatically determined. The proposed model is an incremental version of ELM which adds neurons with the goal of minimization the error of the ELM network. To speed up the model the information of pseudoinverse from previous step is taken into account in the current iteration. To show the ability of the PDI-ELM, it is applied to few benchmark classification datasets in the University of California Irvine (UCI) repository. Compared to ELM learner and two other versions of incremental ELM, the proposed PDI-ELM is faster.