• Title/Summary/Keyword: nearest neighbor classification rule

Search Result 20, Processing Time 0.037 seconds

Learning Reference Vectors by the Nearest Neighbor Network (최근점 이웃망에의한 참조벡터 학습)

  • Kim Baek Sep
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.31B no.7
    • /
    • pp.170-178
    • /
    • 1994
  • The nearest neighbor classification rule is widely used because it is not only simple but the error rate is asymptotically less than twice Bayes theoretical minimum error. But the method basically use the whole training patterns as the reference vectors. so that both storage and classification time increase as the number of training patterns increases. LVQ(Learning Vector Quantization) resolved this problem by training the reference vectors instead of just storing the whole training patterns. But it is a heuristic algorithm which has no theoretic background there is no terminating condition and it requires a lot of iterations to get to meaningful result. This paper is to propose a new training method of the reference vectors. which minimize the given error function. The nearest neighbor network,the network version of the nearest neighbor classification rule is proposed. The network is funtionally identical to the nearest neighbor classification rule is proposed. The network is funtionally identical to the nearest neighbor classification rule and the reference vectors are represented by the weights between the nodes. The network is trained to minimize the error function with respect to the weights by the steepest descent method. The learning algorithm is derived and it is shown that the proposed method can adjust more reference vectors than LVQ in each iteration. Experiment showed that the proposed method requires less iterations and the error rate is smaller than that of LVQ2.

  • PDF

Detection and Classification of Bearing Flaking Defects by Using Kullback Discrimination Information (KDI)

  • Kim, Tae-Gu;Takabumi Fukuda;Hisaji Shimizu
    • International Journal of Safety
    • /
    • v.1 no.1
    • /
    • pp.28-35
    • /
    • 2002
  • Kullback Discrimination Information (KDI) is one of the pattern recognition methods. KDI defined as a measure of the mutual dissimilarity computed between two time series was studied for detection and classification of bearing flaking on outer-race and inner-races. To model the damages, the bearings in normal condition, outer-race flaking condition and inner-races flaking condition were provided. The vibration sensor was attached by the bearing housing. This produced the total 25 pieces of data each condition, and we chose the standard data and measure of distance between standard and tested data. It is difficult to detect the flaking because similar pulses come out when balls pass the defection point. The detection and classification method for inner and outer races are defected by KDI and nearest neighbor classification rule is proposed and its high performance is also shown.

Prototype-Based Classification Using Class Hyperspheres (클래스 초월구를 이용한 프로토타입 기반 분류)

  • Lee, Hyun-Jong;Hwang, Doosung
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.10
    • /
    • pp.483-488
    • /
    • 2016
  • In this paper, we propose a prototype-based classification learning by using the nearest-neighbor rule. The nearest-neighbor is applied to segment the class area of all the training data with hyperspheres, and a hypersphere must cover the data from the same class. The radius of a hypersphere is computed by the mid point of the two distances to the farthest same class point and the nearest other class point. And we transform the prototype selection problem into a set covering problem in order to determine the smallest set of prototypes that cover all the training data. The proposed prototype selection method is designed by a greedy algorithm and applicable to process a large-scale training set in parallel. The prediction rule is the nearest-neighbor rule and the new training data is the set of prototypes. In experiments, the generalization performance of the proposed method is superior to existing methods.

Performance Improvement of Nearest-neighbor Classification Learning through Prototype Selections (프로토타입 선택을 이용한 최근접 분류 학습의 성능 개선)

  • Hwang, Doo-Sung
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.49 no.2
    • /
    • pp.53-60
    • /
    • 2012
  • Nearest-neighbor classification predicts the class of an input data with the most frequent class among the near training data of the input data. Even though nearest-neighbor classification doesn't have a training stage, all of the training data are necessary in a predictive stage and the generalization performance depends on the quality of training data. Therefore, as the training data size increase, a nearest-neighbor classification requires the large amount of memory and the large computation time in prediction. In this paper, we propose a prototype selection algorithm that predicts the class of test data with the new set of prototypes which are near-boundary training data. Based on Tomek links and distance metric, the proposed algorithm selects boundary data and decides whether the selected data is added to the set of prototypes by considering classes and distance relationships. In the experiments, the number of prototypes is much smaller than the size of original training data and we takes advantages of storage reduction and fast prediction in a nearest-neighbor classification.

Hangul Recognition Using a Hierarchical Neural Network (계층구조 신경망을 이용한 한글 인식)

  • 최동혁;류성원;강현철;박규태
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.28B no.11
    • /
    • pp.852-858
    • /
    • 1991
  • An adaptive hierarchical classifier(AHCL) for Korean character recognition using a neural net is designed. This classifier has two neural nets: USACL (Unsupervised Adaptive Classifier) and SACL (Supervised Adaptive Classifier). USACL has the input layer and the output layer. The input layer and the output layer are fully connected. The nodes in the output layer are generated by the unsupervised and nearest neighbor learning rule during learning. SACL has the input layer, the hidden layer and the output layer. The input layer and the hidden layer arefully connected, and the hidden layer and the output layer are partially connected. The nodes in the SACL are generated by the supervised and nearest neighbor learning rule during learning. USACL has pre-attentive effect, which perform partial search instead of full search during SACL classification to enhance processing speed. The input of USACL and SACL is a directional edge feature with a directional receptive field. In order to test the performance of the AHCL, various multi-font printed Hangul characters are used in learning and testing, and its processing its speed and and classification rate are compared with the conventional LVQ(Learning Vector Quantizer) which has the nearest neighbor learning rule.

  • PDF

A Study on Fault Detection and Diagnosis of Gear Damages - A Comparison between Wavelet Transform Analysis and Kullback Discrimination Information - (기어의 이상검지 및 진단에 관한 연구 -Wavelet Transform해석과 KDI의 비교-)

  • Kim, Tae-Gu;Kim, Kwang-Il
    • Journal of the Korean Society of Safety
    • /
    • v.15 no.2
    • /
    • pp.1-7
    • /
    • 2000
  • This paper presents the approach involving fault detection and diagnosis of gears using pattern recognition and Wavelet transform. It describes result of the comparison between KDI (Kullback Discrimination Information) with the nearest neighbor classification rule as one of pattern recognition methods and Wavelet transform to know a way to detect and diagnosis of gear damages experimentally. To model the damages 1) Normal (no defect), 2) one tooth is worn out, 3) All teeth faces are worn out 4) One tooth is broken. The vibration sensor was attached on the bearing housing. This produced the total time history data that is 20 pieces of each condition. We chose the standard data and measure distance between standard and tested data. In Wavelet transform analysis method, the time series data of magnitude in specified frequency (rotary and mesh frequency) were earned. As a result, the monitoring system using Wavelet transform method and KDI with nearest neighbor classification rule successfully detected and classified the damages from the experimental data.

  • PDF

Prototype based Classification by Generating Multidimensional Spheres per Class Area (클래스 영역의 다차원 구 생성에 의한 프로토타입 기반 분류)

  • Shim, Seyong;Hwang, Doosung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.20 no.2
    • /
    • pp.21-28
    • /
    • 2015
  • In this paper, we propose a prototype-based classification learning by using the nearest-neighbor rule. The nearest-neighbor is applied to segment the class area of all the training data into spheres within which the data exist from the same class. Prototypes are the center of spheres and their radii are computed by the mid-point of the two distances to the farthest same class point and the nearest another class point. And we transform the prototype selection problem into a set covering problem in order to determine the smallest set of prototypes that include all the training data. The proposed prototype selection method is based on a greedy algorithm that is applicable to the training data per class. The complexity of the proposed method is not complicated and the possibility of its parallel implementation is high. The prototype-based classification learning takes up the set of prototypes and predicts the class of test data by the nearest neighbor rule. In experiments, the generalization performance of our prototype classifier is superior to those of the nearest neighbor, Bayes classifier, and another prototype classifier.

Guitar Tab Digit Recognition and Play using Prototype based Classification

  • Baek, Byung-Hyun;Lee, Hyun-Jong;Hwang, Doosung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.21 no.9
    • /
    • pp.19-25
    • /
    • 2016
  • This paper is to recognize and play tab chords from guitar musical sheets. The musical chord area of an input image is segmented by changing the image in saturation and applying the Grabcut algorithm. Based on a template matching, our approach detects tab starting sections on a segmented musical area. The virtual block method is introduced to search blanks over chord lines and extract tab fret segments, which doesn't cause the computation loss to remove tab lines. In the experimental tests, the prototype based classification outperforms Bayesian method and the nearest neighbor rule with the whole set of training data and its performance is similar to that of the support vector machine. The experimental result shows that the prediction rate is about 99.0% and the number of selected prototypes is below 3.0%.

Nearest Neighbor Based Prototype Classification Preserving Class Regions

  • Hwang, Doosung;Kim, Daewon
    • Journal of Information Processing Systems
    • /
    • v.13 no.5
    • /
    • pp.1345-1357
    • /
    • 2017
  • A prototype selection method chooses a small set of training points from a whole set of class data. As the data size increases, the selected prototypes play a significant role in covering class regions and learning a discriminate rule. This paper discusses the methods for selecting prototypes in a classification framework. We formulate a prototype selection problem into a set covering optimization problem in which the sets are composed with distance metric and predefined classes. The formulation of our problem makes us draw attention only to prototypes per class, not considering the other class points. A training point becomes a prototype by checking the number of neighbors and whether it is preselected. In this setting, we propose a greedy algorithm which chooses the most relevant points for preserving the class dominant regions. The proposed method is simple to implement, does not have parameters to adapt, and achieves better or comparable results on both artificial and real-world problems.

Plurality Rule-based Density and Correlation Coefficient-based Clustering for K-NN

  • Aung, Swe Swe;Nagayama, Itaru;Tamaki, Shiro
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.6 no.3
    • /
    • pp.183-192
    • /
    • 2017
  • k-nearest neighbor (K-NN) is a well-known classification algorithm, being feature space-based on nearest-neighbor training examples in machine learning. However, K-NN, as we know, is a lazy learning method. Therefore, if a K-NN-based system very much depends on a huge amount of history data to achieve an accurate prediction result for a particular task, it gradually faces a processing-time performance-degradation problem. We have noticed that many researchers usually contemplate only classification accuracy. But estimation speed also plays an essential role in real-time prediction systems. To compensate for this weakness, this paper proposes correlation coefficient-based clustering (CCC) aimed at upgrading the performance of K-NN by leveraging processing-time speed and plurality rule-based density (PRD) to improve estimation accuracy. For experiments, we used real datasets (on breast cancer, breast tissue, heart, and the iris) from the University of California, Irvine (UCI) machine learning repository. Moreover, real traffic data collected from Ojana Junction, Route 58, Okinawa, Japan, was also utilized to lay bare the efficiency of this method. By using these datasets, we proved better processing-time performance with the new approach by comparing it with classical K-NN. Besides, via experiments on real-world datasets, we compared the prediction accuracy of our approach with density peaks clustering based on K-NN and principal component analysis (DPC-KNN-PCA).