• Title/Summary/Keyword: Multiple Classifier Combination

Search Result 22, Processing Time 0.021 seconds

Multiple-Classifier Combination based on Image Degradation Model for Low-Quality Image Recognition (저화질 영상 인식을 위한 화질 저하 모델 기반 다중 인식기 결합)

  • Ryu, Sang-Jin;Kim, In-Jung
    • The KIPS Transactions:PartB
    • /
    • v.17B no.3
    • /
    • pp.233-238
    • /
    • 2010
  • In this paper, we propose a multiple classifier combination method based on image degradation modeling to improve recognition performance on low-quality images. Using an image degradation model, it generates a set of classifiers each of which is specialized for a specific image quality. In recognition, it combines the results of the recognizers by weighted averaging to decide the final result. At this time, the weight of each recognizer is dynamically decided from the estimated quality of the input image. It assigns large weight to the recognizer specialized to the estimated quality of the input image, but small weight to other recognizers. As the result, it can effectively adapt to image quality variation. Moreover, being a multiple-classifier system, it shows more reliable performance then the single-classifier system on low-quality images. In the experiment, the proposed multiple-classifier combination method achieved higher recognition rate than multiple-classifier combination systems not considering the image quality or single classifier systems considering the image quality.

Selecting Classifiers using Mutual Information between Classifiers (인식기 간의 상호정보를 이용한 인식기 선택)

  • Kang, Hee-Joong
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.14 no.3
    • /
    • pp.326-330
    • /
    • 2008
  • The study on combining multiple classifiers in the field of pattern recognition has mainly focused on how to combine multiple classifiers, but it has gradually turned to the study on how to select multiple classifiers from a classifier pool recently. Actually, the performance of multiple classifier system depends on the selected classifiers as well as the combination method of classifiers. Therefore, it is necessary to select a classifier set showing good performance, and an approach based on information theory has been tried to select the classifier set. In this paper, a classifier set candidate is made by the selection of classifiers, on the basis of mutual information between classifiers, and the classifier set candidate is compared with the other classifier sets chosen by the different selection methods in experiments.

A Multiple Classifier System based on Dynamic Classifier Selection having Local Property (지역적 특성을 갖는 동적 선택 방법에 기반한 다중 인식기 시스템)

  • 송혜정;김백섭
    • Journal of KIISE:Software and Applications
    • /
    • v.30 no.3_4
    • /
    • pp.339-346
    • /
    • 2003
  • This paper proposes a multiple classifier system having massive micro classifiers. The micro classifiers are trained by using a local set of training patterns. The k nearest neighboring training patterns of one training pattern comprise the local region for training a micro classifier. Each training pattern is incorporated with one or more micro classifiers. Two types of micro classifiers are adapted in this paper. SVM with linear kernel and SVM with RBF kernel. Classification is done by selecting the best micro classifier among the micro classifiers in vicinity of incoming test pattern. To measure the goodness of each micro classifier, the weighted sum of correctly classified training patterns in vicinity of the test pattern is used. Experiments have been done on Elena database. Results show that the proposed method gives better classification accuracy than any conventional classifiers like SVM, k-NN and the conventional classifier combination/selection scheme.

Fuzzy Behavior Knowledge Space for Integration of Multiple Classifiers (다중 분류기 통합을 위한 퍼지 행위지식 공간)

  • 김봉근;최형일
    • Korean Journal of Cognitive Science
    • /
    • v.6 no.2
    • /
    • pp.27-45
    • /
    • 1995
  • In this paper, we suggest the "Fuzzy Behavior Knowledge Space(FBKS)" and explain how to utilize the FBKS when aggregating decisions of individual classifiers. The concept of "Behavior Knowledge Space(BKS)" is known to be the best method in the context that each classifier offers only one class label as its decision. However. the BKS does not considers measurement value of class label. Furthermore, it does not allow the heuristic knowledge of human experts to be embedded when combining multiple decisions. The FBKS eliminates such drawbacks of the BKS by adapting the fwzy concepts. Our method applies to the classification results that contain both class labels and associated measurement values. Experimental results confirm that the FBKS could be a very promising tool in pattern recognition areas.

  • PDF

Fingerprint Classification using Multiple Decision Templates with SVM (SVM의 다중결정템플릿을 이용한 지문분류)

  • Min Jun-Ki;Hong Jin-Hyuk;Cho Sung-Bae
    • Journal of KIISE:Software and Applications
    • /
    • v.32 no.11
    • /
    • pp.1136-1146
    • /
    • 2005
  • Fingerprint classification is useful in an automated fingerprint identification system (AFIS) to reduce the matching time by categorizing fingerprints. Based on Henry system that classifies fingerprints into S classes, various techniques such as neural networks and support vector machines (SVMs) have been widely used to classify fingerprints. Especially, SVMs of high classification performance have been actively investigated. Since the SVM is binary classifier, we propose a novel classifier-combination model, multiple decision templates (MuDTs), to classily fingerprints. The method extracts several clusters of different characteristics from samples of a class and constructs a suitable combination model to overcome the restriction of the single model, which may be subject to the ambiguous images. With the experimental results of the proposed on the FingerCodes extracted from NIST Database4 for the five-class and four-class problems, we have achieved a classification accuracy of $90.4\%\;and\;94.9\%\;with\;1.8\%$ rejection, respectively.

Study on the influence of Alpha wave music on working memory based on EEG

  • Xu, Xin;Sun, Jiawen
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.2
    • /
    • pp.467-479
    • /
    • 2022
  • Working memory (WM), which plays a vital role in daily activities, is a memory system that temporarily stores and processes information when people are engaged in complex cognitive activities. The influence of music on WM has been widely studied. In this work, we conducted a series of n-back memory experiments with different task difficulties and multiple trials on 14 subjects under the condition of no music and Alpha wave leading music. The analysis of behavioral data show that the change of music condition has significant effect on the accuracy and time of memory reaction (p<0.01), both of which are improved after the stimulation of Alpha wave music. Behavioral results also suggest that short-term training has no significant impact on working memory. In the further analysis of electrophysiology (EEG) data recorded in the experiment, auto-regressive (AR) model is employed to extract features, after which an average classification accuracy of 82.9% is achieved with support vector machine (SVM) classifier in distinguishing between before and after WM enhancement. The above findings indicate that Alpha wave leading music can improve WM, and the combination of AR model and SVM classifier is effective in detecting the brain activity changes resulting from music stimulation.

Classifier Selection using Feature Space Attributes in Local Region (국부적 영역에서의 특징 공간 속성을 이용한 다중 인식기 선택)

  • Shin Dong-Kuk;Song Hye-Jeong;Kim Baeksop
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.12
    • /
    • pp.1684-1690
    • /
    • 2004
  • This paper presents a method for classifier selection that uses distribution information of the training samples in a small region surrounding a sample. The conventional DCS-LA(Dynamic Classifier Selection - Local Accuracy) selects a classifier dynamically by comparing the local accuracy of each classifier at the test time, which inevitably requires long classification time. On the other hand, in the proposed approach, the best classifier in a local region is stored in the FSA(Feature Space Attribute) table during the training time, and the test is done by just referring to the table. Therefore, this approach enables fast classification because classification is not needed during test. Two feature space attributes are used entropy and density of k training samples around each sample. Each sample in the feature space is mapped into a point in the attribute space made by two attributes. The attribute space is divided into regular rectangular cells in which the local accuracy of each classifier is appended. The cells with associated local accuracy comprise the FSA table. During test, when a test sample is applied, the cell to which the test sample belongs is determined first by calculating the two attributes, and then, the most accurate classifier is chosen from the FSA table. To show the effectiveness of the proposed algorithm, it is compared with the conventional DCS -LA using the Elena database. The experiments show that the accuracy of the proposed algorithm is almost same as DCS-LA, but the classification time is about four times faster than that.

CREATING MULTIPLE CLASSIFIERS FOR THE CLASSIFICATION OF HYPERSPECTRAL DATA;FEATURE SELECTION OR FEATURE EXTRACTION

  • Maghsoudi, Yasser;Rahimzadegan, Majid;Zoej, M.J.Valadan
    • Proceedings of the KSRS Conference
    • /
    • 2007.10a
    • /
    • pp.6-10
    • /
    • 2007
  • Classification of hyperspectral images is challenging. A very high dimensional input space requires an exponentially large amount of data to adequately and reliably represent the classes in that space. In other words in order to obtain statistically reliable classification results, the number of necessary training samples increases exponentially as the number of spectral bands increases. However, in many situations, acquisition of the large number of training samples for these high-dimensional datasets may not be so easy. This problem can be overcome by using multiple classifiers. In this paper we compared the effectiveness of two approaches for creating multiple classifiers, feature selection and feature extraction. The methods are based on generating multiple feature subsets by running feature selection or feature extraction algorithm several times, each time for discrimination of one of the classes from the rest. A maximum likelihood classifier is applied on each of the obtained feature subsets and finally a combination scheme was used to combine the outputs of individual classifiers. Experimental results show the effectiveness of feature extraction algorithm for generating multiple classifiers.

  • PDF

Web Mining Using Fuzzy Integration of Multiple Structure Adaptive Self-Organizing Maps (다중 구조적응 자기구성지도의 퍼지결합을 이용한 웹 마이닝)

  • 김경중;조성배
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.1
    • /
    • pp.61-70
    • /
    • 2004
  • It is difficult to find an appropriate web site because exponentially growing web contains millions of web documents. Personalization of web search can be realized by recommending proper web sites using user profile but more efficient method is needed for estimating preference because user's evaluation on web contents presents many aspects of his characteristics. As user profile has a property of non-linearity, estimation by classifier is needed and combination of classifiers is necessary to anticipate diverse properties. Structure adaptive self-organizing map (SASOM) that is suitable for Pattern classification and visualization is an enhanced model of SOM and might be useful for web mining. Fuzzy integral is a combination method using classifiers' relevance that is defined subjectively. In this paper, estimation of user profile is conducted by using ensemble of SASOM's teamed independently based on fuzzy integral and evaluated by Syskill & Webert UCI benchmark data. Experimental results show that the proposed method performs better than previous naive Bayes classifier as well as voting of SASOM's.

Learning Algorithm for Multiple Distribution Data using Haar-like Feature and Decision Tree (다중 분포 학습 모델을 위한 Haar-like Feature와 Decision Tree를 이용한 학습 알고리즘)

  • Kwak, Ju-Hyun;Woen, Il-Young;Lee, Chang-Hoon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.2 no.1
    • /
    • pp.43-48
    • /
    • 2013
  • Adaboost is widely used for Haar-like feature boosting algorithm in Face Detection. It shows very effective performance on single distribution model. But when detecting front and side face images at same time, Adaboost shows it's limitation on multiple distribution data because it uses linear combination of basic classifier. This paper suggest the HDCT, modified decision tree algorithm for Haar-like features. We still tested the performance of HDCT compared with Adaboost on multiple distributed image recognition.