• Title/Summary/Keyword: Feature clustering

Search Result 444, Processing Time 0.027 seconds

Character Recognition of Vehicle Number Plate using Modular Neural Network (모듈라 신경망을 이용한 자동차 번호판 문자인식)

  • Park, Chang-Seok;Kim, Byeong-Man;Seo, Byung-Hoon;Lee, Kwang-Ho
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.13 no.4
    • /
    • pp.409-415
    • /
    • 2003
  • Recently, the modular learning are very popular and receive much attention for pattern classification. The modular learning method based on the "divide and conquer" strategy can not only solve the complex problems, but also reach a better result than a single classifier′s on the learning quality and speed. In the neural network area, some researches that take the modular learning approach also have been made to improve classification performance. In this paper, we propose a simple modular neural network for characters recognition of vehicle number plate and evaluate its performance on the clustering methods of feature vectors used in constructing subnetworks. We implement two clustering method, one is grouping similar feature vectors by K-means clustering algorithm, the other grouping unsimilar feature vectors by our proposed algorithm. The experiment result shows that our algorithm achieves much better performance.

A Clustering Algorithm Using the Ordered Weight of Self-Organizing Feature Maps (자기조직화 신경망의 정렬된 연결강도를 이용한 클러스터링 알고리즘)

  • Lee Jong-Sup;Kang Maing-Kyu
    • Journal of the Korean Operations Research and Management Science Society
    • /
    • v.31 no.3
    • /
    • pp.41-51
    • /
    • 2006
  • Clustering is to group similar objects into clusters. Until now there are a lot of approaches using Self-Organizing feature Maps (SOFMS) But they have problems with a small output-layer nodes and initial weight. For example, one of them is a one-dimension map of c output-layer nodes, if they want to make c clusters. This approach has problems to classify elaboratively. This Paper suggests one-dimensional output-layer nodes in SOFMs. The number of output-layer nodes is more than those of clusters intended to find and the order of output-layer nodes is ascending in the sum of the output-layer node's weight. We un find input data in SOFMs output node and classify input data in output nodes using Euclidean distance. The proposed algorithm was tested on well-known IRIS data and TSPLIB. The results of this computational study demonstrate the superiority of the proposed algorithm.

Land Cover Clustering of NDVI-drived Phenological Features

  • Kim, Dong-Keun;Suh, Myoung-Seok;Park, Kyoung-Yoon
    • Proceedings of the KSRS Conference
    • /
    • 1998.09a
    • /
    • pp.201-206
    • /
    • 1998
  • In this paper, we have considered the method for clustering land cover types over the East Asia from AVHRR data. The feature vectors such that maximum NDVI, amplitude of NDVI, mean NDVI, and NDVI threshold are extracted from the 10-day composite by maximum value composite(MVC) for reducing the effect of cloud contaninations. To find the land cover clusters given by the feature vectors, we are adapted the self-organizing feature map(SOFM) clustering which is the mapping of an input vector space of n-dimensions into a one - or two-dimensional grid of output layer. The approach is to find first the clusters by the first layer SOFM and then merge several clusters of the first layer to a large cluster by the second layer SOFM. In experiments, we were used the 8-km AVHRR data for two years(1992-1993) over the East Asia.

  • PDF

Performance Improvement of Deep Clustering Networks for Multi Dimensional Data (다차원 데이터에 대한 심층 군집 네트워크의 성능향상 방법)

  • Lee, Hyunjin
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.8
    • /
    • pp.952-959
    • /
    • 2018
  • Clustering is one of the most fundamental algorithms in machine learning. The performance of clustering is affected by the distribution of data, and when there are more data or more dimensions, the performance is degraded. For this reason, we use a stacked auto encoder, one of the deep learning algorithms, to reduce the dimension of data which generate a feature vector that best represents the input data. We use k-means, which is a famous algorithm, as a clustering. Sine the feature vector which reduced dimensions are also multi dimensional, we use the Euclidean distance as well as the cosine similarity to increase the performance which calculating the similarity between the center of the cluster and the data as a vector. A deep clustering networks combining a stacked auto encoder and k-means re-trains the networks when the k-means result changes. When re-training the networks, the loss function of the stacked auto encoder and the loss function of the k-means are combined to improve the performance and the stability of the network. Experiments of benchmark image ad document dataset empirically validated the power of the proposed algorithm.

A Hybrid Recommendation System based on Fuzzy C-Means Clustering and Supervised Learning

  • Duan, Li;Wang, Weiping;Han, Baijing
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.7
    • /
    • pp.2399-2413
    • /
    • 2021
  • A recommendation system is an information filter tool, which uses the ratings and reviews of users to generate a personalized recommendation service for users. However, the cold-start problem of users and items is still a major research hotspot on service recommendations. To address this challenge, this paper proposes a high-efficient hybrid recommendation system based on Fuzzy C-Means (FCM) clustering and supervised learning models. The proposed recommendation method includes two aspects: on the one hand, FCM clustering technique has been applied to the item-based collaborative filtering framework to solve the cold start problem; on the other hand, the content information is integrated into the collaborative filtering. The algorithm constructs the user and item membership degree feature vector, and adopts the data representation form of the scoring matrix to the supervised learning algorithm, as well as by combining the subjective membership degree feature vector and the objective membership degree feature vector in a linear combination, the prediction accuracy is significantly improved on the public datasets with different sparsity. The efficiency of the proposed system is illustrated by conducting several experiments on MovieLens dataset.

Clustering Effects in Sparse NMF(Non-negative Matrix Factorization) (Sparse NMF에 의한 클러스터링)

  • Oh, Sang-Hoon
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2008.05a
    • /
    • pp.92-95
    • /
    • 2008
  • NMF(Non-negative Matrix Factorization) has been proposed as an useful algorithm for feature extraction. Using NMF, we can extract low-dimensional feature vectors. Also, we can find clustering effects in the NMF algorithm. Also, it is reported that the sparse NMF algorithm shows better clustering effects. This paper compares the two approaches in the viewpoint of clustering effects.

  • PDF

Online nonparametric Bayesian analysis of parsimonious Gaussian mixture models and scenes clustering

  • Zhou, Ri-Gui;Wang, Wei
    • ETRI Journal
    • /
    • v.43 no.1
    • /
    • pp.74-81
    • /
    • 2021
  • The mixture model is a very powerful and flexible tool in clustering analysis. Based on the Dirichlet process and parsimonious Gaussian distribution, we propose a new nonparametric mixture framework for solving challenging clustering problems. Meanwhile, the inference of the model depends on the efficient online variational Bayesian approach, which enhances the information exchange between the whole and the part to a certain extent and applies to scalable datasets. The experiments on the scene database indicate that the novel clustering framework, when combined with a convolutional neural network for feature extraction, has meaningful advantages over other models.

The extension of the largest generalized-eigenvalue based distance metric Dij1) in arbitrary feature spaces to classify composite data points

  • Daoud, Mosaab
    • Genomics & Informatics
    • /
    • v.17 no.4
    • /
    • pp.39.1-39.20
    • /
    • 2019
  • Analyzing patterns in data points embedded in linear and non-linear feature spaces is considered as one of the common research problems among different research areas, for example: data mining, machine learning, pattern recognition, and multivariate analysis. In this paper, data points are heterogeneous sets of biosequences (composite data points). A composite data point is a set of ordinary data points (e.g., set of feature vectors). We theoretically extend the derivation of the largest generalized eigenvalue-based distance metric Dij1) in any linear and non-linear feature spaces. We prove that Dij1) is a metric under any linear and non-linear feature transformation function. We show the sufficiency and efficiency of using the decision rule $\bar{{\delta}}_{{\Xi}i}$(i.e., mean of Dij1)) in classification of heterogeneous sets of biosequences compared with the decision rules min𝚵iand median𝚵i. We analyze the impact of linear and non-linear transformation functions on classifying/clustering collections of heterogeneous sets of biosequences. The impact of the length of a sequence in a heterogeneous sequence-set generated by simulation on the classification and clustering results in linear and non-linear feature spaces is empirically shown in this paper. We propose a new concept: the limiting dispersion map of the existing clusters in heterogeneous sets of biosequences embedded in linear and nonlinear feature spaces, which is based on the limiting distribution of nucleotide compositions estimated from real data sets. Finally, the empirical conclusions and the scientific evidences are deduced from the experiments to support the theoretical side stated in this paper.

EXTRACTING INSIGHTS OF CLASSIFICATION FOR TURING PATTERN WITH FEATURE ENGINEERING

  • OH, SEOYOUNG;LEE, SEUNGGYU
    • Journal of the Korean Society for Industrial and Applied Mathematics
    • /
    • v.24 no.3
    • /
    • pp.321-330
    • /
    • 2020
  • Data classification and clustering is one of the most common applications of the machine learning. In this paper, we aim to provide the insight of the classification for Turing pattern image, which has high nonlinearity, with feature engineering using the machine learning without a multi-layered algorithm. For a given image data X whose fixel values are defined in [-1, 1], X - X3 and ∇X would be more meaningful feature than X to represent the interface and bulk region for a complex pattern image data. Therefore, we use X - X3 and ∇X in the neural network and clustering algorithm to classification. The results validate the feasibility of the proposed approach.

Feature-Point Extraction by Dynamic Linking Model bas Wavelets and Fuzzy C-Means Clustering Algorithm (Gabor 웨이브렛과 FCM 군집화 알고리즘에 기반한 동적 연결모형에 의한 얼굴표정에서 특징점 추출)

  • 신영숙
    • Korean Journal of Cognitive Science
    • /
    • v.14 no.1
    • /
    • pp.11-16
    • /
    • 2003
  • This Paper extracts the edge of main components of face with Gator wavelets transformation in facial expression images. FCM(Fuzzy C-Means) clustering algorithm then extracts the representative feature points of low dimensionality from the edge extracted in neutral face. The feature-points of the neutral face is used as a template to extract the feature-points of facial expression images. To match point to Point feature points on an expression face against each feature point on a neutral face, it consists of two steps using a dynamic linking model, which are called the coarse mapping and the fine mapping. This paper presents an automatic extraction of feature-points by dynamic linking model based on Gabor wavelets and fuzzy C-means(FCM) algorithm. The result of this study was applied to extract features automatically in facial expression recognition based on dimension[1].

  • PDF