• Title/Summary/Keyword: expectation maximization

Search Result 221, Processing Time 0.02 seconds

A Content-Based Image Retrieval Technique Using the Shape and Color Features of Objects (객체의 모양과 색상특징을 이용한 내용기반 영상검색 기법)

  • 박종현;박순영;오일환
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.10B
    • /
    • pp.1902-1911
    • /
    • 1999
  • In this paper we present a content-based image retrieval algorithm using the visual feature vectors which describe the spatial characteristics of objects. The proposed technique uses the Gaussian mixture model(GMM) to represent multi-colored objects and the expectation maximization(EM) algorithm is employed to estimate the maximum likelihood(ML) parameters of the model. After image segmentation is performed based on GMM, the shape and color features are extracted from each object using Fourier descriptors and color histograms, respectively. Image retrieval consists of two steps: first, the shape-based query is carried out to find the candidate images whose objects have the similar shapes with the query image and second, the color-based query is followed. The experimental results show that the proposed algorithm is effective in image retrieving by using the spatial and visual features of segmented objects.

  • PDF

Speech/Mixed Content Signal Classification Based on GMM Using MFCC (MFCC를 이용한 GMM 기반의 음성/혼합 신호 분류)

  • Kim, Ji-Eun;Lee, In-Sung
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.2
    • /
    • pp.185-192
    • /
    • 2013
  • In this paper, proposed to improve the performance of speech and mixed content signal classification using MFCC based on GMM probability model used for the MPEG USAC(Unified Speech and Audio Coding) standard. For effective pattern recognition, the Gaussian mixture model (GMM) probability model is used. For the optimal GMM parameter extraction, we use the expectation maximization (EM) algorithm. The proposed classification algorithm is divided into two significant parts. The first one extracts the optimal parameters for the GMM. The second distinguishes between speech and mixed content signals using MFCC feature parameters. The performance of the proposed classification algorithm shows better results compared to the conventionally implemented USAC scheme.

Controlling Linkage Disequilibrium in Association Tests: Revisiting APOE Association in Alzheimer's Disease

  • Park, Lee-Young
    • Genomics & Informatics
    • /
    • v.5 no.2
    • /
    • pp.61-67
    • /
    • 2007
  • The allele frequencies of markers as well as linkage disequilibrium (LD) can be changed in cases due to the LD between markers and the disease allele, exhibiting spurious associations of markers. To identify the true association, classical statistical tests for dealing with confounders have been applied to draw a conclusion as to whether the association of variants comes from LD with the known disease allele. However, a more direct test considering LD using estimated haplotype frequencies may be more efficient. The null hypothesis is that the different allele frequencies of a variant between cases and controls come solely from the increased disease allele frequency and the LD relationship with the disease allele. The haplotype frequencies of controls are estimated using the expectation maximization (EM) algorithm from the genotype data. The estimated frequencies are applied to calculate the expected haplotype frequencies in cases corresponding to the increase or decrease of the causative or protective alleles. The suggested method was applied to previously published data, and several APOE variants showed association with Alzheimer's disease independent from the APOE ${\varepsilon}4$ variant, rs429358, regardless of LD showing significant simulated p-values. The test results support the possibility that there may be more than one common disease variant in a locus.

A Study on Emotion Recognition Systems based on the Probabilistic Relational Model Between Facial Expressions and Physiological Responses (생리적 내재반응 및 얼굴표정 간 확률 관계 모델 기반의 감정인식 시스템에 관한 연구)

  • Ko, Kwang-Eun;Sim, Kwee-Bo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.19 no.6
    • /
    • pp.513-519
    • /
    • 2013
  • The current vision-based approaches for emotion recognition, such as facial expression analysis, have many technical limitations in real circumstances, and are not suitable for applications that use them solely in practical environments. In this paper, we propose an approach for emotion recognition by combining extrinsic representations and intrinsic activities among the natural responses of humans which are given specific imuli for inducing emotional states. The intrinsic activities can be used to compensate the uncertainty of extrinsic representations of emotional states. This combination is done by using PRMs (Probabilistic Relational Models) which are extent version of bayesian networks and are learned by greedy-search algorithms and expectation-maximization algorithms. Previous research of facial expression-related extrinsic emotion features and physiological signal-based intrinsic emotion features are combined into the attributes of the PRMs in the emotion recognition domain. The maximum likelihood estimation with the given dependency structure and estimated parameter set is used to classify the label of the target emotional states.

Pattern Classification of Hard Disk Defect Distribution Using Gaussian Mixture Model (가우시안 혼합 모델을 이용한 하드 디스크 결함 분포의 패턴 분류)

  • Jun, Jae-Young;Kim, Jeong-Heon;Moon, Un-Chul;Choi, Kwang-Nam
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2008.06c
    • /
    • pp.482-486
    • /
    • 2008
  • 본 논문에서는 하드 디스크 드라이브(Hard Disk Drive, HDD) 생산 공정 과정에서 발생할 수 있는 불량 HDD의 결함 분포에 대해서 패턴을 자동으로 분류해주는 기법을 제시한다. 이를 위해서 표준 패턴 클래스로 분류되어 있는 불량 HDD의 각 클래스의 확률 모델을 GMM(Gaussian Mixture Model)로 가정한다. 실험은 전문가에 의해 분류된 실제 HDD 결함 분포로부터 5가지의 특징 값들을 추출한 후, 결함 분포의 클래스를 표현할 수 있는 GMM의 파라미터(Parameter)를 학습한다. 각 모델의 파라미터를 추정하기 위해 EM(Expectation Maximization) 알고리즘을 사용한다. 학습된 GMM의 분류 테스트는 학습에 사용되지 않은 HDD 결함 분포에서 5가지의 특징 값을 입력 값으로 추정된 모델들의 파라미터 값에 의해 사후 확률을 구한다. 계산된 확률 값 중 가장 큰 값을 갖는 모델의 클래스를 표준 패턴 클래스로 분류한다. 그 결과 제시된 GMM을 이용한 HDD의 패턴 분류의 결과 96.1%의 정답률을 보여준다.

  • PDF

Semi-Supervised Recursive Learning of Discriminative Mixture Models for Time-Series Classification

  • Kim, Minyoung
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.13 no.3
    • /
    • pp.186-199
    • /
    • 2013
  • We pose pattern classification as a density estimation problem where we consider mixtures of generative models under partially labeled data setups. Unlike traditional approaches that estimate density everywhere in data space, we focus on the density along the decision boundary that can yield more discriminative models with superior classification performance. We extend our earlier work on the recursive estimation method for discriminative mixture models to semi-supervised learning setups where some of the data points lack class labels. Our model exploits the mixture structure in the functional gradient framework: it searches for the base mixture component model in a greedy fashion, maximizing the conditional class likelihoods for the labeled data and at the same time minimizing the uncertainty of class label prediction for unlabeled data points. The objective can be effectively imposed as individual mixture component learning on weighted data, hence our mixture learning typically becomes highly efficient for popular base generative models like Gaussians or hidden Markov models. Moreover, apart from the expectation-maximization algorithm, the proposed recursive estimation has several advantages including the lack of need for a pre-determined mixture order and robustness to the choice of initial parameters. We demonstrate the benefits of the proposed approach on a comprehensive set of evaluations consisting of diverse time-series classification problems in semi-supervised scenarios.

The inference and estimation for latent discrete outcomes with a small sample

  • Choi, Hyung;Chung, Hwan
    • Communications for Statistical Applications and Methods
    • /
    • v.23 no.2
    • /
    • pp.131-146
    • /
    • 2016
  • In research on behavioral studies, significant attention has been paid to the stage-sequential process for longitudinal data. Latent class profile analysis (LCPA) is an useful method to study sequential patterns of the behavioral development by the two-step identification process: identifying a small number of latent classes at each measurement occasion and two or more homogeneous subgroups in which individuals exhibit a similar sequence of latent class membership over time. Maximum likelihood (ML) estimates for LCPA are easily obtained by expectation-maximization (EM) algorithm, and Bayesian inference can be implemented via Markov chain Monte Carlo (MCMC). However, unusual properties in the likelihood of LCPA can cause difficulties in ML and Bayesian inference as well as estimation in small samples. This article describes and addresses erratic problems that involve conventional ML and Bayesian estimates for LCPA with small samples. We argue that these problems can be alleviated with a small amount of prior input. This study evaluates the performance of likelihood and MCMC-based estimates with the proposed prior in drawing inference over repeated sampling. Our simulation shows that estimates from the proposed methods perform better than those from the conventional ML and Bayesian method.

Tsunami-induced Change Detection Using SAR Intensity and Texture Information Based on the Generalized Gaussian Mixture Model

  • Jung, Min-young;Kim, Yong-il
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.34 no.2
    • /
    • pp.195-206
    • /
    • 2016
  • The remote sensing technique using SAR data have many advantages when applied to the disaster site due to its wide coverage and all-weather acquisition availability. Although a single-pol (polarimetric) SAR image cannot represent the land surface better than a quad-pol SAR image can, single-pol SAR data are worth using for disaster-induced change detection. In this paper, an automatic change detection method based on a mixture of GGDs (generalized Gaussian distribution) is proposed, and usability of the textural features and intensity is evaluated by using the proposed method. Three ALOS/PALSAR images were used in the experiments, and the study site was Norita City, which was affected by the 2011 Tohoku earthquake. The experiment results showed that the proposed automatic change detection method is practical for disaster sites where the large areas change. The intensity information is useful for detecting disaster-induced changes with a 68.3% g-mean, but the texture information is not. The autocorrelation and correlation show the interesting implication that they tend not to extract agricultural areas in the change detection map. Therefore, the final tsunami-induced change map is produced by the combination of three maps: one is derived from the intensity information and used as an initial map, and the others are derived from the textural information and used as auxiliary data.

Reducing Decoding Complexity by Improving Motion Field Using Bicubic and Lanczos Interpolation Techniques in Wyner-Ziv Video Coding

  • Widyantara, I Made O.;Wirawan, Wirawan;Hendrantoro, Gamantyo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.6 no.9
    • /
    • pp.2351-2369
    • /
    • 2012
  • This paper describes interpolation method of motion field in the Wyner-Ziv video coding (WZVC) based on Expectation-Maximization (EM) algorithm. In the EM algorithm, the estimated motion field distribution is calculated on a block-by-block basis. Each pixel in the block shares similar probability distribution, producing an undesired blocking artefact on the pixel-based motion field. The proposed interpolation techniques are Bicubic and Lanczos which successively use 16 and 32 neighborhood probability distributions of block-based motion field for one pixel in k-by-k block on pixel-based motion field. EM-based WZVC codec updates the estimated probability distribution on block-based motion field, and interpolates it to pixel resolution. This is required to generate higher-quality soft side information (SI) such that the decoding algorithm is able to make syndrome estimation more quickly. Our experiments showed that the proposed interpolation methods have the capability to reduce EM-based WZVC decoding complexity with small increment of bit rate.

Segmenting Inpatients by Mixture Model and Analytical Hierarchical Process(AHP) Approach In Medical Service (의료서비스에서 혼합모형(Mixture model) 및 분석적 계층과정(AHP)를 이용한 입원환자의 시장세분화에 관한 연구)

  • 백수경;곽영식
    • Health Policy and Management
    • /
    • v.12 no.2
    • /
    • pp.1-22
    • /
    • 2002
  • Since the early 1980s scholars have applied latent structure and other type of finite mixture models from various academic fields. Although the merits of finite mixture model are well documented, the attempt to apply the mixture model to medical service has been relatively rare. The researchers aim to try to fill this gap by introducing finite mixture model and segmenting inpatients DB from one general hospital. In section 2 finite mixture models are compared with clustering, chi-square analysis, and discriminant analysis based on Wedel and Kamakura(2000)'s segmentation methodology schemata. The mixture model shows the optimal segments number and fuzzy classification for each observation by EM(expectation-maximization algorism). The finite mixture model is to unfix the sample, to Identify the groups, and to estimate the parameters of the density function underlying the observed data within each group. In section 3 and 4 we illustrate results of segmenting 4510 patients data including menial and ratio scales. And then, we show AHP can be identify the attractiveness of each segment, in which the decision maker can select the best target segment.