• Title/Summary/Keyword: point dataset

Search Result 195, Processing Time 0.027 seconds

A Novel Method for Hand Posture Recognition Based on Depth Information Descriptor

  • Xu, Wenkai;Lee, Eung-Joo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.2
    • /
    • pp.763-774
    • /
    • 2015
  • Hand posture recognition has been a wide region of applications in Human Computer Interaction and Computer Vision for many years. The problem arises mainly due to the high dexterity of hand and self-occlusions created in the limited view of the camera or illumination variations. To remedy these problems, a hand posture recognition method using 3-D point cloud is proposed to explicitly utilize 3-D information from depth maps in this paper. Firstly, hand region is segmented by a set of depth threshold. Next, hand image normalization will be performed to ensure that the extracted feature descriptors are scale and rotation invariant. By robustly coding and pooling 3-D facets, the proposed descriptor can effectively represent the various hand postures. After that, SVM with Gaussian kernel function is used to address the issue of posture recognition. Experimental results based on posture dataset captured by Kinect sensor (from 1 to 10) demonstrate the effectiveness of the proposed approach and the average recognition rate of our method is over 96%.

Approximation of the Distribution Function for the Number of Innovation Activities Using a Mixture Model (기술혁신 횟수의 분포함수 추정 -혼합모형을 적용하여-)

  • Yoo Seung-Hoon;Park Doo-Ho
    • Journal of Korea Technology Innovation Society
    • /
    • v.8 no.3
    • /
    • pp.887-910
    • /
    • 2005
  • This paper attempts to approximate the distribution function for the number of innovation activities (NIA). To this end, the dataset of 2002 Korean Innovation Survey (KIS 2002) published by Science and Technology Policy Institute is used. To deal with zero NTI values given by a considerable number of firms in the KIS 2002 survey, a mixture model of distributions for NIA is applied. The NIA is specified as a mixture of two distributions, one with a point mass at zero and the other with full support on the positive half of the real line. The model was empirically verified for the KIS 2002 data. The mixture model can easily capture the common bimodality feature of the NIA distribution. In addition, when covariates were added to the mixture model, it was found that the probability that a firm has zero NIA significantly varies with some variables.

  • PDF

A comparative study in Bayesian semiparametric approach to small area estimation

  • Heo, Simyoung;Kim, Dal Ho
    • Journal of the Korean Data and Information Science Society
    • /
    • v.27 no.5
    • /
    • pp.1433-1441
    • /
    • 2016
  • Small area model provides reliable and accurate estimations when the sample size is not sufficient. Our dataset has an inherent nonlinear pattern which signicantly affects our inference. In this case, we could consider semiparametric models such as truncated polynomial basis function and radial basis function. In this paper, we study four Bayesian semiparametric models for small areas to handle this point. Four small area models are based on two kinds of basis function and different knots positions. To evaluate the different estimates, four comparison measurements have been employed as criteria. In these comparison measurements, the truncated polynomial basis function with equal quantile knots has shown the best result. In Bayesian calculation, we use Gibbs sampler to solve the numerical problems.

A Mixed Co-clustering Algorithm Based on Information Bottleneck

  • Liu, Yongli;Duan, Tianyi;Wan, Xing;Chao, Hao
    • Journal of Information Processing Systems
    • /
    • v.13 no.6
    • /
    • pp.1467-1486
    • /
    • 2017
  • Fuzzy co-clustering is sensitive to noise data. To overcome this noise sensitivity defect, possibilistic clustering relaxes the constraints in FCM-type fuzzy (co-)clustering. In this paper, we introduce a new possibilistic fuzzy co-clustering algorithm based on information bottleneck (ibPFCC). This algorithm combines fuzzy co-clustering and possibilistic clustering, and formulates an objective function which includes a distance function that employs information bottleneck theory to measure the distance between feature data point and feature cluster centroid. Many experiments were conducted on three datasets and one artificial dataset. Experimental results show that ibPFCC is better than such prominent fuzzy (co-)clustering algorithms as FCM, FCCM, RFCC and FCCI, in terms of accuracy and robustness.

Exploring an Optimal Feature Selection Method for Effective Opinion Mining Tasks

  • Eo, Kyun Sun;Lee, Kun Chang
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.2
    • /
    • pp.171-177
    • /
    • 2019
  • This paper aims to find the most effective feature selection method for the sake of opinion mining tasks. Basically, opinion mining tasks belong to sentiment analysis, which is to categorize opinions of the online texts into positive and negative from a text mining point of view. By using the five product groups dataset such as apparel, books, DVDs, electronics, and kitchen, TF-IDF and Bag-of-Words(BOW) fare calculated to form the product review feature sets. Next, we applied the feature selection methods to see which method reveals most robust results. The results show that the stacking classifier based on those features out of applying Information Gain feature selection method yields best result.

A Study on Representative Skyline Using Connected Component Clustering

  • Choi, Jong-Hyeok;Nasridinov, Aziz
    • Journal of Multimedia Information System
    • /
    • v.6 no.1
    • /
    • pp.37-42
    • /
    • 2019
  • Skyline queries are used in a variety of fields to make optimal decisions. However, as the volume of data and the dimension of the data increase, the number of skyline points increases with the amount of time it takes to discover them. Mainly, because the number of skylines is essential in many real-life applications, various studies have been proposed. However, previous researches have used the k-parameter methods such as top-k and k-means to discover representative skyline points (RSPs) from entire skyline point set, resulting in high query response time and reduced representativeness due to k dependency. To solve this problem, we propose a new Connected Component Clustering based Representative Skyline Query (3CRS) that can discover RSP quickly even in high-dimensional data through connected component clustering. 3CRS performs fast discovery and clustering of skylines through hash indexes and connected components and selects RSPs from each cluster. This paper proves the superiority of the proposed method by comparing it with representative skyline queries using k-means and DBSCAN with the real-world dataset.

Variational autoencoder for prosody-based speaker recognition

  • Starlet Ben Alex;Leena Mary
    • ETRI Journal
    • /
    • v.45 no.4
    • /
    • pp.678-689
    • /
    • 2023
  • This paper describes a novel end-to-end deep generative model-based speaker recognition system using prosodic features. The usefulness of variational autoencoders (VAE) in learning the speaker-specific prosody representations for the speaker recognition task is examined herein for the first time. The speech signal is first automatically segmented into syllable-like units using vowel onset points (VOP) and energy valleys. Prosodic features, such as the dynamics of duration, energy, and fundamental frequency (F0), are then extracted at the syllable level and used to train/adapt a speaker-dependent VAE from a universal VAE. The initial comparative studies on VAEs and traditional autoencoders (AE) suggest that the former can efficiently learn speaker representations. Investigations on the impact of gender information in speaker recognition also point out that gender-dependent impostor banks lead to higher accuracies. Finally, the evaluation on the NIST SRE 2010 dataset demonstrates the usefulness of the proposed approach for speaker recognition.

Predicting depth value of the future depth-based multivariate record

  • Samaneh Tata;Mohammad Reza Faridrohani
    • Communications for Statistical Applications and Methods
    • /
    • v.30 no.5
    • /
    • pp.453-465
    • /
    • 2023
  • The prediction problem of univariate records, though not addressed in multivariate records, has been discussed by many authors based on records values. There are various definitions for multivariate records among which depth-based records have been selected for the aim of this paper. In this paper, by means of the maximum likelihood and conditional median methods, point and interval predictions of depth values which are related to the future depth-based multivariate records are considered on the basis of the observed ones. The observations derived from some elements of the elliptical distributions are the main reason of studying this problem. Finally, the satisfactory performance of the prediction methods is illustrated via some simulation studies and a real dataset about Kermanshah city drought.

Term Frequency-Inverse Document Frequency (TF-IDF) Technique Using Principal Component Analysis (PCA) with Naive Bayes Classification

  • J.Uma;K.Prabha
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.4
    • /
    • pp.113-118
    • /
    • 2024
  • Pursuance Sentiment Analysis on Twitter is difficult then performance it's used for great review. The present be for the reason to the tweet is extremely small with mostly contain slang, emoticon, and hash tag with other tweet words. A feature extraction stands every technique concerning structure and aspect point beginning particular tweets. The subdivision in a aspect vector is an integer that has a commitment on ascribing a supposition class to a tweet. The cycle of feature extraction is to eradicate the exact quality to get better the accurateness of the classifications models. In this manuscript we proposed Term Frequency-Inverse Document Frequency (TF-IDF) method is to secure Principal Component Analysis (PCA) with Naïve Bayes Classifiers. As the classifications process, the work proposed can produce different aspects from wildly valued feature commencing a Twitter dataset.

Assessment and merging technique for GPM satellite precipitation product using ground based measurement (GPM 위성 강우자료의 검증과 지상관측 자료를 통한 강우 보정 기법)

  • Baik, Jongjin;Park, Jongmin;Kim, Kiyoung;Choi, Minha
    • Journal of Korea Water Resources Association
    • /
    • v.51 no.2
    • /
    • pp.131-140
    • /
    • 2018
  • Precipitation is a key variable to enhance the understanding of water cycle system and secure and manage the water resources efficiently. In this study, we evaluated the feasibility of GPM precipitation datasets through comparison with the 92 ASOS sites in South Korea during 2015. Additionally, three merging techniques (i.e., Geographical Differential Analysis, Geographical Ratio Analysis, Conditional Merging) were applied to improve accuracy of precipitation by fusing the advantages from point and satellite-based datasets. The results of this study are as follows. 1) GPM dataset indicated slightly overestimation with compared ASOS dataset, especially high uncertainties in summer season. 2) Validation of three merging techniques through jackniffe cross-validation showed that uncertainty were decreased as the spatial resolution increased. Especially, conditional merging showed the best performance among three methods.