• Title/Summary/Keyword: Frequency Partition

Search Result 58, Processing Time 0.028 seconds

Atrial Fibrillation Pattern Analysis based on Symbolization and Information Entropy (부호화와 정보 엔트로피에 기반한 심방세동 (Atrial Fibrillation: AF) 패턴 분석)

  • Cho, Ik-Sung;Kwon, Hyeog-Soong
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.16 no.5
    • /
    • pp.1047-1054
    • /
    • 2012
  • Atrial fibrillation (AF) is the most common arrhythmia encountered in clinical practice, and its risk increases with age. Conventionally, the way of detecting AF was the time·frequency domain analysis of RR variability. However, the detection of ECG signal is difficult because of the low amplitude of the P wave and the corruption by the noise. Also, the time·frequency domain analysis of RR variability has disadvantage to get the details of irregular RR interval rhythm. In this study, we describe an atrial fibrillation pattern analysis based on symbolization and information entropy. We transformed RR interval data into symbolic sequence through differential partition, analyzed RR interval pattern, quantified the complexity through Shannon entropy and detected atrial fibrillation. The detection algorithm was tested using the threshold between 10ms and 100ms on two databases, namely the MIT-BIH Atrial Fibrillation Database.

Fuzzy discretization with spatial distribution of data and Its application to feature selection (데이터의 공간적 분포를 고려한 퍼지 이산화와 특징선택에의 응용)

  • Son, Chang-Sik;Shin, A-Mi;Lee, In-Hee;Park, Hee-Joon;Park, Hyoung-Seob;Kim, Yoon-Nyun
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.20 no.2
    • /
    • pp.165-172
    • /
    • 2010
  • In clinical data minig, choosing the optimal subset of features is such important, not only to reduce the computational complexity but also to improve the usefulness of the model constructed from the given data. Moreover the threshold values (i.e., cut-off points) of selected features are used in a clinical decision criteria of experts for differential diagnosis of diseases. In this paper, we propose a fuzzy discretization approach, which is evaluated by measuring the degree of separation of redundant attribute values in overlapping region, based on spatial distribution of data with continuous attributes. The weighted average of the redundant attribute values is then used to determine the threshold value for each feature and rough set theory is utilized to select a subset of relevant features from the overall features. To verify the validity of the proposed method, we compared experimental results, which applied to classification problem using 668 patients with a chief complaint of dyspnea, based on three discretization methods (i.e., equal-width, equal-frequency, and entropy-based) and proposed discretization method. From the experimental results, we confirm that the discretization methods with fuzzy partition give better results in two evaluation measures, average classification accuracy and G-mean, than those with hard partition.

A Big Data Analysis by Between-Cluster Information using k-Modes Clustering Algorithm (k-Modes 분할 알고리즘에 의한 군집의 상관정보 기반 빅데이터 분석)

  • Park, In-Kyoo
    • Journal of Digital Convergence
    • /
    • v.13 no.11
    • /
    • pp.157-164
    • /
    • 2015
  • This paper describes subspace clustering of categorical data for convergence and integration. Because categorical data are not designed for dealing only with numerical data, The conventional evaluation measures are more likely to have the limitations due to the absence of ordering and high dimensional data and scarcity of frequency. Hence, conditional entropy measure is proposed to evaluate close approximation of cohesion among attributes within each cluster. We propose a new objective function that is used to reflect the optimistic clustering so that the within-cluster dispersion is minimized and the between-cluster separation is enhanced. We performed experiments on five real-world datasets, comparing the performance of our algorithms with four algorithms, using three evaluation metrics: accuracy, f-measure and adjusted Rand index. According to the experiments, the proposed algorithm outperforms the algorithms that were considered int the evaluation, regarding the considered metrics.

A Dynamic Partitioning Scheme for Distributed Storage of Large-Scale RDF Data (대규모 RDF 데이터의 분산 저장을 위한 동적 분할 기법)

  • Kim, Cheon Jung;Kim, Ki Yeon;Yoo, Jong Hyeon;Lim, Jong Tae;Bok, Kyoung Soo;Yoo, Jae Soo
    • Journal of KIISE
    • /
    • v.41 no.12
    • /
    • pp.1126-1135
    • /
    • 2014
  • In recent years, RDF partitioning schemes have been studied for the effective distributed storage and management of large-scale RDF data. In this paper, we propose an RDF dynamic partitioning scheme to support load balancing in dynamic environments where the RDF data is continuously inserted and updated. The proposed scheme creates clusters and sub-clusters according to the frequency of the RDF data used by queries to set graph partitioning criteria. We partition the created clusters and sub-clusters by considering the workloads and data sizes for the servers. Therefore, we resolve the data concentration of a specific server, resulting from the continuous insertion and update of the RDF data, in such a way that the load is distributed among servers in dynamic environments. It is shown through performance evaluation that the proposed scheme significantly improves the query processing time over the existing scheme.

A Collusion-secure Fingerprinting Scheme for Three-dimensional Mesh Models (삼차원 메쉬 모델에 적용한 공모방지 핑거프린팅 기법)

  • Hur, Yung;Jeon, Jeong-Hee;Ho, Yo-Sung
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.41 no.4
    • /
    • pp.113-123
    • /
    • 2004
  • This paper presents a new collusion-secure fingerprinting scheme to embed fingerprints into three-dimensional(3-D) mesh models efficiently. In the proposed scheme, we make the same number of fingerprints as the number of customers based on the finite projective geometry, partition a 3-D mesh model related to the number of bits assigned to each fingerprint and then embed a watermark representing copyright information into each submesh to be marked. Considering imperceptibility and robustness of the watermarking algorithm we embed the watermark signal into mid-frequency DCT coefficients obtained by transforming vertex coordinates in the triangle strips which are generated from the submeshes to be marked. Experimental results show that our scheme is robust to additive random noises, MPEG-4 SNHC 3-D mesh coding, geometrical transformations, and fingerprint attacks by two traitors' collusion. In addition, we can reduce the number of bits assigned to each fingerprint significantly.

Bin Packing Algorithm for Equitable Partitioning Problem with Skill Levels (기량수준 동등분할 문제의 상자 채우기 알고리즘)

  • Lee, Sang-Un
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.20 no.2
    • /
    • pp.209-214
    • /
    • 2020
  • The equitable partitioning problem(EPP) is classified as [0/1] binary skill existence or nonexistence and integer skill levels such as [1,2,3,4,5]. There is well-known a polynomial-time optimal solution finding algorithm for binary skill EPP. On the other hand, tabu search a kind of metaheuristic has apply to integer skill level EPP is due to unknown polynomial-time algorithm for it and this problem is NP-hard. This paper suggests heuristic greedy algorithm with polynomial-time to find the optimal solution for integer skill level EPP. This algorithm descending sorts of skill level frequency for each field and decides the lower bound(LB) that more than the number of group, packing for each group bins first, than the students with less than LB allocates to each bin additionally. As a result of experimental data, this algorithm shows performance improvement than the result of tabu search.

The Effect of Punsimgieumgamibang on Sleep Disorder and Emotionality in Animals (분심기음가미방(分心氣飮加味方)이 타면장애(唾眠障碍)와 정서성(情緖性)에 미치는 영향(影響))

  • Hu Yong-Suk;Kim Jong-Woo;Whang Wei-Wan;Kim Hyun-Taek;Park Soon-Kwon;Kim Hyun-Ju
    • Journal of Oriental Neuropsychiatry
    • /
    • v.11 no.2
    • /
    • pp.43-52
    • /
    • 2000
  • This sudy aimed to evaluate the effect of Punsimgieum-gamibang(分心氣飮加味方) on emotionality and sleep disorder. Punsimgieum(分心氣飮) has been used for insomnia, trembling, emotionality caused by strong mental stimulation and continuous stress in Oriental Medicine. And it was reported that Punsimgieum had anti-stress and anti-depression effect. Animals were devided into two groups : control group and Punsimgieum-gamibang group.Emotionality was tested in open-field with five indexes : walking, rearing, grooming, excretion and start latency. For the study of sleep disorder, after two groups had been given caffeine into abdomen, activity amount in animals was assessed for daytime and then the average percentage of sleep in two groups was calculated.The following results were observed.1. In peripheral and central partition, meditaion group walked more than control group. The difference between two groups was statistically significant.2. In rearing and grooming frequency, there was difference between two groups, but it was not statistically significant. 3. In grooming. excretion and start latency. there was no difference between two groups.4. There was no difference in the average percentage of sleep between two groups.

  • PDF

Transmission and Reflection Characteristics Measurements at the 60GHz for the Various Obstacles (다양한 장애물에 대한 60GHz 대역에서의 투과 및 반사 특성 측정)

  • Song, Ki-Hong
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.12 no.1
    • /
    • pp.25-32
    • /
    • 2008
  • This paper presents the reflection and transmission measurements conducted at the 60GHz suitable to provide a high speed wide band service. Mean received power and standard deviation are calculated and used to compare the characteristics of radio wave propagation to the various obstacles between transmitting and receiving antennas at the frequency. The results show that the transmitted signal strength by the steel door and copper plate are about 40dB lower than in free space, those by the rubber plate, glass and styroform are about 3dB lower than in free space. Also, the re(looted signal strengths at the 60 degree grazing angle show that in case by the partition is about 23dB lower, by the surface of a wall is about 6dB lower than by the copper plate. The presented results can be used for the design of 60 GHz picocell communication network that the reflected and transmitted waves affect to the service area.

Product Recommender Systems using Multi-Model Ensemble Techniques (다중모형조합기법을 이용한 상품추천시스템)

  • Lee, Yeonjeong;Kim, Kyoung-Jae
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.2
    • /
    • pp.39-54
    • /
    • 2013
  • Recent explosive increase of electronic commerce provides many advantageous purchase opportunities to customers. In this situation, customers who do not have enough knowledge about their purchases, may accept product recommendations. Product recommender systems automatically reflect user's preference and provide recommendation list to the users. Thus, product recommender system in online shopping store has been known as one of the most popular tools for one-to-one marketing. However, recommender systems which do not properly reflect user's preference cause user's disappointment and waste of time. In this study, we propose a novel recommender system which uses data mining and multi-model ensemble techniques to enhance the recommendation performance through reflecting the precise user's preference. The research data is collected from the real-world online shopping store, which deals products from famous art galleries and museums in Korea. The data initially contain 5759 transaction data, but finally remain 3167 transaction data after deletion of null data. In this study, we transform the categorical variables into dummy variables and exclude outlier data. The proposed model consists of two steps. The first step predicts customers who have high likelihood to purchase products in the online shopping store. In this step, we first use logistic regression, decision trees, and artificial neural networks to predict customers who have high likelihood to purchase products in each product group. We perform above data mining techniques using SAS E-Miner software. In this study, we partition datasets into two sets as modeling and validation sets for the logistic regression and decision trees. We also partition datasets into three sets as training, test, and validation sets for the artificial neural network model. The validation dataset is equal for the all experiments. Then we composite the results of each predictor using the multi-model ensemble techniques such as bagging and bumping. Bagging is the abbreviation of "Bootstrap Aggregation" and it composite outputs from several machine learning techniques for raising the performance and stability of prediction or classification. This technique is special form of the averaging method. Bumping is the abbreviation of "Bootstrap Umbrella of Model Parameter," and it only considers the model which has the lowest error value. The results show that bumping outperforms bagging and the other predictors except for "Poster" product group. For the "Poster" product group, artificial neural network model performs better than the other models. In the second step, we use the market basket analysis to extract association rules for co-purchased products. We can extract thirty one association rules according to values of Lift, Support, and Confidence measure. We set the minimum transaction frequency to support associations as 5%, maximum number of items in an association as 4, and minimum confidence for rule generation as 10%. This study also excludes the extracted association rules below 1 of lift value. We finally get fifteen association rules by excluding duplicate rules. Among the fifteen association rules, eleven rules contain association between products in "Office Supplies" product group, one rules include the association between "Office Supplies" and "Fashion" product groups, and other three rules contain association between "Office Supplies" and "Home Decoration" product groups. Finally, the proposed product recommender systems provides list of recommendations to the proper customers. We test the usability of the proposed system by using prototype and real-world transaction and profile data. For this end, we construct the prototype system by using the ASP, Java Script and Microsoft Access. In addition, we survey about user satisfaction for the recommended product list from the proposed system and the randomly selected product lists. The participants for the survey are 173 persons who use MSN Messenger, Daum Caf$\acute{e}$, and P2P services. We evaluate the user satisfaction using five-scale Likert measure. This study also performs "Paired Sample T-test" for the results of the survey. The results show that the proposed model outperforms the random selection model with 1% statistical significance level. It means that the users satisfied the recommended product list significantly. The results also show that the proposed system may be useful in real-world online shopping store.

Data-centric XAI-driven Data Imputation of Molecular Structure and QSAR Model for Toxicity Prediction of 3D Printing Chemicals (3D 프린팅 소재 화학물질의 독성 예측을 위한 Data-centric XAI 기반 분자 구조 Data Imputation과 QSAR 모델 개발)

  • ChanHyeok Jeong;SangYoun Kim;SungKu Heo;Shahzeb Tariq;MinHyeok Shin;ChangKyoo Yoo
    • Korean Chemical Engineering Research
    • /
    • v.61 no.4
    • /
    • pp.523-541
    • /
    • 2023
  • As accessibility to 3D printers increases, there is a growing frequency of exposure to chemicals associated with 3D printing. However, research on the toxicity and harmfulness of chemicals generated by 3D printing is insufficient, and the performance of toxicity prediction using in silico techniques is limited due to missing molecular structure data. In this study, quantitative structure-activity relationship (QSAR) model based on data-centric AI approach was developed to predict the toxicity of new 3D printing materials by imputing missing values in molecular descriptors. First, MissForest algorithm was utilized to impute missing values in molecular descriptors of hazardous 3D printing materials. Then, based on four different machine learning models (decision tree, random forest, XGBoost, SVM), a machine learning (ML)-based QSAR model was developed to predict the bioconcentration factor (Log BCF), octanol-air partition coefficient (Log Koa), and partition coefficient (Log P). Furthermore, the reliability of the data-centric QSAR model was validated through the Tree-SHAP (SHapley Additive exPlanations) method, which is one of explainable artificial intelligence (XAI) techniques. The proposed imputation method based on the MissForest enlarged approximately 2.5 times more molecular structure data compared to the existing data. Based on the imputed dataset of molecular descriptor, the developed data-centric QSAR model achieved approximately 73%, 76% and 92% of prediction performance for Log BCF, Log Koa, and Log P, respectively. Lastly, Tree-SHAP analysis demonstrated that the data-centric-based QSAR model achieved high prediction performance for toxicity information by identifying key molecular descriptors highly correlated with toxicity indices. Therefore, the proposed QSAR model based on the data-centric XAI approach can be extended to predict the toxicity of potential pollutants in emerging printing chemicals, chemical process, semiconductor or display process.