• Title/Summary/Keyword: Rough Entropy

Search Result 21, Processing Time 0.031 seconds

Clustering Algorithm for Data Mining using Posterior Probability-based Information Entropy (데이터마이닝을 위한 사후확률 정보엔트로피 기반 군집화알고리즘)

  • Park, In-Kyoo
    • Journal of Digital Convergence
    • /
    • v.12 no.12
    • /
    • pp.293-301
    • /
    • 2014
  • In this paper, we propose a new measure based on the confidence of Bayesian posterior probability so as to reduce unimportant information in the clustering process. Because the performance of clustering is up to selecting the important degree of attributes within the databases, the concept of information entropy is added to posterior probability for attributes discernibility. Hence, The same value of attributes in the confidence of the proposed measure is considerably much less due to the natural logarithm. Therefore posterior probability-based clustering algorithm selects the minimum of attribute reducts and improves the efficiency of clustering. Analysis of the validation of the proposed algorithms compared with others shows their discernibility as well as ability of clustering to handle uncertainty with ACME categorical data.

Mechanism for Shape Control of Cu Nanocrystals Prepared by Polyol Process

  • Cha, Seung-I.;Kim, Kyung-T.;Mo, Chan-B.;Jeong, Yong-J.;Hong, Soon-H.
    • Proceedings of the Korean Powder Metallurgy Institute Conference
    • /
    • 2006.09a
    • /
    • pp.154-155
    • /
    • 2006
  • This study investigated a mechanism for controlling the shape of Cu nanocrystals fabricated using the polyol process, which considers the thermodynamic transition from a facetted surface to a rough surface and the growth mechanisms of nanocrystals with facetted or rough surfaces. The facetted surfaces were stable at relatively low temperatures due to the low entropy of perfectly facetted surfaces. Nanocrystals fabricated using a coordinative surfactant stabilized the facetted surface at a higher temperature than those fabricated using a non-coordinative surfactant. The growth rate of the surface under a given driving force was dependent on the surface structure, i.e., facetted or rough, and the growth of a facetted surface was a thermally activated process. Surface twins decreased the activation energy for growth of the facetted surface and resulted in rod- or wire-shaped nanocrystals

  • PDF

Rough Set Analysis for Stock Market Timing (러프집합분석을 이용한 매매시점 결정)

  • Huh, Jin-Nyung;Kim, Kyoung-Jae;Han, In-Goo
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.3
    • /
    • pp.77-97
    • /
    • 2010
  • Market timing is an investment strategy which is used for obtaining excessive return from financial market. In general, detection of market timing means determining when to buy and sell to get excess return from trading. In many market timing systems, trading rules have been used as an engine to generate signals for trade. On the other hand, some researchers proposed the rough set analysis as a proper tool for market timing because it does not generate a signal for trade when the pattern of the market is uncertain by using the control function. The data for the rough set analysis should be discretized of numeric value because the rough set only accepts categorical data for analysis. Discretization searches for proper "cuts" for numeric data that determine intervals. All values that lie within each interval are transformed into same value. In general, there are four methods for data discretization in rough set analysis including equal frequency scaling, expert's knowledge-based discretization, minimum entropy scaling, and na$\ddot{i}$ve and Boolean reasoning-based discretization. Equal frequency scaling fixes a number of intervals and examines the histogram of each variable, then determines cuts so that approximately the same number of samples fall into each of the intervals. Expert's knowledge-based discretization determines cuts according to knowledge of domain experts through literature review or interview with experts. Minimum entropy scaling implements the algorithm based on recursively partitioning the value set of each variable so that a local measure of entropy is optimized. Na$\ddot{i}$ve and Booleanreasoning-based discretization searches categorical values by using Na$\ddot{i}$ve scaling the data, then finds the optimized dicretization thresholds through Boolean reasoning. Although the rough set analysis is promising for market timing, there is little research on the impact of the various data discretization methods on performance from trading using the rough set analysis. In this study, we compare stock market timing models using rough set analysis with various data discretization methods. The research data used in this study are the KOSPI 200 from May 1996 to October 1998. KOSPI 200 is the underlying index of the KOSPI 200 futures which is the first derivative instrument in the Korean stock market. The KOSPI 200 is a market value weighted index which consists of 200 stocks selected by criteria on liquidity and their status in corresponding industry including manufacturing, construction, communication, electricity and gas, distribution and services, and financing. The total number of samples is 660 trading days. In addition, this study uses popular technical indicators as independent variables. The experimental results show that the most profitable method for the training sample is the na$\ddot{i}$ve and Boolean reasoning but the expert's knowledge-based discretization is the most profitable method for the validation sample. In addition, the expert's knowledge-based discretization produced robust performance for both of training and validation sample. We also compared rough set analysis and decision tree. This study experimented C4.5 for the comparison purpose. The results show that rough set analysis with expert's knowledge-based discretization produced more profitable rules than C4.5.

An Improvement of the Decision-Making of Categorical Data in Rough Set Analysis (범주형 데이터의 러프집합 분석을 통한 의사결정 향상기법)

  • Park, In-Kyu
    • Journal of Digital Convergence
    • /
    • v.13 no.6
    • /
    • pp.157-164
    • /
    • 2015
  • An efficient retrieval of useful information is a prerequisite of an optimal decision making system. Hence, A research of data mining techniques finding useful patterns from the various forms of data has been progressed with the increase of the application of Big Data for convergence and integration with other industries. Each technique is more likely to have its drawback so that the generalization of retrieving useful information is weak. Another integrated technique is essential for retrieving useful information. In this paper, a uncertainty measure of information is calculated such that algebraic probability is measured by Bayesian theory and then information entropy of the probability is measured. The proposed measure generates the effective reduct set (i.e., reduced set of necessary attributes) and formulating the core of the attribute set. Hence, the optimal decision rules are induced. Through simulation deciding contact lenses, the proposed approach is compared with the equivalence and value-reduct theories. As the result, the proposed is more general than the previous theories in useful decision-making.

Missing Pattern Matching of Rough Set Based on Attribute Variations Minimization in Rough Set (속성 변동 최소화에 의한 러프집합 누락 패턴 부합)

  • Lee, Young-Cheon
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.10 no.6
    • /
    • pp.683-690
    • /
    • 2015
  • In Rough set, attribute missing values have several problems such as reduct and core estimation. Further, they do not give some discernable pattern for decision tree construction. Now, there are several methods such as substitutions of typical attribute values, assignment of every possible value, event covering, C4.5 and special LEMS algorithm. However, they are mainly substitutions into frequently appearing values or common attribute ones. Thus, decision rules with high information loss are derived in case that important attribute values are missing in pattern matching. In particular, there is difficult to implement cross validation of the decision rules. In this paper we suggest new method for substituting the missing attribute values into high information gain by using entropy variation among given attributes, and thereby completing the information table. The suggested method is validated by conducting the same rough set analysis on the incomplete information system using the software ROSE.

Discretization of Numerical Attributes and Approximate Reasoning by using Rough Membership Function) (러프 소속 함수를 이용한 수치 속성의 이산화와 근사 추론)

  • Kwon, Eun-Ah;Kim, Hong-Gi
    • Journal of KIISE:Databases
    • /
    • v.28 no.4
    • /
    • pp.545-557
    • /
    • 2001
  • In this paper we propose a hierarchical classification algorithm based on rough membership function which can reason a new object approximately. We use the fuzzy reasoning method that substitutes fuzzy membership value for linguistic uncertainty and reason approximately based on the composition of membership values of conditional sttributes Here we use the rough membership function instead of the fuzzy membership function It can reduce the process that the fuzzy algorithm using fuzzy membership function produces fuzzy rules In addition, we transform the information system to the understandable minimal decision information system In order to do we, study the discretization of continuous valued attributes and propose the discretization algorithm based on the rough membership function and the entropy of the information theory The test shows a good partition that produce the smaller decision system We experimented the IRIS data etc. using our proposed algorithm The experimental results with IRIS data shows 96%~98% rate of classification.

  • PDF

System Modeling based on Genetic Algorithms for Image Restoration : Rough-Fuzzy Entropy (영상복원을 위한 유전자기반 시스템 모델링 : 러프-퍼지엔트로피)

  • 박인규;황상문;진달복
    • Science of Emotion and Sensibility
    • /
    • v.1 no.2
    • /
    • pp.93-103
    • /
    • 1998
  • 효율적이고 체계적인 퍼지제어를 위해 조작자의 제어동작을 모델링하거나 공정을 모델링하는 기법이 필요하고, 또한 퍼지 추론시에 조건부의 기여도(contribution factor)의 결정과 동작부의 제어량의 결정이 추론의 결과에 중요하다. 본 논문에서는 추론시 조건부의 기여도와 동작부의 세어량이 퍼지 엔트로피의 개념하에서 수행되는 적응 퍼지 추론시스템을 제시한다. 제시된 시스템은 전방향 신경회로망의 토대위에서 구현되며 주건부의 기여도가 퍼지 엔트로피에 의하여 구해지고, 동작부의 제어량은 확장된 퍼지 엔트로피에 의하여 구해진다. 이를 위한 학습 알고리즘으로는 역전파 알고리즘을 이용하여 조건부의 파라미터의 동정을 하고 동작부 파라미터의 동정에는 국부해에 보다 강인한 유전자 알고리즘을 이용하다. 이러한 모델링 기법을 임펄스 잡음과 가우시안 잡음이 첨가된 영상에 적용하여 본 결과, 영상복원시에 발생되는 여러 가지의 경우에 대한 적응성이 보다 양호하게 유지되었고, 전체영상의 20%의 데이터만으로도 객관적 화질에 있어서 기존의 추론 방법에 비해 향상을 보였다.

  • PDF

A Property Analysis of Parameters for Effective Watermarking based on Human Visual System (HVS 기반 효과적인 워터마킹을 위한 파라미터 성능분석)

  • Park Ki-Hong;Yoon Byung-Min;Kim YoonHo
    • Journal of Digital Contents Society
    • /
    • v.6 no.2
    • /
    • pp.119-123
    • /
    • 2005
  • We proposed a method of effective watermarking based m HVS. To speak generally, HVS parameters are contrast sensitivity, texture degree, entropy sensitivity and standard of deviation, etc. The aim of this paper is evaluated by the experiments of imperceptibility and correctness of watermark. According to some experimental results, contrast sensitivity function is superior in smooth image. On the other hand, standard of deviation provides good results in rough images. Consequently, how to select the parameters considering image attribute is key problem in effective watermarking.

  • PDF

Methodology to Verify the Unpredictability of True Random Number Generators (실난수 발생기 통계적 예측 불가능성 확인 방법)

  • Moon-Seok Kim;Seung-Bae Jeon
    • Convergence Security Journal
    • /
    • v.24 no.2
    • /
    • pp.123-132
    • /
    • 2024
  • In the era of the Internet of Things, 7 billion diverse devices have been interconnected worldwide. Ensuring information security across these varied devices is crucial in this hyper-connected age. To achieve essential security functions such as confidentiality, integrity, and authentication, it is imperative to implement true random number generators (TRNGs). Therefore, this study proposes a method to rapidly characterize the randomness of TRNGs. While there are international standards for formally characterizing the randomness of TRNGs, adhering to these standards often requires significant time and resources. This study aims to help TRNG developers enhance efficiency in both time and cost by characterizing rough randomness and unpredictability. Firstly, we propose applying auto-correlation and cross-correlation metrics for analog signals. Secondly, we suggest adopting joint entropy and mutual information metrics for digital signals.

Application of Chiu's Two Dimensional Velocity Distribution Equations to Natural Rivers (Chiu가 제안한 2차원 유속분포식의 자연하천 적용성 분석)

  • Lee, Chan-Joo;Seo, Il-Won;Kim, Chang-Wan;Kim, Won
    • Journal of Korea Water Resources Association
    • /
    • v.40 no.12
    • /
    • pp.957-968
    • /
    • 2007
  • It is essential to obtain accurate and highly reliable streamflow data for quantitative management for water resources. Thereafter such real-time streamflow gauging methods as ultrasonic flowmeter and index-velocity are introduced recently. Since these methods calculate flowrate through entire cross-section by measuring partial velocities of it, rational and theoretical basis are necessary for accurate estimation of discharge. The purpose of the present study lies in analysis on the applicability of Chiu#s(1987, 1988) two dimensional velocity distribution equations by applying them to natural rivers and by comparing simulated velocity distributions with observed ones obtained with ADCP. Maximum and mean velocities are calculated from observed data to estimate entropy parameter M. Such isovel shape parameters as h and $\beta_i$ are estimated by object function based on least squares criterion. In case optimized parameters are applied, Chiu#s velocity distributions fairly well simulate observed ones. By using 14 simulated data sets which have relatively high correlation coefficients, properties of parameters are analyzed and h, $\beta_i$ are estimated for velocity-unknown river sections. When estimated parameters are adopted for verification, simulated velocity distributions well reproduce real ones. Finally, calculated discharges display rough agreement with measured data. The results of the present study mean that if parameters related are properly estimated, Chiu#s velocity distribution is likely to reproduce the real one of natural rivers.