• 제목/요약/키워드: Data Models

검색결과 13,916건 처리시간 0.038초

A Study of Data Mining Optimization Model for the Credit Evaluation

  • Kim, Kap-Sik;Lee, Chang-Soon
    • Journal of the Korean Data and Information Science Society
    • /
    • 제14권4호
    • /
    • pp.825-836
    • /
    • 2003
  • Based on customer information and financing processes in capital market, we derived individual models by applying multi-layered perceptrons, MDA, and decision tree. Further, the results from the existing single models were compared with the results from the integrated model that was developed using genetic algorithm. This study contributes not only to verifying the existing individual models and but also to overcoming the limitations of the existing approaches. We have depended upon the approaches that compare individual models and search for the best-fit model. However, this study presents a methodology to build an integrated data mining model using genetic algorithm.

  • PDF

부정확한 데이터를 가지는 자료포락분석을 위한 로버스트 최적화 모형의 적용 (Data Envelopment Analysis with Imprecise Data Based on Robust Optimization)

  • 임성묵
    • 산업경영시스템학회지
    • /
    • 제38권4호
    • /
    • pp.117-131
    • /
    • 2015
  • Conventional data envelopment analysis (DEA) models require that inputs and outputs are given as crisp values. Very often, however, some of inputs and outputs are given as imprecise data where they are only known to lie within bounded intervals. While a typical approach to addressing this situation for optimization models such as DEA is to conduct sensitivity analysis, it provides only a limited ex-post measure against the data imprecision. Robust optimization provides a more effective ex-ante measure where the data imprecision is directly incorporated into the model. This study aims to apply robust optimization approach to DEA models with imprecise data. Based upon a recently developed robust optimization framework which allows a flexible adjustment of the level of conservatism, we propose two robust optimization DEA model formulations with imprecise data; multiplier and envelopment models. We demonstrate that the two models consider different risks regarding imprecise efficiency scores, and that the existing DEA models with imprecise data are special cases of the proposed models. We show that the robust optimization for the multiplier DEA model considers the risk that estimated efficiency scores exceed true values, while the one for the envelopment DEA model deals with the risk that estimated efficiency scores fall short of true values. We also show that efficiency scores stratified in terms of probabilistic bounds of constraint violations can be obtained from the proposed models. We finally illustrate the proposed approach using a sample data set and show how the results can be used for ranking DMUs.

Wavelet-like convolutional neural network structure for time-series data classification

  • Park, Seungtae;Jeong, Haedong;Min, Hyungcheol;Lee, Hojin;Lee, Seungchul
    • Smart Structures and Systems
    • /
    • 제22권2호
    • /
    • pp.175-183
    • /
    • 2018
  • Time-series data often contain one of the most valuable pieces of information in many fields including manufacturing. Because time-series data are relatively cheap to acquire, they (e.g., vibration signals) have become a crucial part of big data even in manufacturing shop floors. Recently, deep-learning models have shown state-of-art performance for analyzing big data because of their sophisticated structures and considerable computational power. Traditional models for a machinery-monitoring system have highly relied on features selected by human experts. In addition, the representational power of such models fails as the data distribution becomes complicated. On the other hand, deep-learning models automatically select highly abstracted features during the optimization process, and their representational power is better than that of traditional neural network models. However, the applicability of deep-learning models to the field of prognostics and health management (PHM) has not been well investigated yet. This study integrates the "residual fitting" mechanism inherently embedded in the wavelet transform into the convolutional neural network deep-learning structure. As a result, the architecture combines a signal smoother and classification procedures into a single model. Validation results from rotor vibration data demonstrate that our model outperforms all other off-the-shelf feature-based models.

범주형 재무자료에 대한 신용평가모형 검증 비교 (Validation Comparison of Credit Rating Models for Categorized Financial Data)

  • 홍종선;이창혁;김지훈
    • Communications for Statistical Applications and Methods
    • /
    • 제15권4호
    • /
    • pp.615-631
    • /
    • 2008
  • 재무자료에 대한 신용평가모형은 각각의 재무변수를 평활한 예측부도율로 변환하여 사용한다. 본 연구에서는 연속형 재무자료를 변환하여 설정된 신용평가모형의 문제점을 살펴보고, 연속형 재무변수를 다양한 형태로 범주화한 신용평가모형들을 제안한다. 범주형 재무자료를 사용해서 개발한 여러 종류의 신용평가모형들의 성과를 다양한 적합성 검증 방법으로 비교하고, 범주형 재무자료를 이용한 신용평가모형의 유용성을 토론한다.

효율적인 신용평가를 위한 데이터마이닝 모형의 비교.분석에 관한 연구 (Study on the Comparison and Analysis of Data Mining Models for the Efficient Customer Credit Evaluation)

  • 김갑식
    • Journal of Information Technology Applications and Management
    • /
    • 제11권1호
    • /
    • pp.161-174
    • /
    • 2004
  • This study is intended to suggest1 the optimized data mining model for the efficient customer credit evaluation in the capital finance industry. To accomplish the research objective, various data mining models for the customer credit evaluation are compared and analyzed. Furthermore, existing models such as Multi-Layered Perceptrons, Multivariate Discrimination Analysis, Radial Basis Function, Decision Tree, and Logistic Regression are employed for analyzing the customer information in the capital finance market and the detailed data of capital financing transactions. Finally, the data from the integrated model utilizing a genetic algorithm is compared with those of each individual model mentioned above. The results reveals that the integrated model is superior to other existing models.

  • PDF

A Temporal Data model and a Query Language Based on the OO data model

  • Shu, Yongmoo
    • 경영과학
    • /
    • 제14권1호
    • /
    • pp.87-105
    • /
    • 1997
  • There have been lots of research on temporal data management for the past two decades. Most of them are based on some logical data model, especially on the relational data model, although there are some conceptual data models which are independent of logical data models. Also, many properties or issues regarding temporal data models and temporal query languages have been studied. But some of them were shown to be incompatible, which means there could not be a complete temporal data model, satisfying all the desired properties at the same time. Many modeling issues discussed in the papers, do not have to be done so, if they take object-oriented data model as a base model. Therefore, this paper proposes a temporal data model, which is based on the object-oriented data model, mainly discussing the most essential issues that are common to many temporal data models. Our new temporal data model and query language will be illustrated with a small database, created by a set of sample transaction.

  • PDF

A Temporal Data model and a Query Language Based on the OO data model

  • 서용무
    • 한국경영과학회지
    • /
    • 제14권1호
    • /
    • pp.87-87
    • /
    • 1989
  • There have been lots of research on temporal data management for the past two decades. Most of them are based on some logical data model, especially on the relational data model, although there are some conceptual data models which are independent of logical data models. Also, many properties or issues regarding temporal data models and temporal query languages have been studied. But some of them were shown to be incompatible, which means there could not be a complete temporal data model, satisfying all the desired properties at the same time. Many modeling issues discussed in the papers, do not have to be done so, if they take object-oriented data model as a base model. Therefore, this paper proposes a temporal data model, which is based on the object-oriented data model, mainly discussing the most essential issues that are common to many temporal data models. Our new temporal data model and query language will be illustrated with a small database, created by a set of sample transaction.

An XPDL-Based Workflow Control-Structure and Data-Sequence Analyzer

  • Kim, Kwanghoon Pio
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제13권3호
    • /
    • pp.1702-1721
    • /
    • 2019
  • A workflow process (or business process) management system helps to define, execute, monitor and manage workflow models deployed on a workflow-supported enterprise, and the system is compartmentalized into a modeling subsystem and an enacting subsystem, in general. The modeling subsystem's functionality is to discover and analyze workflow models via a theoretical modeling methodology like ICN, to graphically define them via a graphical representation notation like BPMN, and to systematically deploy those graphically defined models onto the enacting subsystem by transforming into their textual models represented by a standardized workflow process definition language like XPDL. Before deploying those defined workflow models, it is very important to inspect its syntactical correctness as well as its structural properness to minimize the loss of effectiveness and the depreciation of efficiency in managing the corresponding workflow models. In this paper, we are particularly interested in verifying very large-scale and massively parallel workflow models, and so we need a sophisticated analyzer to automatically analyze those specialized and complex styles of workflow models. One of the sophisticated analyzers devised in this paper is able to analyze not only the structural complexity but also the data-sequence complexity, especially. The structural complexity is based upon combinational usages of those control-structure constructs such as subprocesses, exclusive-OR, parallel-AND and iterative-LOOP primitives with preserving matched pairing and proper nesting properties, whereas the data-sequence complexity is based upon combinational usages of those relevant data repositories such as data definition sequences and data use sequences. Through the devised and implemented analyzer in this paper, we are able eventually to achieve the systematic verifications of the syntactical correctness as well as the effective validation of the structural properness on those complicate and large-scale styles of workflow models. As an experimental study, we apply the implemented analyzer to an exemplary large-scale and massively parallel workflow process model, the Large Bank Transaction Workflow Process Model, and show the structural complexity analysis results via a series of operational screens captured from the implemented analyzer.

Mixed Linear Models with Censored Data

  • Ha, Il-do;Lee, Youngjo-;Song, Jae-Kee
    • Journal of the Korean Statistical Society
    • /
    • 제28권2호
    • /
    • pp.211-223
    • /
    • 1999
  • We propose a simple estimation procedure in the mixed linear models with censored normal data, using both Buckly and James(1979) type pseudo random variables and Lee and Nelder's(1996) estimation procedure. The proposed method is illustrated with the matched pairs data in Pettitt(1986).

  • PDF

Semi-Supervised Recursive Learning of Discriminative Mixture Models for Time-Series Classification

  • Kim, Minyoung
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • 제13권3호
    • /
    • pp.186-199
    • /
    • 2013
  • We pose pattern classification as a density estimation problem where we consider mixtures of generative models under partially labeled data setups. Unlike traditional approaches that estimate density everywhere in data space, we focus on the density along the decision boundary that can yield more discriminative models with superior classification performance. We extend our earlier work on the recursive estimation method for discriminative mixture models to semi-supervised learning setups where some of the data points lack class labels. Our model exploits the mixture structure in the functional gradient framework: it searches for the base mixture component model in a greedy fashion, maximizing the conditional class likelihoods for the labeled data and at the same time minimizing the uncertainty of class label prediction for unlabeled data points. The objective can be effectively imposed as individual mixture component learning on weighted data, hence our mixture learning typically becomes highly efficient for popular base generative models like Gaussians or hidden Markov models. Moreover, apart from the expectation-maximization algorithm, the proposed recursive estimation has several advantages including the lack of need for a pre-determined mixture order and robustness to the choice of initial parameters. We demonstrate the benefits of the proposed approach on a comprehensive set of evaluations consisting of diverse time-series classification problems in semi-supervised scenarios.