• Title/Summary/Keyword: Rule-Based Model

Search Result 1,025, Processing Time 0.03 seconds

Fraud Detection System Model Using Generative Adversarial Networks and Deep Learning (생성적 적대 신경망과 딥러닝을 활용한 이상거래탐지 시스템 모형)

  • Ye Won Kim;Ye Lim Yu;Hong Yong Choi
    • Information Systems Review
    • /
    • v.22 no.1
    • /
    • pp.59-72
    • /
    • 2020
  • Artificial Intelligence is establishing itself as a familiar tool from an intractable concept. In this trend, financial sector is also looking to improve the problem of existing system which includes Fraud Detection System (FDS). It is being difficult to detect sophisticated cyber financial fraud using original rule-based FDS. This is because diversification of payment environment and increasing number of electronic financial transactions has been emerged. In order to overcome present FDS, this paper suggests 3 types of artificial intelligence models, Generative Adversarial Network (GAN), Deep Neural Network (DNN), and Convolutional Neural Network (CNN). GAN proves how data imbalance problem can be developed while DNN and CNN show how abnormal financial trading patterns can be precisely detected. In conclusion, among the experiments on this paper, WGAN has the highest improvement effects on data imbalance problem. DNN model reflects more effects on fraud classification comparatively.

Improved Sentence Boundary Detection Method for Web Documents (웹 문서를 위한 개선된 문장경계인식 방법)

  • Lee, Chung-Hee;Jang, Myung-Gil;Seo, Young-Hoon
    • Journal of KIISE:Software and Applications
    • /
    • v.37 no.6
    • /
    • pp.455-463
    • /
    • 2010
  • In this paper, we present an approach to sentence boundary detection for web documents that builds on statistical-based methods and uses rule-based correction. The proposed system uses the classification model learned offline using a training set of human-labeled web documents. The web documents have many word-spacing errors and frequently no punctuation mark that indicates the end of sentence boundary. As sentence boundary candidates, the proposed method considers every Ending Eomis as well as punctuation marks. We optimize engine performance by selecting the best feature, the best training data, and the best classification algorithm. For evaluation, we made two test sets; Set1 consisting of articles and blog documents and Set2 of web community documents. We use F-measure to compare results on a large variety of tasks, Detecting only periods as sentence boundary, our basis engine showed 96.5% in Set1 and 56.7% in Set2. We improved our basis engine by adapting features and the boundary search algorithm. For the final evaluation, we compared our adaptation engine with our basis engine in Set2. As a result, the adaptation engine obtained improvements over the basis engine by 39.6%. We proved the effectiveness of the proposed method in sentence boundary detection.

Identifying Security Requirement using Reusable State Transition Diagram at Security Threat Location (보안 위협위치에서 재사용 가능한 상태전이도를 이용한 보안요구사항 식별)

  • Seo Seong-Chae;You Jin-Ho;Kim Young-Dae;Kim Byung-Ki
    • The KIPS Transactions:PartD
    • /
    • v.13D no.1 s.104
    • /
    • pp.67-74
    • /
    • 2006
  • The security requirements identification in the software development has received some attention recently. However, previous methods do not provide clear method and process of security requirements identification. We propose a process that software developers can build application specific security requirements from state transition diagrams at the security threat location. The proposed process consists of building model and identifying application specific security requirements. The state transition diagram is constructed through subprocesses i) the identification of security threat locations using security failure data based on the point that attackers exploit software vulnerabilities and attack system assets, ii) the construction of a state transition diagram which is usable to protect, mitigate, and remove vulnerabilities of security threat locations. The identification Process of application specific security requirements consist of i) the analysis of the functional requirements of the software, which are decomposed into a DFD(Data Flow Diagram; the identification of the security threat location; and the appliance of the corresponding state transition diagram into the security threat locations, ii) the construction of the application specific state transition diagram, iii) the construction of security requirements based on the rule of the identification of security requirements. The proposed method is helpful to identify the security requirements easily at an early phase of software development.

Development and assessment of framework for selecting multi-GCMs considering Asia monsoon characteristics (아시아 몬순특성을 고려한 다중 GCMs 선정방법 개발 및 평가)

  • Kim, Jeong-Bae;Kim, Jin-Hoon;Bae, Deg-Hyo
    • Journal of Korea Water Resources Association
    • /
    • v.53 no.9
    • /
    • pp.647-660
    • /
    • 2020
  • The objectives of this study are to develop a framework for selecting multi-GCMs considering Asia monsoon characteristics and assess it's applicability. 12 climate variables related to monsoon climates are selected for GCM selection. The framework for selecting multi-GCMs includes the evaluation matrix of GCM performance based on their capability to simulate historical climate features. The climatological patterns of 12 variables derived from individual GCM over the summer monsoon season during the past period (1976-2005) and they are compared against observations to evaluate GCM performance. For objective evaluation, a rigorous scoring rule is implemented by comparing the GCM performance based on the results of statistics between historical simulation derived from individual GCM and observations. Finally, appropriate 5 GCMs (NorESM1-M, bcc-csm1-m, CNRM-CM5, CMCC-CMS, and CanESM2) are selected in consideration of the ranking of GCM and precipitation performance of each GCM. The selected 5 GCMs are compared with the historical observations in terms of monsoon season and monthly mean to validate their applicability. The 5 GCMs well capture the observational climate characteristics of Asia for the 12 climate variables also they reduce the bias between the entire GCM simulations and the observational data. This study demonstrates that it is necessary to consider various climate variables for GCM selection and, the method introduced in this study can be used to select more reliable climate change scenarios for climate change assessment in the Asia region.

An Implementation of Lighting Control System using Interpretation of Context Conflict based on Priority (우선순위 기반의 상황충돌 해석 조명제어시스템 구현)

  • Seo, Won-Il;Kwon, Sook-Youn;Lim, Jae-Hyun
    • Journal of Internet Computing and Services
    • /
    • v.17 no.1
    • /
    • pp.23-33
    • /
    • 2016
  • The current smart lighting is shaped to offer the lighting environment suitable for current context, after identifying user's action and location through a sensor. The sensor-based context awareness technology just considers a single user, and the studies to interpret many users' various context occurrences and conflicts lack. In existing studies, a fuzzy theory and algorithm including ReBa have been used as the methodology to solve context conflict. The fuzzy theory and algorithm including ReBa just avoid an opportunity of context conflict that may occur by providing services by each area, after the spaces where users are located are classified into many areas. Therefore, they actually cannot be regarded as customized service type that can offer personal preference-based context conflict. This paper proposes a priority-based LED lighting control system interpreting multiple context conflicts, which decides services, based on the granted priority according to context type, when service conflict is faced with, due to simultaneous occurrence of various contexts to many users. This study classifies the residential environment into such five areas as living room, 'bed room, study room, kitchen and bath room, and the contexts that may occur within each area are defined as 20 contexts such as exercising, doing makeup, reading, dining and entering, targeting several users. The proposed system defines various contexts of users using an ontology-based model and gives service of user oriented lighting environment through rule based on standard and context reasoning engine. To solve the issue of various context conflicts among users in the same space and at the same time point, the context in which user concentration is required is set in the highest priority. Also, visual comfort is offered as the best alternative priority in the case of the same priority. In this manner, they are utilized as the criteria for service selection upon conflict occurrence.

Improving the Accuracy of Document Classification by Learning Heterogeneity (이질성 학습을 통한 문서 분류의 정확성 향상 기법)

  • Wong, William Xiu Shun;Hyun, Yoonjin;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.3
    • /
    • pp.21-44
    • /
    • 2018
  • In recent years, the rapid development of internet technology and the popularization of smart devices have resulted in massive amounts of text data. Those text data were produced and distributed through various media platforms such as World Wide Web, Internet news feeds, microblog, and social media. However, this enormous amount of easily obtained information is lack of organization. Therefore, this problem has raised the interest of many researchers in order to manage this huge amount of information. Further, this problem also required professionals that are capable of classifying relevant information and hence text classification is introduced. Text classification is a challenging task in modern data analysis, which it needs to assign a text document into one or more predefined categories or classes. In text classification field, there are different kinds of techniques available such as K-Nearest Neighbor, Naïve Bayes Algorithm, Support Vector Machine, Decision Tree, and Artificial Neural Network. However, while dealing with huge amount of text data, model performance and accuracy becomes a challenge. According to the type of words used in the corpus and type of features created for classification, the performance of a text classification model can be varied. Most of the attempts are been made based on proposing a new algorithm or modifying an existing algorithm. This kind of research can be said already reached their certain limitations for further improvements. In this study, aside from proposing a new algorithm or modifying the algorithm, we focus on searching a way to modify the use of data. It is widely known that classifier performance is influenced by the quality of training data upon which this classifier is built. The real world datasets in most of the time contain noise, or in other words noisy data, these can actually affect the decision made by the classifiers built from these data. In this study, we consider that the data from different domains, which is heterogeneous data might have the characteristics of noise which can be utilized in the classification process. In order to build the classifier, machine learning algorithm is performed based on the assumption that the characteristics of training data and target data are the same or very similar to each other. However, in the case of unstructured data such as text, the features are determined according to the vocabularies included in the document. If the viewpoints of the learning data and target data are different, the features may be appearing different between these two data. In this study, we attempt to improve the classification accuracy by strengthening the robustness of the document classifier through artificially injecting the noise into the process of constructing the document classifier. With data coming from various kind of sources, these data are likely formatted differently. These cause difficulties for traditional machine learning algorithms because they are not developed to recognize different type of data representation at one time and to put them together in same generalization. Therefore, in order to utilize heterogeneous data in the learning process of document classifier, we apply semi-supervised learning in our study. However, unlabeled data might have the possibility to degrade the performance of the document classifier. Therefore, we further proposed a method called Rule Selection-Based Ensemble Semi-Supervised Learning Algorithm (RSESLA) to select only the documents that contributing to the accuracy improvement of the classifier. RSESLA creates multiple views by manipulating the features using different types of classification models and different types of heterogeneous data. The most confident classification rules will be selected and applied for the final decision making. In this paper, three different types of real-world data sources were used, which are news, twitter and blogs.

Knowledge Extraction Methodology and Framework from Wikipedia Articles for Construction of Knowledge-Base (지식베이스 구축을 위한 한국어 위키피디아의 학습 기반 지식추출 방법론 및 플랫폼 연구)

  • Kim, JaeHun;Lee, Myungjin
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.43-61
    • /
    • 2019
  • Development of technologies in artificial intelligence has been rapidly increasing with the Fourth Industrial Revolution, and researches related to AI have been actively conducted in a variety of fields such as autonomous vehicles, natural language processing, and robotics. These researches have been focused on solving cognitive problems such as learning and problem solving related to human intelligence from the 1950s. The field of artificial intelligence has achieved more technological advance than ever, due to recent interest in technology and research on various algorithms. The knowledge-based system is a sub-domain of artificial intelligence, and it aims to enable artificial intelligence agents to make decisions by using machine-readable and processible knowledge constructed from complex and informal human knowledge and rules in various fields. A knowledge base is used to optimize information collection, organization, and retrieval, and recently it is used with statistical artificial intelligence such as machine learning. Recently, the purpose of the knowledge base is to express, publish, and share knowledge on the web by describing and connecting web resources such as pages and data. These knowledge bases are used for intelligent processing in various fields of artificial intelligence such as question answering system of the smart speaker. However, building a useful knowledge base is a time-consuming task and still requires a lot of effort of the experts. In recent years, many kinds of research and technologies of knowledge based artificial intelligence use DBpedia that is one of the biggest knowledge base aiming to extract structured content from the various information of Wikipedia. DBpedia contains various information extracted from Wikipedia such as a title, categories, and links, but the most useful knowledge is from infobox of Wikipedia that presents a summary of some unifying aspect created by users. These knowledge are created by the mapping rule between infobox structures and DBpedia ontology schema defined in DBpedia Extraction Framework. In this way, DBpedia can expect high reliability in terms of accuracy of knowledge by using the method of generating knowledge from semi-structured infobox data created by users. However, since only about 50% of all wiki pages contain infobox in Korean Wikipedia, DBpedia has limitations in term of knowledge scalability. This paper proposes a method to extract knowledge from text documents according to the ontology schema using machine learning. In order to demonstrate the appropriateness of this method, we explain a knowledge extraction model according to the DBpedia ontology schema by learning Wikipedia infoboxes. Our knowledge extraction model consists of three steps, document classification as ontology classes, proper sentence classification to extract triples, and value selection and transformation into RDF triple structure. The structure of Wikipedia infobox are defined as infobox templates that provide standardized information across related articles, and DBpedia ontology schema can be mapped these infobox templates. Based on these mapping relations, we classify the input document according to infobox categories which means ontology classes. After determining the classification of the input document, we classify the appropriate sentence according to attributes belonging to the classification. Finally, we extract knowledge from sentences that are classified as appropriate, and we convert knowledge into a form of triples. In order to train models, we generated training data set from Wikipedia dump using a method to add BIO tags to sentences, so we trained about 200 classes and about 2,500 relations for extracting knowledge. Furthermore, we evaluated comparative experiments of CRF and Bi-LSTM-CRF for the knowledge extraction process. Through this proposed process, it is possible to utilize structured knowledge by extracting knowledge according to the ontology schema from text documents. In addition, this methodology can significantly reduce the effort of the experts to construct instances according to the ontology schema.

Development and Application of Learning Materials for the Law of Planetary Motion using the Kepler's Abductive Reasoning (행성운동법칙에 관한 케플러의 귀추적 사고를 도입한 학습자료의 개발 및 적용)

  • Park, Su-Gyeong
    • Journal of the Korean earth science society
    • /
    • v.33 no.2
    • /
    • pp.170-182
    • /
    • 2012
  • The purpose of this study was to develop learning materials based on the Kepler's abductive reasoning and to identify high school students' rule-inferring strategies on the law of planetary motion. The learning materials including the concepts of solar magnetic field, conservation of figure skater's angular momentum and Kepler's polyhedral theory were developed and the questions about Kepler's 2nd and 3rd law of planetary motion were also created. The participants were 79science high school students and 83general high school students. The patterns and properties of their abductive inference were analyzed. The findings revealed that the students showed 'incomplete analogy abduction', 'analogy abduction' and 'reconstruction' to generate the hypotheses concerning the Mars' motion related to the solar magnetic field. There were more general high school students who showed the incomplete analogy abduction than science high school students. On the other hand, there were more science high school students who showed the analogy abduction and reconstruction strategy than general high school students. Also, they showed 'incomplete analogy abduction', 'analogy abduction' and 'model construction and manipulation' to generate the hypotheses concerning Kepler's second law. A number of general high school students showed the incomplete analogy. It is suggested that because the analogy of figure skater cause the students' alternative framework to use, more detailed demonstration is necessary in class. In addition, students combined Kepler's polyhedral theory with their prior knowledge to infer Kepler's third law.

Life Prediction of Composite Pressure Vessels Using Multi-Scale Approach (멀티 스케일 접근법을 이용한 복합재 압력용기의 수명 예측)

  • Jin, Kyo-Kook;Ha, Sung-Kyu;Kim, Jae-Hyuk;Han, Hoon-Hee;Kim, Seong-Jong
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.11 no.9
    • /
    • pp.3176-3183
    • /
    • 2010
  • A multi-scale fatigue life prediction methodology of composite pressure vessels subjected to multi-axial loading has been proposed in this paper. The multi-scale approach starts from the constituents, fiber, matrix and interface, leading to predict behavior of ply, laminates and eventually the composite structures. The multi-scale fatigue life prediction methodology is composed of two steps: macro stress analysis and micro mechanics of failure based on fatigue analysis. In the macro stress analysis, multi-axial fatigue loading acting at laminate is determined from finite element analysis of composite pressure vessel, and ply stresses are computed using a classical laminate theory. The micro stresses are calculated in each constituent from ply stresses using a micromechanical model. Three methods are employed in predicting fatigue life of each constituent, i.e. a maximum stress method for fiber, an equivalent stress method for multi-axially loaded matrix, and a critical plane method for the interface. A modified Goodman diagram is used to take into account the generic mean stresses. Damages from each loading cycle are accumulated using Miner's rule. Monte Carlo simulation has been performed to predict the overall fatigue life of a composite pressure vessel considering statistical distribution of material properties of each constituent, fiber volume fraction and manufacturing winding angle.

The Impact of an Ontological Knowledge Representation on Information Retrieval: An Evaluation Study of OCLC's FRBR-Based FictionFinder (정보검색에 온톨로지 지식 표현이 미치는 영향에 대한 연구: OCLC의 FRBR기반 FictionFinder의 평가를 중심으로)

  • Cho, Myung-Dae
    • Journal of the Korean Society for information Management
    • /
    • v.25 no.2
    • /
    • pp.183-198
    • /
    • 2008
  • With the purpose of enriching existing catalogues with FRBR, which is the Functional Requirements for Bibliographic Records, in mind, this paper aims to evaluate the impact of bibliographic ontology on the overall system's performance in the field of literature. In doing this, OCLC's FictionFinder(http://fictionfinder.oclc.org) was selected and qualitatively evaluated. In this study 40 university seniors evaluated the following three aspects using the 'transferring thoughts onto paper method': 1) In which ways is this FRBR-aware bibliographical ontology helpful? 2) Are the things which are initially attempted to be helped being helped? 3) Would users seeking one work in particular also see all other related works? In conclusion, this study revealed that, as Cutter claimed in his $2^{nd}$ rule of the library, collocations give added-value to the users and overall ontology provides better interface and usefulness. It also revealed that a system's evaluation with qualitative methodology helped to build full pictures of the system and to grip the information needs of the users when the system is developed. Qualitative evaluations, therefore, could be used as indicators for the evaluation of any information retrieval systems.