• Title/Summary/Keyword: Inference System

Search Result 1,622, Processing Time 0.034 seconds

Design of Optimized pRBFNNs-based Face Recognition Algorithm Using Two-dimensional Image and ASM Algorithm (최적 pRBFNNs 패턴분류기 기반 2차원 영상과 ASM 알고리즘을 이용한 얼굴인식 알고리즘 설계)

  • Oh, Sung-Kwun;Ma, Chang-Min;Yoo, Sung-Hoon
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.21 no.6
    • /
    • pp.749-754
    • /
    • 2011
  • In this study, we propose the design of optimized pRBFNNs-based face recognition system using two-dimensional Image and ASM algorithm. usually the existing 2 dimensional face recognition methods have the effects of the scale change of the image, position variation or the backgrounds of an image. In this paper, the face region information obtained from the detected face region is used for the compensation of these defects. In this paper, we use a CCD camera to obtain a picture frame directly. By using histogram equalization method, we can partially enhance the distorted image influenced by natural as well as artificial illumination. AdaBoost algorithm is used for the detection of face image between face and non-face image area. We can butt up personal profile by extracting the both face contour and shape using ASM(Active Shape Model) and then reduce dimension of image data using PCA. The proposed pRBFNNs consists of three functional modules such as the condition part, the conclusion part, and the inference part. In the condition part of fuzzy rules, input space is partitioned with Fuzzy C-Means clustering. In the conclusion part of rules, the connection weight of RBFNNs is represented as three kinds of polynomials such as constant, linear, and quadratic. The essential design parameters (including learning rate, momentum coefficient and fuzzification coefficient) of the networks are optimized by means of Differential Evolution. The proposed pRBFNNs are applied to real-time face image database and then demonstrated from viewpoint of the output performance and recognition rate.

A Performance Analysis of Large ABox Reasoning in OWL-DL Reasoners (다양한 OWL-DL 추론 엔진에서 대용량 ABox 추론에 대한 성능평가)

  • Seo, Eun-Seok;Park, Young-Tack
    • Journal of KIISE:Software and Applications
    • /
    • v.34 no.7
    • /
    • pp.655-666
    • /
    • 2007
  • Reasoners using typical Tableaux algorithm such as RacerPro, Pellet have a problem in Tableaux algorithm large ABox reasoning. Researches to solve these Problems are dealt with Instance Store of University of Manchester which uses Tableaux algorithm based reasoner and DBMS and KAON2 of University of Karlsruhe using Disjunctive Datalog approach. An evaluation experiment for present reasoners is the experiment of TBox reasoning in most of Tableaux algorithm based one. The most of benchmarking tests in reasoning systems haven't done with ABox reasoning based Tableaux Algorithm but done with TBox reasoning based Tableaux Algorithm. Especially, rarely reported benchmarking tests in reasoners have been issued nowadays. Therefore, this thesis evaluates systems with theory of each reasoners for large ABox reasoning that becomes issues recently with typical reasoners. The large AoBx reasoning engine will be analyzed using Instance Store and KAON2 of Manchester University for large ABox processing. At the analysing method, LUBM(Lehigh University BenchMark), benchmarking test method, and it's test system will be introduced. In conclusion, I recommend appropriate reasoner in various environment with experiment result and characteristic of algorithm used for each reasoner.

Development of Empathy Education Program Using Brain-Based Education Principles in Home Economics (뇌기반 교육원리를 적용한 가정과 공감교육 프로그램 개발)

  • Lee, Eunjin;Choi, Saeeun
    • Journal of Korean Home Economics Education Association
    • /
    • v.33 no.2
    • /
    • pp.153-172
    • /
    • 2021
  • The purpose of this study is to develop the Empathy Education Program by employing the brain-based education principle in home economics education. For this purpose, the study used the "ADDE" (Analysis, Design, Development, and Evaluation) method to develope a Empathy Education Program for middle-school students. Analysis from previous literature derived that four principles of brain-based empathy education in home economics education: 1) understanding through imitation; 2) inference through imagination; 3) interaction through experience, and 4) internalization through practice. Based on these brain-based four principles and three components of empathy, a total of 15 lessons were designed and developed for the educational program for middle-school students, "FEEL" (For Empathy Education & Learning). The contents of this program were selected from the analysis of the 2015-revised technology and home economics education curriculum and textbooks. Results from expert evaluation of the validity and feasibility of the program showed that this program is highly valid with an average validity score of 4.88 and 0.98 for the validity index ("CVI"). This study is meaningful in that it can be effectively employed in the middle school free-year policy system. Also, this study contributes to shedding light on the possibility of combining empathy education with home economics education by presenting newly presented elements of empathy and brain-based education principles in home economics education.

Design and Implementation of BNN based Human Identification and Motion Classification System Using CW Radar (연속파 레이다를 활용한 이진 신경망 기반 사람 식별 및 동작 분류 시스템 설계 및 구현)

  • Kim, Kyeong-min;Kim, Seong-jin;NamKoong, Ho-jung;Jung, Yun-ho
    • Journal of Advanced Navigation Technology
    • /
    • v.26 no.4
    • /
    • pp.211-218
    • /
    • 2022
  • Continuous wave (CW) radar has the advantage of reliability and accuracy compared to other sensors such as camera and lidar. In addition, binarized neural network (BNN) has a characteristic that dramatically reduces memory usage and complexity compared to other deep learning networks. Therefore, this paper proposes binarized neural network based human identification and motion classification system using CW radar. After receiving a signal from CW radar, a spectrogram is generated through a short-time Fourier transform (STFT). Based on this spectrogram, we propose an algorithm that detects whether a person approaches a radar. Also, we designed an optimized BNN model that can support the accuracy of 90.0% for human identification and 98.3% for motion classification. In order to accelerate BNN operation, we designed BNN hardware accelerator on field programmable gate array (FPGA). The accelerator was implemented with 1,030 logics, 836 registers, and 334.904 Kbit block memory, and it was confirmed that the real-time operation was possible with a total calculation time of 6 ms from inference to transferring result.

Ontology-Based Process-Oriented Knowledge Map Enabling Referential Navigation between Knowledge (지식 간 상호참조적 네비게이션이 가능한 온톨로지 기반 프로세스 중심 지식지도)

  • Yoo, Kee-Dong
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.2
    • /
    • pp.61-83
    • /
    • 2012
  • A knowledge map describes the network of related knowledge into the form of a diagram, and therefore underpins the structure of knowledge categorizing and archiving by defining the relationship of the referential navigation between knowledge. The referential navigation between knowledge means the relationship of cross-referencing exhibited when a piece of knowledge is utilized by a user. To understand the contents of the knowledge, a user usually requires additionally information or knowledge related with each other in the relation of cause and effect. This relation can be expanded as the effective connection between knowledge increases, and finally forms the network of knowledge. A network display of knowledge using nodes and links to arrange and to represent the relationship between concepts can provide a more complex knowledge structure than a hierarchical display. Moreover, it can facilitate a user to infer through the links shown on the network. For this reason, building a knowledge map based on the ontology technology has been emphasized to formally as well as objectively describe the knowledge and its relationships. As the necessity to build a knowledge map based on the structure of the ontology has been emphasized, not a few researches have been proposed to fulfill the needs. However, most of those researches to apply the ontology to build the knowledge map just focused on formally expressing knowledge and its relationships with other knowledge to promote the possibility of knowledge reuse. Although many types of knowledge maps based on the structure of the ontology were proposed, no researches have tried to design and implement the referential navigation-enabled knowledge map. This paper addresses a methodology to build the ontology-based knowledge map enabling the referential navigation between knowledge. The ontology-based knowledge map resulted from the proposed methodology can not only express the referential navigation between knowledge but also infer additional relationships among knowledge based on the referential relationships. The most highlighted benefits that can be delivered by applying the ontology technology to the knowledge map include; formal expression about knowledge and its relationships with others, automatic identification of the knowledge network based on the function of self-inference on the referential relationships, and automatic expansion of the knowledge-base designed to categorize and store knowledge according to the network between knowledge. To enable the referential navigation between knowledge included in the knowledge map, and therefore to form the knowledge map in the format of a network, the ontology must describe knowledge according to the relation with the process and task. A process is composed of component tasks, while a task is activated after any required knowledge is inputted. Since the relation of cause and effect between knowledge can be inherently determined by the sequence of tasks, the referential relationship between knowledge can be circuitously implemented if the knowledge is modeled to be one of input or output of each task. To describe the knowledge with respect to related process and task, the Protege-OWL, an editor that enables users to build ontologies for the Semantic Web, is used. An OWL ontology-based knowledge map includes descriptions of classes (process, task, and knowledge), properties (relationships between process and task, task and knowledge), and their instances. Given such an ontology, the OWL formal semantics specifies how to derive its logical consequences, i.e. facts not literally present in the ontology, but entailed by the semantics. Therefore a knowledge network can be automatically formulated based on the defined relationships, and the referential navigation between knowledge is enabled. To verify the validity of the proposed concepts, two real business process-oriented knowledge maps are exemplified: the knowledge map of the process of 'Business Trip Application' and 'Purchase Management'. By applying the 'DL-Query' provided by the Protege-OWL as a plug-in module, the performance of the implemented ontology-based knowledge map has been examined. Two kinds of queries to check whether the knowledge is networked with respect to the referential relations as well as the ontology-based knowledge network can infer further facts that are not literally described were tested. The test results show that not only the referential navigation between knowledge has been correctly realized, but also the additional inference has been accurately performed.

A Study of Anomaly Detection for ICT Infrastructure using Conditional Multimodal Autoencoder (ICT 인프라 이상탐지를 위한 조건부 멀티모달 오토인코더에 관한 연구)

  • Shin, Byungjin;Lee, Jonghoon;Han, Sangjin;Park, Choong-Shik
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.3
    • /
    • pp.57-73
    • /
    • 2021
  • Maintenance and prevention of failure through anomaly detection of ICT infrastructure is becoming important. System monitoring data is multidimensional time series data. When we deal with multidimensional time series data, we have difficulty in considering both characteristics of multidimensional data and characteristics of time series data. When dealing with multidimensional data, correlation between variables should be considered. Existing methods such as probability and linear base, distance base, etc. are degraded due to limitations called the curse of dimensions. In addition, time series data is preprocessed by applying sliding window technique and time series decomposition for self-correlation analysis. These techniques are the cause of increasing the dimension of data, so it is necessary to supplement them. The anomaly detection field is an old research field, and statistical methods and regression analysis were used in the early days. Currently, there are active studies to apply machine learning and artificial neural network technology to this field. Statistically based methods are difficult to apply when data is non-homogeneous, and do not detect local outliers well. The regression analysis method compares the predictive value and the actual value after learning the regression formula based on the parametric statistics and it detects abnormality. Anomaly detection using regression analysis has the disadvantage that the performance is lowered when the model is not solid and the noise or outliers of the data are included. There is a restriction that learning data with noise or outliers should be used. The autoencoder using artificial neural networks is learned to output as similar as possible to input data. It has many advantages compared to existing probability and linear model, cluster analysis, and map learning. It can be applied to data that does not satisfy probability distribution or linear assumption. In addition, it is possible to learn non-mapping without label data for teaching. However, there is a limitation of local outlier identification of multidimensional data in anomaly detection, and there is a problem that the dimension of data is greatly increased due to the characteristics of time series data. In this study, we propose a CMAE (Conditional Multimodal Autoencoder) that enhances the performance of anomaly detection by considering local outliers and time series characteristics. First, we applied Multimodal Autoencoder (MAE) to improve the limitations of local outlier identification of multidimensional data. Multimodals are commonly used to learn different types of inputs, such as voice and image. The different modal shares the bottleneck effect of Autoencoder and it learns correlation. In addition, CAE (Conditional Autoencoder) was used to learn the characteristics of time series data effectively without increasing the dimension of data. In general, conditional input mainly uses category variables, but in this study, time was used as a condition to learn periodicity. The CMAE model proposed in this paper was verified by comparing with the Unimodal Autoencoder (UAE) and Multi-modal Autoencoder (MAE). The restoration performance of Autoencoder for 41 variables was confirmed in the proposed model and the comparison model. The restoration performance is different by variables, and the restoration is normally well operated because the loss value is small for Memory, Disk, and Network modals in all three Autoencoder models. The process modal did not show a significant difference in all three models, and the CPU modal showed excellent performance in CMAE. ROC curve was prepared for the evaluation of anomaly detection performance in the proposed model and the comparison model, and AUC, accuracy, precision, recall, and F1-score were compared. In all indicators, the performance was shown in the order of CMAE, MAE, and AE. Especially, the reproduction rate was 0.9828 for CMAE, which can be confirmed to detect almost most of the abnormalities. The accuracy of the model was also improved and 87.12%, and the F1-score was 0.8883, which is considered to be suitable for anomaly detection. In practical aspect, the proposed model has an additional advantage in addition to performance improvement. The use of techniques such as time series decomposition and sliding windows has the disadvantage of managing unnecessary procedures; and their dimensional increase can cause a decrease in the computational speed in inference.The proposed model has characteristics that are easy to apply to practical tasks such as inference speed and model management.

Auxiliary Reinforcement Method for the Safety of Tunnelling Face (터널 막장안정성에 따른 보강공법 적용)

  • Kim, Chang-Yong;Park, Chi-Hyun;Bae, Gyu-Jin;Hong, Sung-Wan;Oh, Myung-Ryul
    • Journal of Korean Tunnelling and Underground Space Association
    • /
    • v.2 no.2
    • /
    • pp.11-21
    • /
    • 2000
  • Tunnelling has been created as a great extent in view of less land space available because the growth of population in metropolitan has been accelerated at a faster pace than the development of the cities. In tunnelling, it is often faced that measures are obliged to be taken without confirmation for such abnormality as diverged movement of surrounding rock mass, growing crack of shotcrete and yielding of rockbolts. In this case, it is usually said that the judgments of experienced engineers for the selection of measure are importance and allowed us to get over the situations in many construction sites. But decrease of such experienced engineers need us to develop the new system to assist the selection of measures for the abnormality without any experiences of similar tunnelling sites. In this study, After a lot of tunnelling reinforcement methods were surveyed and the detail application were studied, an expert system was developed to predict the safety of tunnel and choose proper tunnel reinforcement system using fuzzy quantification theory and fuzzy inference rule based on tunnel information database. The expert system developed in this study have two main parts named pre-module and post-module. Pre-module decides tunnel information imput items based on the tunnel face mapping information which can be easily obtained in-situ site. Then, using fuzzy quantification theory II, fuzzy membership function is composed and tunnel safety level is inferred through this membership function. The comparison result between the predicted reinforcement system level and measured ones was very similar. In-situ data were obtained in three tunnel sites including subway tunnel under Han river. This system will be very helpful to make the most of in-situ data and suggest proper applicability of tunnel reinforcement system developing more resonable tunnel support method from dependance of some experienced experts for the absent of guide.

  • PDF

Ontology-based User Customized Search Service Considering User Intention (온톨로지 기반의 사용자 의도를 고려한 맞춤형 검색 서비스)

  • Kim, Sukyoung;Kim, Gunwoo
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.4
    • /
    • pp.129-143
    • /
    • 2012
  • Recently, the rapid progress of a number of standardized web technologies and the proliferation of web users in the world bring an explosive increase of producing and consuming information documents on the web. In addition, most companies have produced, shared, and managed a huge number of information documents that are needed to perform their businesses. They also have discretionally raked, stored and managed a number of web documents published on the web for their business. Along with this increase of information documents that should be managed in the companies, the need of a solution to locate information documents more accurately among a huge number of information sources have increased. In order to satisfy the need of accurate search, the market size of search engine solution market is becoming increasingly expended. The most important functionality among much functionality provided by search engine is to locate accurate information documents from a huge information sources. The major metric to evaluate the accuracy of search engine is relevance that consists of two measures, precision and recall. Precision is thought of as a measure of exactness, that is, what percentage of information considered as true answer are actually such, whereas recall is a measure of completeness, that is, what percentage of true answer are retrieved as such. These two measures can be used differently according to the applied domain. If we need to exhaustively search information such as patent documents and research papers, it is better to increase the recall. On the other hand, when the amount of information is small scale, it is better to increase precision. Most of existing web search engines typically uses a keyword search method that returns web documents including keywords which correspond to search words entered by a user. This method has a virtue of locating all web documents quickly, even though many search words are inputted. However, this method has a fundamental imitation of not considering search intention of a user, thereby retrieving irrelevant results as well as relevant ones. Thus, it takes additional time and effort to set relevant ones out from all results returned by a search engine. That is, keyword search method can increase recall, while it is difficult to locate web documents which a user actually want to find because it does not provide a means of understanding the intention of a user and reflecting it to a progress of searching information. Thus, this research suggests a new method of combining ontology-based search solution with core search functionalities provided by existing search engine solutions. The method enables a search engine to provide optimal search results by inferenceing the search intention of a user. To that end, we build an ontology which contains concepts and relationships among them in a specific domain. The ontology is used to inference synonyms of a set of search keywords inputted by a user, thereby making the search intention of the user reflected into the progress of searching information more actively compared to existing search engines. Based on the proposed method we implement a prototype search system and test the system in the patent domain where we experiment on searching relevant documents associated with a patent. The experiment shows that our system increases the both recall and precision in accuracy and augments the search productivity by using improved user interface that enables a user to interact with our search system effectively. In the future research, we will study a means of validating the better performance of our prototype system by comparing other search engine solution and will extend the applied domain into other domains for searching information such as portal.

Inferential Structure and Reality Problem in Diagnosis of Oriental Medicine (한의 진단의 추론형식과 실재성)

  • Park, Geong-Mo;Choi, Seong-Hoon;Ahn, Gyu-Seok
    • Journal of The Association for Neo Medicine
    • /
    • v.2 no.1
    • /
    • pp.55-84
    • /
    • 1997
  • Inferential structure and reality problem is a serious issue to O.M.(oriental medicine). The study will analyze this issue through a philosophical and historical comparative study of W.M.M(Western modern medicine) and O.M. First, I presuppose some basic ideas. The first is the division of the 'the philosophy of medicine' and 'the medicine itself'. Second, there is a 'visibility' that discriminate between 'the abstractive concept' and 'the concrete object' in diagnostic terminology. The third is the separation of disease, the entity and disease, the phenomenon. Finally, the distinction between the cause of disease and the nature of disease. Through these basic concepts, this study will analyze O.M's diagnostic methodology, 'Pattern identification of the S.A.S(sign and symptom)'. The results are follows: 1. O.M's views disease as a phenomenon. So, the S.A.S, which is visible, is the disease itself. Tough the analysis and inference of the S.A.S, 證(zheng) the essence is derived. 2. 證(zheng) can be considered as 'the abstractive concept' reflecting the essence of a disease. 3. 證(zheng) is not arrived through causal sequence reasoning but rather by analogical reasoning. 4. 證(zheng) is 'the non-random correlative combination of S.A.S', pattern. These patterns secure the abstractive deduction in reality. that is, The causality, the positivism, the view of disease as entity, and anatomical knowledge are the traits peculiar to W.M.M. But, these properties can not be applied universally to every medical systems. Also, these properties do not indicate the superiority or inferiority of any medical system. 5. 證(zheng) summarizes the patients condition simultaneously with the S.A.S. However, 證(zheng) doesn't necessarily indicate the knowledge about the actual internal organ. That is, Early in O.M.'s history, the diagnostic terminologies including 證(zheng) were analogical reflections of a naive knowledge of internal organs and external environmental factors. Later, the naive knowledge in 證(zheng) changed int new nature, an abstractive concept. The confusion of the concept of disease, the indiscriminate acceptance of Western anatomical knowledge, and the O.M.'s theoretical evolution et are the challenge facing modern O.M. To find solutions, this study looks at the sequence of the birth of W.M.M. and then compares it's system with the O.M. system. The confusion of the concept of disease, the indiscriminate acceptance of Western anatomical knowledge, and the O.M.'s theoretical evolution et are the challenge facing modern O.M. To find solutions, this study looks at the sequence of the birth of W.M.M. and then compares it's system with the O.M. system. It is recommended that O.M. diagnostics should pay close attention to the ambiguity of the diagnostic methodology in order to further development. At present time, the concept and the system peculiar to O.M. can not be explained by common language. but O.M.'s practitioner can not persist in this manner an: longer. Along with the internal development of O.M., the adjustment of O.M.'s diagnostic terminology needs to be adopted.

  • PDF

A Study on Web-based Technology Valuation System (웹기반 지능형 기술가치평가 시스템에 관한 연구)

  • Sung, Tae-Eung;Jun, Seung-Pyo;Kim, Sang-Gook;Park, Hyun-Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.23-46
    • /
    • 2017
  • Although there have been cases of evaluating the value of specific companies or projects which have centralized on developed countries in North America and Europe from the early 2000s, the system and methodology for estimating the economic value of individual technologies or patents has been activated on and on. Of course, there exist several online systems that qualitatively evaluate the technology's grade or the patent rating of the technology to be evaluated, as in 'KTRS' of the KIBO and 'SMART 3.1' of the Korea Invention Promotion Association. However, a web-based technology valuation system, referred to as 'STAR-Value system' that calculates the quantitative values of the subject technology for various purposes such as business feasibility analysis, investment attraction, tax/litigation, etc., has been officially opened and recently spreading. In this study, we introduce the type of methodology and evaluation model, reference information supporting these theories, and how database associated are utilized, focusing various modules and frameworks embedded in STAR-Value system. In particular, there are six valuation methods, including the discounted cash flow method (DCF), which is a representative one based on the income approach that anticipates future economic income to be valued at present, and the relief-from-royalty method, which calculates the present value of royalties' where we consider the contribution of the subject technology towards the business value created as the royalty rate. We look at how models and related support information (technology life, corporate (business) financial information, discount rate, industrial technology factors, etc.) can be used and linked in a intelligent manner. Based on the classification of information such as International Patent Classification (IPC) or Korea Standard Industry Classification (KSIC) for technology to be evaluated, the STAR-Value system automatically returns meta data such as technology cycle time (TCT), sales growth rate and profitability data of similar company or industry sector, weighted average cost of capital (WACC), indices of industrial technology factors, etc., and apply adjustment factors to them, so that the result of technology value calculation has high reliability and objectivity. Furthermore, if the information on the potential market size of the target technology and the market share of the commercialization subject refers to data-driven information, or if the estimated value range of similar technologies by industry sector is provided from the evaluation cases which are already completed and accumulated in database, the STAR-Value is anticipated that it will enable to present highly accurate value range in real time by intelligently linking various support modules. Including the explanation of the various valuation models and relevant primary variables as presented in this paper, the STAR-Value system intends to utilize more systematically and in a data-driven way by supporting the optimal model selection guideline module, intelligent technology value range reasoning module, and similar company selection based market share prediction module, etc. In addition, the research on the development and intelligence of the web-based STAR-Value system is significant in that it widely spread the web-based system that can be used in the validation and application to practices of the theoretical feasibility of the technology valuation field, and it is expected that it could be utilized in various fields of technology commercialization.