• Title/Summary/Keyword: Model interpretability

Search Result 49, Processing Time 0.022 seconds

Explainable Artificial Intelligence (XAI) Surrogate Models for Chemical Process Design and Analysis (화학 공정 설계 및 분석을 위한 설명 가능한 인공지능 대안 모델)

  • Yuna Ko;Jonggeol Na
    • Korean Chemical Engineering Research
    • /
    • v.61 no.4
    • /
    • pp.542-549
    • /
    • 2023
  • Since the growing interest in surrogate modeling, there has been continuous research aimed at simulating nonlinear chemical processes using data-driven machine learning. However, the opaque nature of machine learning models, which limits their interpretability, poses a challenge for their practical application in industry. Therefore, this study aims to analyze chemical processes using Explainable Artificial Intelligence (XAI), a concept that improves interpretability while ensuring model accuracy. While conventional sensitivity analysis of chemical processes has been limited to calculating and ranking the sensitivity indices of variables, we propose a methodology that utilizes XAI to not only perform global and local sensitivity analysis, but also examine the interactions among variables to gain physical insights from the data. For the ammonia synthesis process, which is the target process of the case study, we set the temperature of the preheater leading to the first reactor and the split ratio of the cold shot to the three reactors as process variables. By integrating Matlab and Aspen Plus, we obtained data on ammonia production and the maximum temperatures of the three reactors while systematically varying the process variables. We then trained tree-based models and performed sensitivity analysis using the SHAP technique, one of the XAI methods, on the most accurate model. The global sensitivity analysis showed that the preheater temperature had the greatest effect, and the local sensitivity analysis provided insights for defining the ranges of process variables to improve productivity and prevent overheating. By constructing alternative models for chemical processes and using XAI for sensitivity analysis, this work contributes to providing both quantitative and qualitative feedback for process optimization.

Comparative Analysis for Real-Estate Price Index Prediction Models using Machine Learning Algorithms: LIME's Interpretability Evaluation (기계학습 알고리즘을 활용한 지역 별 아파트 실거래가격지수 예측모델 비교: LIME 해석력 검증)

  • Jo, Bo-Geun;Park, Kyung-Bae;Ha, Sung-Ho
    • The Journal of Information Systems
    • /
    • v.29 no.3
    • /
    • pp.119-144
    • /
    • 2020
  • Purpose Real estate usually takes charge of the highest proportion of physical properties which individual, organizations, and government hold and instability of real estate market affects the economic condition seriously for each economic subject. Consequently, practices for predicting the real estate market have attention for various reasons, such as financial investment, administrative convenience, and wealth management. Additionally, development of machine learning algorithms and computing hardware enhances the expectation for more precise and useful prediction models in real estate market. Design/methodology/approach In response to the demand, this paper aims to provide a framework for forecasting the real estate market with machine learning algorithms. The framework consists of demonstrating the prediction efficiency of each machine learning algorithm, interpreting the interior feature effects of prediction model with a state-of-art algorithm, LIME(Local Interpretable Model-agnostic Explanation), and comparing the results in different cities. Findings This research could not only enhance the academic base for information system and real estate fields, but also resolve information asymmetry on real estate market among economic subjects. This research revealed that macroeconomic indicators, real estate-related indicators, and Google Trends search indexes can predict real-estate prices quite well.

Knowledge-guided artificial intelligence technologies for decoding complex multiomics interactions in cells

  • Lee, Dohoon;Kim, Sun
    • Clinical and Experimental Pediatrics
    • /
    • v.65 no.5
    • /
    • pp.239-249
    • /
    • 2022
  • Cells survive and proliferate through complex interactions among diverse molecules across multiomics layers. Conventional experimental approaches for identifying these interactions have built a firm foundation for molecular biology, but their scalability is gradually becoming inadequate compared to the rapid accumulation of multiomics data measured by high-throughput technologies. Therefore, the need for data-driven computational modeling of interactions within cells has been highlighted in recent years. The complexity of multiomics interactions is primarily due to their nonlinearity. That is, their accurate modeling requires intricate conditional dependencies, synergies, or antagonisms between considered genes or proteins, which retard experimental validations. Artificial intelligence (AI) technologies, including deep learning models, are optimal choices for handling complex nonlinear relationships between features that are scalable and produce large amounts of data. Thus, they have great potential for modeling multiomics interactions. Although there exist many AI-driven models for computational biology applications, relatively few explicitly incorporate the prior knowledge within model architectures or training procedures. Such guidance of models by domain knowledge will greatly reduce the amount of data needed to train models and constrain their vast expressive powers to focus on the biologically relevant space. Therefore, it can enhance a model's interpretability, reduce spurious interactions, and prove its validity and utility. Thus, to facilitate further development of knowledge-guided AI technologies for the modeling of multiomics interactions, here we review representative bioinformatics applications of deep learning models for multiomics interactions developed to date by categorizing them by guidance mode.

Identification Methodology of FCM-based Fuzzy Model Using Particle Swarm Optimization (입자 군집 최적화를 이용한 FCM 기반 퍼지 모델의 동정 방법론)

  • Oh, Sung-Kwun;Kim, Wook-Dong;Park, Ho-Sung;Son, Myung-Hee
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.60 no.1
    • /
    • pp.184-192
    • /
    • 2011
  • In this study, we introduce a identification methodology for FCM-based fuzzy model. The two underlying design mechanisms of such networks involve Fuzzy C-Means (FCM) clustering method and Particle Swarm Optimization(PSO). The proposed algorithm is based on FCM clustering method for efficient processing of data and the optimization of model was carried out using PSO. The premise part of fuzzy rules does not construct as any fixed membership functions such as triangular, gaussian, ellipsoidal because we build up the premise part of fuzzy rules using FCM. As a result, the proposed model can lead to the compact architecture of network. In this study, as the consequence part of fuzzy rules, we are able to use four types of polynomials such as simplified, linear, quadratic, modified quadratic. In addition, a Weighted Least Square Estimation to estimate the coefficients of polynomials, which are the consequent parts of fuzzy model, can decouple each fuzzy rule from the other fuzzy rules. Therefore, a local learning capability and an interpretability of the proposed fuzzy model are improved. Also, the parameters of the proposed fuzzy model such as a fuzzification coefficient of FCM clustering, the number of clusters of FCM clustering, and the polynomial type of the consequent part of fuzzy rules are adjusted using PSO. The proposed model is illustrated with the use of Automobile Miles per Gallon(MPG) and Boston housing called Machine Learning dataset. A comparative analysis reveals that the proposed FCM-based fuzzy model exhibits higher accuracy and superb predictive capability in comparison to some previous models available in the literature.

EMD based hybrid models to forecast the KOSPI (코스피 예측을 위한 EMD를 이용한 혼합 모형)

  • Kim, Hyowon;Seong, Byeongchan
    • The Korean Journal of Applied Statistics
    • /
    • v.29 no.3
    • /
    • pp.525-537
    • /
    • 2016
  • The paper considers a hybrid model to analyze and forecast time series data based on an empirical mode decomposition (EMD) that accommodates complex characteristics of time series such as nonstationarity and nonlinearity. We aggregate IMFs using the concept of cumulative energy to improve the interpretability of intrinsic mode functions (IMFs) from EMD. We forecast aggregated IMFs and residue with a hybrid model that combines the ARIMA model and an exponential smoothing method (ETS). The proposed method is applied to forecast KOSPI time series and is compared to traditional forecast models. Aggregated IMFs and residue provide a convenience to interpret the short, medium and long term dynamics of the KOSPI. It is also observed that the hybrid model with ARIMA and ETS is superior to traditional and other types of hybrid models.

Structural Design of FCM-based Fuzzy Inference System : A Comparative Study of WLSE and LSE (FCM기반 퍼지추론 시스템의 구조 설계: WLSE 및 LSE의 비교 연구)

  • Park, Wook-Dong;Oh, Sung-Kwun;Kim, Hyun-Ki
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.59 no.5
    • /
    • pp.981-989
    • /
    • 2010
  • In this study, we introduce a new architecture of fuzzy inference system. In the fuzzy inference system, we use Fuzzy C-Means clustering algorithm to form the premise part of the rules. The membership functions standing in the premise part of fuzzy rules do not assume any explicit functional forms, but for any input the resulting activation levels of such radial basis functions directly depend upon the distance between data points by means of the Fuzzy C-Means clustering. As the consequent part of fuzzy rules of the fuzzy inference system (being the local model representing input output relation in the corresponding sub-space), four types of polynomial are considered, namely constant, linear, quadratic and modified quadratic. This offers a significant level of design flexibility as each rule could come with a different type of the local model in its consequence. Either the Least Square Estimator (LSE) or the weighted Least Square Estimator (WLSE)-based learning is exploited to estimate the coefficients of the consequent polynomial of fuzzy rules. In fuzzy modeling, complexity and interpretability (or simplicity) as well as accuracy of the obtained model are essential design criteria. The performance of the fuzzy inference system is directly affected by some parameters such as e.g., the fuzzification coefficient used in the FCM, the number of rules(clusters) and the order of polynomial in the consequent part of the rules. Accordingly we can obtain preferred model structure through an adjustment of such parameters of the fuzzy inference system. Moreover the comparative experimental study between WLSE and LSE is analyzed according to the change of the number of clusters(rules) as well as polynomial type. The superiority of the proposed model is illustrated and also demonstrated with the use of Automobile Miles per Gallon(MPG), Boston housing called Machine Learning dataset, and Mackey-glass time series dataset.

Development of an Automatic Classification Model for Construction Site Photos with Semantic Analysis based on Korean Construction Specification (표준시방서 기반의 의미론적 분석을 반영한 건설 현장 사진 자동 분류 모델 개발)

  • Park, Min-Geon;Kim, Kyung-Hwan
    • Korean Journal of Construction Engineering and Management
    • /
    • v.25 no.3
    • /
    • pp.58-67
    • /
    • 2024
  • In the era of the fourth industrial revolution, data plays a vital role in enhancing the productivity of industries. To advance digitalization in the construction industry, which suffers from a lack of available data, this study proposes a model that classifies construction site photos by work types. Unlike traditional image classification models that solely rely on visual data, the model in this study includes semantic analysis of construction work types. This is achieved by extracting the significance of relationships between objects and work types from the standard construction specification. These relationships are then used to enhance the classification process by correlating them with objects detected in photos. This model improves the interpretability and reliability of classification results, offering convenience to field operators in photo categorization tasks. Additionally, the model's practical utility has been validated through integration into a classification program. As a result, this study is expected to contribute to the digitalization of the construction industry.

Hourly electricity demand forecasting based on innovations state space exponential smoothing models (이노베이션 상태공간 지수평활 모형을 이용한 시간별 전력 수요의 예측)

  • Won, Dayoung;Seong, Byeongchan
    • The Korean Journal of Applied Statistics
    • /
    • v.29 no.4
    • /
    • pp.581-594
    • /
    • 2016
  • We introduce innovations state space exponential smoothing models (ISS-ESM) that can analyze time series with multiple seasonal patterns. Especially, in order to control complex structure existing in the multiple patterns, the model equations use a matrix consisting of seasonal updating parameters. It enables us to group the seasonal parameters according to their similarity. Because of the grouped parameters, we can accomplish the principle of parsimony. Further, the ISS-ESM can potentially accommodate any number of multiple seasonal patterns. The models are applied to predict electricity demand in Korea that is observed on hourly basis, and we compare their performance with that of the traditional exponential smoothing methods. It is observed that the ISS-ESM are superior to the traditional methods in terms of the prediction and the interpretability of seasonal patterns.

A Securities Company's Customer Churn Prediction Model and Causal Inference with SHAP Value (증권 금융 상품 거래 고객의 이탈 예측 및 원인 추론)

  • Na, Kwangtek;Lee, Jinyoung;Kim, Eunchan;Lee, Hyochan
    • The Journal of Bigdata
    • /
    • v.5 no.2
    • /
    • pp.215-229
    • /
    • 2020
  • The interest in machine learning is growing in all industries, but it is difficult to apply it to real-world tasks because of inexplicability. This paper introduces a case of developing a financial customer churn prediction model for a securities company, and introduces the research results on an attempt to develop a machine learning model that can be explained using the SHAP Value methodology and derivation of interpretability. In this study, a total of six customer churn models are compared and analyzed, and the cause of customer churn is inferred through the classification and data analysis of SHAP Value and the type of customer asset change. Based on the results of this study, it would be possible to use it as a basis for comprehensive judgment, such as using the Value of the deviation prediction result that can infer the cause of the marketing manager's actual customer marketing in the future and establishing a target marketing strategy for each customer.

Hourly Prediction of Particulate Matter (PM2.5) Concentration Using Time Series Data and Random Forest (시계열 데이터와 랜덤 포레스트를 활용한 시간당 초미세먼지 농도 예측)

  • Lee, Deukwoo;Lee, Soowon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.9 no.4
    • /
    • pp.129-136
    • /
    • 2020
  • PM2.5 which is a very tiny air particulate matter even smaller than PM10 has been issued in the environmental problem. Since PM2.5 can cause eye diseases or respiratory problems and infiltrate even deep blood vessels in the brain, it is important to predict PM2.5. However, it is difficult to predict PM2.5 because there is no clear explanation yet regarding the creation and the movement of PM2.5. Thus, prediction methods which not only predict PM2.5 accurately but also have the interpretability of the result are needed. To predict hourly PM2.5 of Seoul city, we propose a method using random forest with the adjusted bootstrap number from the time series ground data preprocessed on different sources. With this method, the prediction model can be trained uniformly on hourly information and the result has the interpretability. To evaluate the prediction performance, we conducted comparative experiments. As a result, the performance of the proposed method was superior against other models in all labels. Also, the proposed method showed the importance of the variables regarding the creation of PM2.5 and the effect of China.