• Title/Summary/Keyword: Machine Learning Procedure

Search Result 116, Processing Time 0.029 seconds

One-probe P300 based concealed information test with machine learning (기계학습을 이용한 단일 관련자극 P300기반 숨김정보검사)

  • Hyuk Kim;Hyun-Taek Kim
    • Korean Journal of Cognitive Science
    • /
    • v.35 no.1
    • /
    • pp.49-95
    • /
    • 2024
  • Polygraph examination, statement validity analysis and P300-based concealed information test are major three examination tools, which are use to determine a person's truthfulness and credibility in criminal procedure. Although polygraph examination is most common in criminal procedure, but it has little admissibility of evidence due to the weakness of scientific basis. In 1990s to support the weakness of scientific basis about polygraph, Farwell and Donchin proposed the P300-based concealed information test technique. The P300-based concealed information test has two strong points. First, the P300-based concealed information test is easy to conduct with polygraph. Second, the P300-based concealed information test has plentiful scientific basis. Nevertheless, the utilization of P300-based concealed information test is infrequent, because of the quantity of probe stimulus. The probe stimulus contains closed information that is relevant to the crime or other investigated situation. In tradition P300-based concealed information test protocol, three or more probe stimuli are necessarily needed. But it is hard to acquire three or more probe stimuli, because most of the crime relevant information is opened in investigative situation. In addition, P300-based concealed information test uses oddball paradigm, and oddball paradigm makes imbalance between the number of probe and irrelevant stimulus. Thus, there is a possibility that the unbalanced number of probe and irrelevant stimulus caused systematic underestimation of P300 amplitude of irrelevant stimuli. To overcome the these two limitation of P300-based concealed information test, one-probe P300-based concealed information test protocol is explored with various machine learning algorithms. According to this study, parameters of the modified one-probe protocol are as follows. In the condition of female and male face stimuli, the duration of stimuli are encouraged 400ms, the repetition of stimuli are encouraged 60 times, the analysis method of P300 amplitude is encouraged peak to peak method, the cut-off of guilty condition is encouraged 90% and the cut-off of innocent condition is encouraged 30%. In the condition of two-syllable word stimulus, the duration of stimulus is encouraged 300ms, the repetition of stimulus is encouraged 60 times, the analysis method of P300 amplitude is encouraged peak to peak method, the cut-off of guilty condition is encouraged 90% and the cut-off of innocent condition is encouraged 30%. It was also conformed that the logistic regression (LR), linear discriminant analysis (LDA), K Neighbors (KNN) algorithms were probable methods for analysis of P300 amplitude. The one-probe P300-based concealed information test with machine learning protocol is helpful to increase utilization of P300-based concealed information test, and supports to determine a person's truthfulness and credibility with the polygraph examination in criminal procedure.

Predicting flux of forward osmosis membrane module using deep learning (딥러닝을 이용한 정삼투 막모듈의 플럭스 예측)

  • Kim, Jaeyoon;Jeon, Jongmin;Kim, Noori;Kim, Suhan
    • Journal of Korean Society of Water and Wastewater
    • /
    • v.35 no.1
    • /
    • pp.93-100
    • /
    • 2021
  • Forward osmosis (FO) process is a chemical potential driven process, where highly concentrated draw solution (DS) is used to take water through semi-permeable membrane from feed solution (FS) with lower concentration. Recently, commercial FO membrane modules have been developed so that full-scale FO process can be applied to seawater desalination or water reuse. In order to design a real-scale FO plant, the performance prediction of FO membrane modules installed in the plant is essential. Especially, the flux prediction is the most important task because the amount of diluted draw solution and concentrate solution flowing out of FO modules can be expected from the flux. Through a previous study, a theoretical based FO module model to predict flux was developed. However it needs an intensive numerical calculation work and a fitting process to reflect a complex module geometry. The idea of this work is to introduce deep learning to predict flux of FO membrane modules using 116 experimental data set, which include six input variables (flow rate, pressure, and ion concentration of DS and FS) and one output variable (flux). The procedure of optimizing a deep learning model to minimize prediction error and overfitting problem was developed and tested. The optimized deep learning model (error of 3.87%) was found to predict flux better than the theoretical based FO module model (error of 10.13%) in the data set which were not used in machine learning.

Multi Label Deep Learning classification approach for False Data Injection Attacks in Smart Grid

  • Prasanna Srinivasan, V;Balasubadra, K;Saravanan, K;Arjun, V.S;Malarkodi, S
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.6
    • /
    • pp.2168-2187
    • /
    • 2021
  • The smart grid replaces the traditional power structure with information inventiveness that contributes to a new physical structure. In such a field, malicious information injection can potentially lead to extreme results. Incorrect, FDI attacks will never be identified by typical residual techniques for false data identification. Most of the work on the detection of FDI attacks is based on the linearized power system model DC and does not detect attacks from the AC model. Also, the overwhelming majority of current FDIA recognition approaches focus on FDIA, whilst significant injection location data cannot be achieved. Building on the continuous developments in deep learning, we propose a Deep Learning based Locational Detection technique to continuously recognize the specific areas of FDIA. In the development area solver gap happiness is a False Data Detector (FDD) that incorporates a Convolutional Neural Network (CNN). The FDD is established enough to catch the fake information. As a multi-label classifier, the following CNN is utilized to evaluate the irregularity and cooccurrence dependency of power flow calculations due to the possible attacks. There are no earlier statistical assumptions in the architecture proposed, as they are "model-free." It is also "cost-accommodating" since it does not alter the current FDD framework and it is only several microseconds on a household computer during the identification procedure. We have shown that ANN-MLP, SVM-RBF, and CNN can conduct locational detection under different noise and attack circumstances through broad experience in IEEE 14, 30, 57, and 118 bus systems. Moreover, the multi-name classification method used successfully improves the precision of the present identification.

Under Sampling for Imbalanced Data using Minor Class based SVM (MCSVM) in Semiconductor Process (MCSVM을 이용한 반도체 공정데이터의 과소 추출 기법)

  • Pak, Sae-Rom;Kim, Jun Seok;Park, Cheong-Sool;Park, Seung Hwan;Baek, Jun-Geol
    • Journal of Korean Institute of Industrial Engineers
    • /
    • v.40 no.4
    • /
    • pp.404-414
    • /
    • 2014
  • Yield prediction is important to manage semiconductor quality. Many researches with machine learning algorithms such as SVM (support vector machine) are conducted to predict yield precisely. However, yield prediction using SVM is hard because extremely imbalanced and big data are generated by final test procedure in semiconductor manufacturing process. Using SVM algorithm with imbalanced data sometimes cause unnecessary support vectors from major class because of unselected support vectors from minor class. So, decision boundary at target class can be overwhelmed by effect of observations in major class. For this reason, we propose a under-sampling method with minor class based SVM (MCSVM) which overcomes the limitations of ordinary SVM algorithm. MCSVM constructs the model that fixes some of data from minor class as support vectors, and they can be good samples representing the nature of target class. Several experimental studies with using the data sets from UCI and real manufacturing process represent that our proposed method performs better than existing sampling methods.

FAULT DIAGNOSIS OF ROLLING BEARINGS USING UNSUPERVISED DYNAMIC TIME WARPING-AIDED ARTIFICIAL IMMUNE SYSTEM

  • LUCAS VERONEZ GOULART FERREIRA;LAXMI RATHOUR;DEVIKA DABKE;FABIO ROBERTO CHAVARETTE;VISHNU NARAYAN MISHRA
    • Journal of applied mathematics & informatics
    • /
    • v.41 no.6
    • /
    • pp.1257-1274
    • /
    • 2023
  • Rotating machines heavily rely on an intricate network of interconnected sub-components, with bearing failures accounting for a substantial proportion (40% to 90%) of all such failures. To address this issue, intelligent algorithms have been developed to evaluate vibrational signals and accurately detect faults, thereby reducing the reliance on expert knowledge and lowering maintenance costs. Within the field of machine learning, Artificial Immune Systems (AIS) have exhibited notable potential, with applications ranging from malware detection in computer systems to fault detection in bearings, which is the primary focus of this study. In pursuit of this objective, we propose a novel procedure for detecting novel instances of anomalies in varying operating conditions, utilizing only the signals derived from the healthy state of the analyzed machine. Our approach incorporates AIS augmented by Dynamic Time Warping (DTW). The experimental outcomes demonstrate that the AIS-DTW method yields a considerable improvement in anomaly detection rates (up to 53.83%) compared to the conventional AIS. In summary, our findings indicate that our method represents a significant advancement in enhancing the resilience of AIS-based novelty detection, thereby bolstering the reliability of rotating machines and reducing the need for expertise in bearing fault detection.

Suggestions for the Development of RegTech Based Ontology and Deep Learning Technology to Interpret Capital Market Regulations (레그테크 기반의 자본시장 규제 해석 온톨로지 및 딥러닝 기술 개발을 위한 제언)

  • Choi, Seung Uk;Kwon, Oh Byung
    • The Journal of Information Systems
    • /
    • v.30 no.1
    • /
    • pp.65-84
    • /
    • 2021
  • Purpose Based on the development of artificial intelligence and big data technologies, the RegTech has been emerged to reduce regulatory costs and to enable efficient supervision by regulatory bodies. The word RegTech is a combination of regulation and technology, which means using the technological methods to facilitate the implementation of regulations and to make efficient surveillance and supervision of regulations. The purpose of this study is to describe the recent adoption of RegTech and to provide basic examples of applying RegTech to capital market regulations. Design/methodology/approach English-based ontology and deep learning technologies are quite developed in practice, and it will not be difficult to expand it to European or Latin American languages that are grammatically similar to English. However, it is not easy to use it in most Asian languages such as Korean, which have different grammatical rules. In addition, in the early stages of adoption, companies, financial institutions and regulators will not be familiar with this machine-based reporting system. There is a need to establish an ecosystem which facilitates the adoption of RegTech by consulting and supporting the stakeholders. In this paper, we provide a simple example that shows a procedure of applying RegTech to recognize and interpret Korean language-based capital market regulations. Specifically, we present the process of converting sentences in regulations into a meta-language through the morpheme analyses. We next conduct deep learning analyses to determine whether a regulatory sentence exists in each regulatory paragraph. Findings This study illustrates the applicability of RegTech-based ontology and deep learning technologies in Korean-based capital market regulations.

A Study on the Construction of Stable Clustering by Minimizing the Order Bias (순서 바이어스 최소화에 의한 안정적 클러스터링 구축에 관한 연구)

  • Lee, Gye-Seong
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.6
    • /
    • pp.1571-1580
    • /
    • 1999
  • When a hierarchical structure is derived from data set for data mining and machine learning, using a conceptual clustering algorithm, one of the unsupervised learning paradigms, it is not unusual to have a different set of outcomes with respect to the order of processing data objects. To overcome this problem, the first classification process is proceeded to construct an initial partition. The partition is expected to imply the possible range in the number of final classes. We apply center sorting to the data objects in the classes of the partition for new data ordering and build a new partition using ITERATE clustering procedure. We developed an algorithm, REIT that leads to the final partition with stable and best partition score. A number of experiments were performed to show the minimization of order bias effects using the algorithm.

  • PDF

Optimal Placement of Measurement Using GAs in Harmonic State Estimation of Power System (전력시스템 고조파 상태 춘정에서 GA를 미용한 최적 측정위치 선정)

  • 정형환;왕용필;박희철;안병철
    • The Transactions of the Korean Institute of Electrical Engineers A
    • /
    • v.52 no.8
    • /
    • pp.471-480
    • /
    • 2003
  • The design of a measurement system to perform Harmonic State Estimation (HSE) is a very complex problem. Among the reasons for its complexity are the system size, conflicting requirements of estimator accuracy, reliability in the presence of transducer noise and data communication failures, adaptability to change in the network topology and cost minimization. In particular, the number of harmonic instruments available is always limited. Therefore, a systematic procedure is needed to design the optimal placement of measurement points. This paper presents a new HSE algorithm which is based on an optimal placement of measurement points using Genetic Algorithms (GAs) which is widely used in areas such as: optimization of the objective function, learning of neural networks, tuning of fuzzy membership functions, machine learning, system identification and control. This HSE has been applied to the Simulation Test Power System for the validation of the new HSE algorithm. The study results have indicated an economical and effective method for optimal placement of measurement points using Genetic Algorithms (GAs) in the Harmonic State Estimation (HSE).

The Effect of Bias in Data Set for Conceptual Clustering Algorithms

  • Lee, Gye Sung
    • International journal of advanced smart convergence
    • /
    • v.8 no.3
    • /
    • pp.46-53
    • /
    • 2019
  • When a partitioned structure is derived from a data set using a clustering algorithm, it is not unusual to have a different set of outcomes when it runs with a different order of data. This problem is known as the order bias problem. Many algorithms in machine learning fields try to achieve optimized result from available training and test data. Optimization is determined by an evaluation function which has also a tendency toward a certain goal. It is inevitable to have a tendency in the evaluation function both for efficiency and for consistency in the result. But its preference for a specific goal in the evaluation function may sometimes lead to unfavorable consequences in the final result of the clustering. To overcome this bias problems, the first clustering process proceeds to construct an initial partition. The initial partition is expected to imply the possible range in the number of final clusters. We apply the data centric sorting to the data objects in the clusters of the partition to rearrange them in a new order. The same clustering procedure is reapplied to the newly arranged data set to build a new partition. We have developed an algorithm that reduces bias effect resulting from how data is fed into the algorithm. Experiment results have been presented to show that the algorithm helps minimize the order bias effects. We have also shown that the current evaluation measure used for the clustering algorithm is biased toward favoring a smaller number of clusters and a larger size of clusters as a result.

A new Design of Granular-oriented Self-organizing Polynomial Neural Networks (입자화 중심 자기구성 다항식 신경 회로망의 새로운 설계)

  • Oh, Sung-Kwun;Park, Ho-Sung
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.61 no.2
    • /
    • pp.312-320
    • /
    • 2012
  • In this study, we introduce a new design methodology of a granular-oriented self-organizing polynomial neural networks (GoSOPNNs) that is based on multi-layer perceptron with Context-based Polynomial Neurons (CPNs) or Polynomial Neurons (PNs). In contrast to the typical architectures encountered in polynomial neural networks (PNN), our main objective is to develop a methodological design strategy of GoSOPNNs as follows : (a) The 1st layer of the proposed network consists of Context-based Polynomial Neuron (CPN). In here, CPN is fully reflective of the structure encountered in numeric data which are granulated with the aid of Context-based Fuzzy C-Means (C-FCM) clustering method. The context-based clustering supporting the design of information granules is completed in the space of the input data while the build of the clusters is guided by a collection of some predefined fuzzy sets (so-called contexts) defined in the output space. (b) The proposed design procedure being applied at each layer of GoSOPNN leads to the selection of preferred nodes of the network (CPNs or PNs) whose local characteristics (such as the number of contexts, the number of clusters, a collection of the specific subset of input variables, and the order of the polynomial) can be easily adjusted. These options contribute to the flexibility as well as simplicity and compactness of the resulting architecture of the network. For the evaluation of performance of the proposed GoSOPNN network, we describe a detailed characteristic of the proposed model using a well-known learning machine data(Automobile Miles Per Gallon Data, Boston Housing Data, Medical Image System Data).