• 제목/요약/키워드: backpropagation method

검색결과 262건 처리시간 0.029초

FIGURE ALPHABET HYPOTHESIS INSPIRED NEURAL NETWORK RECOGNITION MODEL

  • Ohira, Ryoji;Saiki, Kenji;Nagao, Tomoharu
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 한국방송공학회 2009년도 IWAIT
    • /
    • pp.547-550
    • /
    • 2009
  • The object recognition mechanism of human being is not well understood yet. On research of animal experiment using an ape, however, neurons that respond to simple shape (e.g. circle, triangle, square and so on) were found. And Hypothesis has been set up as human being may recognize object as combination of such simple shapes. That mechanism is called Figure Alphabet Hypothesis, and those simple shapes are called Figure Alphabet. As one way to research object recognition algorithm, we focused attention to this Figure Alphabet Hypothesis. Getting idea from it, we proposed the feature extraction algorithm for object recognition. In this paper, we described recognition of binarized images of multifont alphabet characters by the recognition model which combined three-layered neural network in the feature extraction algorithm. First of all, we calculated the difference between the learning image data set and the template by the feature extraction algorithm. The computed finite difference is a feature quantity of the feature extraction algorithm. We had it input the feature quantity to the neural network model and learn by backpropagation (BP method). We had the recognition model recognize the unknown image data set and found the correct answer rate. To estimate the performance of the contriving recognition model, we had the unknown image data set recognized by a conventional neural network. As a result, the contriving recognition model showed a higher correct answer rate than a conventional neural network model. Therefore the validity of the contriving recognition model could be proved. We'll plan the research a recognition of natural image by the contriving recognition model in the future.

  • PDF

Recognition of Online Handwritten Digit using Zernike Moment and Neural Network (Zerinke 모멘트와 신경망을 이용한 온라인 필기체 숫자 인식)

  • Mun, Won-Ho;Choi, Yeon-Suk;Cha, Eui-Young
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 한국해양정보통신학회 2010년도 춘계학술대회
    • /
    • pp.205-208
    • /
    • 2010
  • We introduce a novel feature extraction scheme for online handwritten digit based on utilizing Zernike moment and angulation feature. The time sequential signal from mouse movement on the writing pad is described as a sequence of consecutive points on the x-y plane. So, we can create data-set which are successive and time-sequential pixel position data by preprocessing. Data preprocessed is used for Zernike moment and angulation feature extraction. this feature is scale-, translation-, and rotation-invariant. The extracted specific feature is fed to a BP(backpropagation) neural network, which in turn classifies it as one of the nine digits. In this paper, proposed method not noly show high recognition rate but also need less learning data for 200 handwritten digit data.

  • PDF

Parameter Extraction for Based on AR and Arrhythmia Classification through Deep Learning (AR 기반의 특징점 추출과 딥러닝을 통한 부정맥 분류)

  • Cho, Ik-sung;Kwon, Hyeog-soong
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • 제24권10호
    • /
    • pp.1341-1347
    • /
    • 2020
  • Legacy studies for classifying arrhythmia have been studied in order to improve the accuracy of classification, Neural Network, Fuzzy, Machine Learning, etc. In particular, deep learning is most frequently used for arrhythmia classification using error backpropagation algorithm by solving the limit of hidden layer number, which is a problem of neural network. In order to apply a deep learning model to an ECG signal, it is necessary to select an optimal model and parameters. In this paper, we propose parameter extraction based on AR and arrhythmia classification through a deep learning. For this purpose, the R-wave is detected in the ECG signal from which noise has been removed, QRS and RR interval is modelled. And then, the weights were learned by supervised learning method through deep learning and the model was evaluated by the verification data. The classification rate of PVC is evaluated through MIT-BIH arrhythmia database. The achieved scores indicate arrhythmia classification rate of over 97%.

A study on the construction of the quality prediction model by artificial neural intelligence through integrated learning of CAE-based data and experimental data in the injection molding process (사출성형공정에서 CAE 기반 품질 데이터와 실험 데이터의 통합 학습을 통한 인공지능 품질 예측 모델 구축에 대한 연구)

  • Lee, Jun-Han;Kim, Jong-Sun
    • Design & Manufacturing
    • /
    • 제15권4호
    • /
    • pp.24-31
    • /
    • 2021
  • In this study, an artificial neural network model was constructed to convert CAE analysis data into similar experimental data. In the analysis and experiment, the injection molding data for 50 conditions were acquired through the design of experiment and random selection method. The injection molding conditions and the weight, height, and diameter of the product derived from CAE results were used as the input parameters for learning of the convert model. Also the product qualities of experimental results were used as the output parameters for learning of the convert model. The accuracy of the convert model showed RMSE values of 0.06g, 0.03mm, and 0.03mm in weight, height, and diameter, respectively. As the next step, additional randomly selected conditions were created and CAE analysis was performed. Then, the additional CAE analysis data were converted to similar experimental data through the conversion model. An artificial neural network model was constructed to predict the quality of injection molded product by using converted similar experimental data and injection molding experiment data. The injection molding conditions were used as input parameters for learning of the predicted model and weight, height, and diameter of the product were used as output parameters for learning. As a result of evaluating the performance of the prediction model, the predicted weight, height, and diameter showed RMSE values of 0.11g, 0.03mm, and 0.05mm and in terms of quality criteria of the target product, all of them showed accurate results satisfying the criteria range.

A computational estimation model for the subgrade reaction modulus of soil improved with DCM columns

  • Dehghanbanadaki, Ali;Rashid, Ahmad Safuan A.;Ahmad, Kamarudin;Yunus, Nor Zurairahetty Mohd;Said, Khairun Nissa Mat
    • Geomechanics and Engineering
    • /
    • 제28권4호
    • /
    • pp.385-396
    • /
    • 2022
  • The accurate determination of the subgrade reaction modulus (Ks) of soil is an important factor for geotechnical engineers. This study estimated the Ks of soft soil improved with floating deep cement mixing (DCM) columns. A novel prediction model was developed that emphasizes the accuracy of identifying the most significant parameters of Ks. Several multi-layer perceptron (MLP) models that were trained using the Levenberg Marquardt (LM) backpropagation method were developed to estimate Ks. The models were trained using a reliable database containing the results of 36 physical modelling tests. The input parameters were the undrained shear strength of the DCM columns, undrained shear strength of soft soil, area improvement ratio and length-to-diameter ratio of the DCM columns. Grey wolf optimization (GWO) was coupled with the MLPs to improve the performance indices of the MLPs. Sensitivity tests were carried out to determine the importance of the input parameters for prediction of Ks. The results showed that both the MLP-LM and MLP-GWO methods showed high ability to predict Ks. However, it was shown that MLP-GWO (R = 0.9917, MSE = 0.28 (MN/m2/m)) performed better than MLP-LM (R =0.9126, MSE =6.1916 (MN/m2/m)). This proves the greater reliability of the proposed hybrid model of MLP-GWO in approximating the subgrade reaction modulus of soft soil improved with floating DCM columns. The results revealed that the undrained shear strength of the soil was the most effective factor for estimation of Ks.

Label Embedding for Improving Classification Accuracy UsingAutoEncoderwithSkip-Connections (다중 레이블 분류의 정확도 향상을 위한 스킵 연결 오토인코더 기반 레이블 임베딩 방법론)

  • Kim, Museong;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • 제27권3호
    • /
    • pp.175-197
    • /
    • 2021
  • Recently, with the development of deep learning technology, research on unstructured data analysis is being actively conducted, and it is showing remarkable results in various fields such as classification, summary, and generation. Among various text analysis fields, text classification is the most widely used technology in academia and industry. Text classification includes binary class classification with one label among two classes, multi-class classification with one label among several classes, and multi-label classification with multiple labels among several classes. In particular, multi-label classification requires a different training method from binary class classification and multi-class classification because of the characteristic of having multiple labels. In addition, since the number of labels to be predicted increases as the number of labels and classes increases, there is a limitation in that performance improvement is difficult due to an increase in prediction difficulty. To overcome these limitations, (i) compressing the initially given high-dimensional label space into a low-dimensional latent label space, (ii) after performing training to predict the compressed label, (iii) restoring the predicted label to the high-dimensional original label space, research on label embedding is being actively conducted. Typical label embedding techniques include Principal Label Space Transformation (PLST), Multi-Label Classification via Boolean Matrix Decomposition (MLC-BMaD), and Bayesian Multi-Label Compressed Sensing (BML-CS). However, since these techniques consider only the linear relationship between labels or compress the labels by random transformation, it is difficult to understand the non-linear relationship between labels, so there is a limitation in that it is not possible to create a latent label space sufficiently containing the information of the original label. Recently, there have been increasing attempts to improve performance by applying deep learning technology to label embedding. Label embedding using an autoencoder, a deep learning model that is effective for data compression and restoration, is representative. However, the traditional autoencoder-based label embedding has a limitation in that a large amount of information loss occurs when compressing a high-dimensional label space having a myriad of classes into a low-dimensional latent label space. This can be found in the gradient loss problem that occurs in the backpropagation process of learning. To solve this problem, skip connection was devised, and by adding the input of the layer to the output to prevent gradient loss during backpropagation, efficient learning is possible even when the layer is deep. Skip connection is mainly used for image feature extraction in convolutional neural networks, but studies using skip connection in autoencoder or label embedding process are still lacking. Therefore, in this study, we propose an autoencoder-based label embedding methodology in which skip connections are added to each of the encoder and decoder to form a low-dimensional latent label space that reflects the information of the high-dimensional label space well. In addition, the proposed methodology was applied to actual paper keywords to derive the high-dimensional keyword label space and the low-dimensional latent label space. Using this, we conducted an experiment to predict the compressed keyword vector existing in the latent label space from the paper abstract and to evaluate the multi-label classification by restoring the predicted keyword vector back to the original label space. As a result, the accuracy, precision, recall, and F1 score used as performance indicators showed far superior performance in multi-label classification based on the proposed methodology compared to traditional multi-label classification methods. This can be seen that the low-dimensional latent label space derived through the proposed methodology well reflected the information of the high-dimensional label space, which ultimately led to the improvement of the performance of the multi-label classification itself. In addition, the utility of the proposed methodology was identified by comparing the performance of the proposed methodology according to the domain characteristics and the number of dimensions of the latent label space.

Data Mining using Instance Selection in Artificial Neural Networks for Bankruptcy Prediction (기업부도예측을 위한 인공신경망 모형에서의 사례선택기법에 의한 데이터 마이닝)

  • Kim, Kyoung-jae
    • Journal of Intelligence and Information Systems
    • /
    • 제10권1호
    • /
    • pp.109-123
    • /
    • 2004
  • Corporate financial distress and bankruptcy prediction is one of the major application areas of artificial neural networks (ANNs) in finance and management. ANNs have showed high prediction performance in this area, but sometimes are confronted with inconsistent and unpredictable performance for noisy data. In addition, it may not be possible to train ANN or the training task cannot be effectively carried out without data reduction when the amount of data is so large because training the large data set needs much processing time and additional costs of collecting data. Instance selection is one of popular methods for dimensionality reduction and is directly related to data reduction. Although some researchers have addressed the need for instance selection in instance-based learning algorithms, there is little research on instance selection for ANN. This study proposes a genetic algorithm (GA) approach to instance selection in ANN for bankruptcy prediction. In this study, we use ANN supported by the GA to optimize the connection weights between layers and select relevant instances. It is expected that the globally evolved weights mitigate the well-known limitations of gradient descent algorithm of backpropagation algorithm. In addition, genetically selected instances will shorten the learning time and enhance prediction performance. This study will compare the proposed model with other major data mining techniques. Experimental results show that the GA approach is a promising method for instance selection in ANN.

  • PDF

Model Analysis of AI-Based Water Pipeline Improved Decision (AI기반 상수도시설 개량 의사결정 모델 분석)

  • Kim, Gi-Tae;Min, Byung-Won;Oh, Yong-Sun
    • Journal of Internet of Things and Convergence
    • /
    • 제8권5호
    • /
    • pp.11-16
    • /
    • 2022
  • As an interest in the development of artificial intelligence(AI) technology in the water supply sector increases, we have developed an AI algorithm that can predict improvement decision-making ratings through repetitive learning using the data of pipe condition evaluation results, and present the most reliable prediction model through a verification process. We have developed the algorithm that can predict pipe ratings by pre-processing 12 indirect evaluation items based on the 2020 Han River Basin's basic plan and applying the AI algorithm to update weighting factors through backpropagation. This method ensured that the concordance rate between the direct evaluation result value and the calculated result value through repetitive learning and verification was more than 90%. As a result of the algorithm accuracy verification process, it was confirmed that all water pipe type data were evenly distributed, and the more learning data, the higher prediction accuracy. If data from all across the country is collected, the reliability of the prediction technique for pipe ratings using AI algorithm will be improved, and therefore, it is expected that the AI algorithm will play a role in supporting decision-making in the objective evaluation of the condition of aging pipes.

On-Line Determination Steady State in Simulation Output (시뮬레이션 출력의 안정상태 온라인 결정에 관한 연구)

  • 이영해;정창식;경규형
    • Proceedings of the Korea Society for Simulation Conference
    • /
    • 한국시뮬레이션학회 1996년도 춘계학술대회
    • /
    • pp.1-3
    • /
    • 1996
  • 시뮬레이션 기법을 이용한 시스템의 분석에 있어서 실험의 자동화는 현재 많은 연구와 개발이 진행 중인 분야이다. 컴퓨터와 정보통신 시스템에 대한 시뮬레이션의 예를 들어 보면, 수많은 모델을 대한 시뮬레이션을 수행할 경우 자동화된 실험의 제어가 요구되고 있다. 시뮬레이션 수행회수, 수행길이, 데이터 수집방법 등과 관련하여 시뮬레이션 실험방법이 자동화가 되지 않으면, 시뮬레이션 실험에 필요한 시간과 인적 자원이 상당히 커지게 되며 출력데이터에 대한 분석에 있어서도 어려움이 따르게 된다. 시뮬레이션 실험방법을 자동화하면서 효율적인 시뮬레이션 출력분석을 위해서는 시뮬레이션을 수행하는 경우에 항상 발생하는 초기편의 (initial bias)를 제거하는 문제가 선결되어야 한다. 시뮬레이션 출력분석에 사용되는 데이터들이 초기편의를 반영하지 않는 안정상태에서 수집된 것이어야만 실제 시스템에 대한 올바른 해석이 가능하다. 실제로 시뮬레이션 출력분석과 관련하여 가장 중요하면서도 어려운 문제는 시뮬레이션의 출력데이터가 이루는 추계적 과정 (stochastic process)의 안정상태 평균과 이 평균에 대한 신뢰구간(confidence interval: c. i.)을 구하는 것이다. 한 신뢰구간에 포함되어 있는 정보는 의사결정자에게 얼마나 정확하게 평균을 추정할 구 있는지 알려 준다. 그러나, 신뢰구간을 구성하는 일은 하나의 시뮬레이션으로부터 얻어진 출력데이터가 일반적으로 비정체상태(nonstationary)이고 자동상관(autocorrelated)되어 있기 때문에, 전통적인 통계적인 기법을 직접적으로 이용할 수 없다. 이러한 문제를 해결하기 위해 시뮬레이션 출력데이터 분석기법이 사용된다.본 논문에서는 초기편의를 제거하기 위해서 필요한 출력데이터의 제거시점을 찾는 새로운 기법으로, 유클리드 거리(Euclidean distance: ED)를 이용한 방법과 현재 패턴 분류(pattern classification) 문제에 널리 사용 중인 역전파 신경망(backpropagation neural networks: BNN) 알고리듬을 이용하는 방법을 제시한다. 이 기법들은 대다수의 기존의 기법과는 달리 시험수행(pilot run)이 필요 없으며, 시뮬레이션의 단일수행(single run) 중에 제거시점을 결정할 수 있다. 제거시점과 관련된 기존 연구는 다음과 같다. 콘웨이방법은 현재의 데이터가 이후 데이터의 최대값이나 최소값이 아니면 이 데이터를 제거시점으로 결정하는데, 알고기듬 구조상 온라인으로 제거시점 결정이 불가능하다. 콘웨이방법이 알고리듬의 성격상 온라인이 불가능한 반면, 수정콘웨이방법 (Modified Conway Rule: MCR)은 현재의 데이터가 이전 데이터와 비교했을 때 최대값이나 최소값이 아닌 경우 현재의 데이터를 제거시점으로 결정하기 때문에 온라인이 가능하다. 평균교차방법(Crossings-of-the-Mean Rule: CMR)은 누적평균을 이용하면서 이 평균을 중심으로 관측치가 위에서 아래로, 또는 아래서 위로 교차하는 회수로 결정한다. 이 기법을 사용하려면 교차회수를 결정해야 하는데, 일반적으로 결정된 교차회수가 시스템에 상관없이 일반적으로 적용가능하지 않다는 문제점이 있다. 누적평균방법(Cumulative-Mean Rule: CMR2)은 여러 번의 시험수행을 통해서 얻어진 출력데이터에 대한 총누적평균(grand cumulative mean)을 그래프로 그린 다음, 안정상태인 점을 육안으로 결정한다. 이 방법은 여러 번의 시뮬레이션을 수행에서 얻어진 데이터들의 평균들에 대한 누적평균을 사용하기 매문에 온라인 제거시점 결정이 불가능하며, 작업자가 그래프를 보고 임의로 결정해야 하는 단점이 있다. Welch방법(Welch's Method: WM)은 브라운 브리지(Brownian bridge) 통계량()을 사용하는데, n이 무한에 가까워질 때, 이 브라운 브리지 분포(Brownian bridge distribution)에 수렴하는 성질을 이용한다. 시뮬레이션 출력데이터를 가지고 배치를 구성한 후 하나의 배치를 표본으로 사용한다. 이 기법은 알고리듬이 복잡하고, 값을 추정해야 하는 단점이 있다. Law-Kelton방법(Law-Kelton's Method: LKM)은 회귀 (regression)이론에 기초하는데, 시뮬레이션이 종료된 후 누적평균데이터에 대해서 회귀직선을 적합(fitting)시킨다. 회귀직선의 기울기가 0이라는 귀무가설이 채택되면 그 시점을 제거시점으로 결정한다. 일단 시뮬레이션이 종료된 다음, 데이터가 모아진 순서의 반대 순서로 데이터를 이용하기 때문에 온라인이 불가능하다. Welch절차(Welch's Procedure: WP)는 5회이상의 시뮬레이션수행을 통해 수집한 데이터의 이동평균을 이용해서 시각적으로 제거시점을 결정해야 하며, 반복제거방법을 사용해야 하기 때문에 온라인 제거시점의 결정이 불가능하다. 또한, 한번에 이동할 데이터의 크기(window size)를 결정해야 한다. 지금까지 알아 본 것처럼, 기존의 방법들은 시뮬레이션의 단일 수행 중의 온라인 제거시점 결정의 관점에서는 미약한 면이 있다. 또한, 현재의 시뮬레이션 상용소프트웨어는 작업자로 하여금 제거시점을 임의로 결정하도록 하기 때문에, 실험중인 시스템에 대해서 정확하고도 정량적으로 제거시점을 결정할 수 없게 되어 있다. 사용자가 임의로 제거시점을 결정하게 되면, 초기편의 문제를 효과적으로 해결하기 어려울 뿐만 아니라, 필요 이상으로 너무 많은 양을 제거하거나 초기편의를 해결하지 못할 만큼 너무 적은 양을 제거할 가능성이 커지게 된다. 또한, 기존의 방법들의 대부분은 제거시점을 찾기 위해서 시험수행이 필요하다. 즉, 안정상태 시점만을 찾기 위한 시뮬레이션 수행이 필요하며, 이렇게 사용된 시뮬레이션은 출력분석에 사용되지 않기 때문에 시간적인 손실이 크게 된다.

  • PDF

An Intelligent Decision Support System for Selecting Promising Technologies for R&D based on Time-series Patent Analysis (R&D 기술 선정을 위한 시계열 특허 분석 기반 지능형 의사결정지원시스템)

  • Lee, Choongseok;Lee, Suk Joo;Choi, Byounggu
    • Journal of Intelligence and Information Systems
    • /
    • 제18권3호
    • /
    • pp.79-96
    • /
    • 2012
  • As the pace of competition dramatically accelerates and the complexity of change grows, a variety of research have been conducted to improve firms' short-term performance and to enhance firms' long-term survival. In particular, researchers and practitioners have paid their attention to identify promising technologies that lead competitive advantage to a firm. Discovery of promising technology depends on how a firm evaluates the value of technologies, thus many evaluating methods have been proposed. Experts' opinion based approaches have been widely accepted to predict the value of technologies. Whereas this approach provides in-depth analysis and ensures validity of analysis results, it is usually cost-and time-ineffective and is limited to qualitative evaluation. Considerable studies attempt to forecast the value of technology by using patent information to overcome the limitation of experts' opinion based approach. Patent based technology evaluation has served as a valuable assessment approach of the technological forecasting because it contains a full and practical description of technology with uniform structure. Furthermore, it provides information that is not divulged in any other sources. Although patent information based approach has contributed to our understanding of prediction of promising technologies, it has some limitations because prediction has been made based on the past patent information, and the interpretations of patent analyses are not consistent. In order to fill this gap, this study proposes a technology forecasting methodology by integrating patent information approach and artificial intelligence method. The methodology consists of three modules : evaluation of technologies promising, implementation of technologies value prediction model, and recommendation of promising technologies. In the first module, technologies promising is evaluated from three different and complementary dimensions; impact, fusion, and diffusion perspectives. The impact of technologies refers to their influence on future technologies development and improvement, and is also clearly associated with their monetary value. The fusion of technologies denotes the extent to which a technology fuses different technologies, and represents the breadth of search underlying the technology. The fusion of technologies can be calculated based on technology or patent, thus this study measures two types of fusion index; fusion index per technology and fusion index per patent. Finally, the diffusion of technologies denotes their degree of applicability across scientific and technological fields. In the same vein, diffusion index per technology and diffusion index per patent are considered respectively. In the second module, technologies value prediction model is implemented using artificial intelligence method. This studies use the values of five indexes (i.e., impact index, fusion index per technology, fusion index per patent, diffusion index per technology and diffusion index per patent) at different time (e.g., t-n, t-n-1, t-n-2, ${\cdots}$) as input variables. The out variables are values of five indexes at time t, which is used for learning. The learning method adopted in this study is backpropagation algorithm. In the third module, this study recommends final promising technologies based on analytic hierarchy process. AHP provides relative importance of each index, leading to final promising index for technology. Applicability of the proposed methodology is tested by using U.S. patents in international patent class G06F (i.e., electronic digital data processing) from 2000 to 2008. The results show that mean absolute error value for prediction produced by the proposed methodology is lower than the value produced by multiple regression analysis in cases of fusion indexes. However, mean absolute error value of the proposed methodology is slightly higher than the value of multiple regression analysis. These unexpected results may be explained, in part, by small number of patents. Since this study only uses patent data in class G06F, number of sample patent data is relatively small, leading to incomplete learning to satisfy complex artificial intelligence structure. In addition, fusion index per technology and impact index are found to be important criteria to predict promising technology. This study attempts to extend the existing knowledge by proposing a new methodology for prediction technology value by integrating patent information analysis and artificial intelligence network. It helps managers who want to technology develop planning and policy maker who want to implement technology policy by providing quantitative prediction methodology. In addition, this study could help other researchers by proving a deeper understanding of the complex technological forecasting field.