• Title/Summary/Keyword: Machine Learning Procedure

Search Result 116, Processing Time 0.031 seconds

Sparse Representation Learning of Kernel Space Using the Kernel Relaxation Procedure (커널 이완절차에 의한 커널 공간의 저밀도 표현 학습)

  • 류재홍;정종철
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2001.12a
    • /
    • pp.60-64
    • /
    • 2001
  • In this paper, a new learning methodology for Kernel Methods is suggested that results in a sparse representation of kernel space from the training patterns for classification problems. Among the traditional algorithms of linear discriminant function(perceptron, relaxation, LMS(least mean squared), pseudoinverse), this paper shows that the relaxation procedure can obtain the maximum margin separating hyperplane of linearly separable pattern classification problem as SVM(Support Vector Machine) classifier does. The original relaxation method gives only the necessary condition of SV patterns. We suggest the sufficient condition to identify the SV patterns in the learning epochs. Experiment results show the new methods have the higher or equivalent performance compared to the conventional approach.

  • PDF

Utilization of deep learning-based metamodel for probabilistic seismic damage analysis of railway bridges considering the geometric variation

  • Xi Song;Chunhee Cho;Joonam Park
    • Earthquakes and Structures
    • /
    • v.25 no.6
    • /
    • pp.469-479
    • /
    • 2023
  • A probabilistic seismic damage analysis is an essential procedure to identify seismically vulnerable structures, prioritize the seismic retrofit, and ultimately minimize the overall seismic risk. To assess the seismic risk of multiple structures within a region, a large number of nonlinear time-history structural analyses must be conducted and studied. As a result, each assessment requires high computing resources. To overcome this limitation, we explore a deep learning-based metamodel to enable the prediction of the mean and the standard deviation of the seismic damage distribution of track-on steel-plate girder railway bridges in Korea considering the geometric variation. For machine learning training, nonlinear dynamic time-history analyses are performed to generate 800 high-fidelity datasets on the seismic response. Through intensive trial and error, the study is concentrated on developing an optimal machine learning architecture with the pre-identified variables of the physical configuration of the bridge. Additionally, the prediction performance of the proposed method is compared with a previous, well-defined, response surface model. Finally, the statistical testing results indicate that the overall performance of the deep-learning model is improved compared to the response surface model, as its errors are reduced by as much as 61%. In conclusion, the model proposed in this study can be effectively deployed for the seismic fragility and risk assessment of a region with a large number of structures.

A Deep Learning Application for Automated Feature Extraction in Transaction-based Machine Learning (트랜잭션 기반 머신러닝에서 특성 추출 자동화를 위한 딥러닝 응용)

  • Woo, Deock-Chae;Moon, Hyun Sil;Kwon, Suhnbeom;Cho, Yoonho
    • Journal of Information Technology Services
    • /
    • v.18 no.2
    • /
    • pp.143-159
    • /
    • 2019
  • Machine learning (ML) is a method of fitting given data to a mathematical model to derive insights or to predict. In the age of big data, where the amount of available data increases exponentially due to the development of information technology and smart devices, ML shows high prediction performance due to pattern detection without bias. The feature engineering that generates the features that can explain the problem to be solved in the ML process has a great influence on the performance and its importance is continuously emphasized. Despite this importance, however, it is still considered a difficult task as it requires a thorough understanding of the domain characteristics as well as an understanding of source data and the iterative procedure. Therefore, we propose methods to apply deep learning for solving the complexity and difficulty of feature extraction and improving the performance of ML model. Unlike other techniques, the most common reason for the superior performance of deep learning techniques in complex unstructured data processing is that it is possible to extract features from the source data itself. In order to apply these advantages to the business problems, we propose deep learning based methods that can automatically extract features from transaction data or directly predict and classify target variables. In particular, we applied techniques that show high performance in existing text processing based on the structural similarity between transaction data and text data. And we also verified the suitability of each method according to the characteristics of transaction data. Through our study, it is possible not only to search for the possibility of automated feature extraction but also to obtain a benchmark model that shows a certain level of performance before performing the feature extraction task by a human. In addition, it is expected that it will be able to provide guidelines for choosing a suitable deep learning model based on the business problem and the data characteristics.

Defect Diagnostics of Gas Turbine Engine Using Support Vector Machine and Artificial Neural Network (Support Vector Machine과 인공신경망을 이용한 가스터빈 엔진의 결함 진단에 관한 연구)

  • Park Jun-Cheol;Roh Tae-Seong;Choi Dong-Whan;Lee Chang-Ho
    • Journal of the Korean Society of Propulsion Engineers
    • /
    • v.10 no.2
    • /
    • pp.102-109
    • /
    • 2006
  • In this Paper, Support Vector Machine(SVM) and Artificial Neural Network(ANN) are used for developing the defect diagnostic algorithm of the aircraft turbo-shaft engine. The system that uses the ANN falls in a local minima when it learns many nonlinear data, and its classification accuracy ratio becomes low. To make up for this risk, the Separate Learning Algorithm(SLA) of ANN has been proposed by using SVM. This is the method that ANN learns selectively after discriminating the defect position by SVM, then more improved performance estimation can be obtained than using ANN only. The proposed SLA can make the higher classification accuracy by decreasing the nonlinearity of the massive data during the training procedure.

An Efficient Machine Learning-based Text Summarization in the Malayalam Language

  • P Haroon, Rosna;Gafur M, Abdul;Nisha U, Barakkath
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.6
    • /
    • pp.1778-1799
    • /
    • 2022
  • Automatic text summarization is a procedure that packs enormous content into a more limited book that incorporates significant data. Malayalam is one of the toughest languages utilized in certain areas of India, most normally in Kerala and in Lakshadweep. Natural language processing in the Malayalam language is relatively low due to the complexity of the language as well as the scarcity of available resources. In this paper, a way is proposed to deal with the text summarization process in Malayalam documents by training a model based on the Support Vector Machine classification algorithm. Different features of the text are taken into account for training the machine so that the system can output the most important data from the input text. The classifier can classify the most important, important, average, and least significant sentences into separate classes and based on this, the machine will be able to create a summary of the input document. The user can select a compression ratio so that the system will output that much fraction of the summary. The model performance is measured by using different genres of Malayalam documents as well as documents from the same domain. The model is evaluated by considering content evaluation measures precision, recall, F score, and relative utility. Obtained precision and recall value shows that the model is trustable and found to be more relevant compared to the other summarizers.

Export-Import Value Nowcasting Procedure Using Big Data-AIS and Machine Learning Techniques

  • NICKELSON, Jimmy;NOORAENI, Rani;EFLIZA, EFLIZA
    • Asian Journal of Business Environment
    • /
    • v.12 no.3
    • /
    • pp.1-12
    • /
    • 2022
  • Purpose: This study aims to investigate whether AIS data can be used as a supporting indicator or as an initial signal to describe Indonesia's export-import conditions in real-time. Research design, data, and methodology: This study performs several stages of data selection to obtain indicators from AIS that truly reflect export-import activities in Indonesia. Also, investigate the potential of AIS indicators in producing forecasts of the value and volume of Indonesian export-import using conventional statistical methods and machine learning techniques. Results: The six preprocessing stages defined in this study filtered AIS data from 661.8 million messages to 73.5 million messages. Seven predictors were formed from the selected AIS data. The AIS indicator can be used to provide an initial signal about Indonesia's import-export activities. Each export or import activity has its own predictor. Conventional statistical methods and machine learning techniques have the same ability both in forecasting Indonesia's exports and imports. Conclusions: Big data AIS can be used as a supporting indicator as a signal of the condition of export-import values in Indonesia. The right method of building indicators can make the data valuable for the performance of the forecasting model.

Coupling numerical modeling and machine-learning for back analysis of cantilever retaining wall failure

  • Amichai Mitelman;Gili Lifshitz Sherzer
    • Computers and Concrete
    • /
    • v.31 no.4
    • /
    • pp.307-314
    • /
    • 2023
  • In this paper we back-analyze a failure event of a 9 m high concrete cantilever wall subjected to earth loading. Granular soil was deposited into the space between the wall and a nearby rock slope. The wall segments were not designed to carry lateral earth loading and collapsed due to excessive bending. As many geotechnical programs rely on the Mohr-Coulomb (MC) criterion for elastoplastic analysis, it is useful to apply this failure criterion to the concrete material. Accordingly, the back-analysis is aimed to search for the suitable MC parameters of the concrete. For this study, we propose a methodology for accelerating the back-analysis task by automating the numerical modeling procedure and applying a machine-learning (ML) analysis on FE model results. Through this analysis it is found that the residual cohesion and friction angle have a highly significant impact on model results. Compared to traditional back-analysis studies where good agreement between model and reality are deemed successful based on a limited number of models, the current ML analysis demonstrate that a range of possible combinations of parameters can yield similar results. The proposed methodology can be modified for similar calibration and back-analysis tasks.

Two-Agent Single-Machine Scheduling with Linear Job-Dependent Position-Based Learning Effects (작업 종속 및 위치기반 선형학습효과를 갖는 2-에이전트 단일기계 스케줄링)

  • Choi, Jin Young
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.38 no.3
    • /
    • pp.169-180
    • /
    • 2015
  • Recently, scheduling problems with position-dependent processing times have received considerable attention in the literature, where the processing times of jobs are dependent on the processing sequences. However, they did not consider cases in which each processed job has different learning or aging ratios. This means that the actual processing time for a job can be determined not only by the processing sequence, but also by the learning/aging ratio, which can reflect the degree of processing difficulties in subsequent jobs. Motivated by these remarks, in this paper, we consider a two-agent single-machine scheduling problem with linear job-dependent position-based learning effects, where two agents compete to use a common single machine and each job has a different learning ratio. Specifically, we take into account two different objective functions for two agents: one agent minimizes the total weighted completion time, and the other restricts the makespan to less than an upper bound. After formally defining the problem by developing a mixed integer non-linear programming formulation, we devise a branch-and-bound (B&B) algorithm to give optimal solutions by developing four dominance properties based on a pairwise interchange comparison and four properties regarding the feasibility of a considered sequence. We suggest a lower bound to speed up the search procedure in the B&B algorithm by fathoming any non-prominent nodes. As this problem is at least NP-hard, we suggest efficient genetic algorithms using different methods to generate the initial population and two crossover operations. Computational results show that the proposed algorithms are efficient to obtain near-optimal solutions.

A Study on Personalized Recommendation Method Based on Contents Using Activity and Location Information (이용자 이용행위 및 콘텐츠 위치정보에 기반한 개인화 추천방법에 관한 연구)

  • Kim, Yong;Kim, Mun-Seok;Kim, Yoon-Beom;Park, Jae-Hong
    • Journal of the Korean Society for information Management
    • /
    • v.26 no.1
    • /
    • pp.81-105
    • /
    • 2009
  • In this paper, we propose user contents using behavior and location information on contents on various channels, such as web, IPTV, for contents distribution. With methods to build user and contents profiles, contents using behavior as an implicit user feedback was applied into machine learning procedure for updating user profiles and contents preference. In machine learning procedure, contents-based and collaborative filtering methods were used to analyze user's contents preference. This study proposes contents location information on web sites for final recommendation contents as well. Finally, we refer to a generalized recommender system for personalization. With those methods, more effective and accurate recommendation service can be possible.

Comparison of Prediction Accuracy Between Classification and Convolution Algorithm in Fault Diagnosis of Rotatory Machines at Varying Speed (회전수가 변하는 기기의 고장진단에 있어서 특성 기반 분류와 합성곱 기반 알고리즘의 예측 정확도 비교)

  • Moon, Ki-Yeong;Kim, Hyung-Jin;Hwang, Se-Yun;Lee, Jang Hyun
    • Journal of Navigation and Port Research
    • /
    • v.46 no.3
    • /
    • pp.280-288
    • /
    • 2022
  • This study examined the diagnostics of abnormalities and faults of equipment, whose rotational speed changes even during regular operation. The purpose of this study was to suggest a procedure that can properly apply machine learning to the time series data, comprising non-stationary characteristics as the rotational speed changes. Anomaly and fault diagnosis was performed using machine learning: k-Nearest Neighbor (k-NN), Support Vector Machine (SVM), and Random Forest. To compare the diagnostic accuracy, an autoencoder was used for anomaly detection and a convolution based Conv1D was additionally used for fault diagnosis. Feature vectors comprising statistical and frequency attributes were extracted, and normalization & dimensional reduction were applied to the extracted feature vectors. Changes in the diagnostic accuracy of machine learning according to feature selection, normalization, and dimensional reduction are explained. The hyperparameter optimization process and the layered structure are also described for each algorithm. Finally, results show that machine learning can accurately diagnose the failure of a variable-rotation machine under the appropriate feature treatment, although the convolution algorithms have been widely applied to the considered problem.