• Title/Summary/Keyword: Conventional machine learning

Search Result 295, Processing Time 0.03 seconds

Image Dehazing Algorithm Using Near-infrared Image Characteristics (근적외선 영상의 특성을 활용한 안개 제거 알고리즘)

  • Yu, Jae Taeg;Ra, Sung Woong;Lee, Sungmin;Jung, Seung-Won
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.52 no.11
    • /
    • pp.115-123
    • /
    • 2015
  • The infrared light is known to be less dependent on background light compared to the visible light, and thus many applications such as remote sensing and image surveillance use the infrared image. Similar to color images, infrared images can also be degraded by hazy weather condition, and consequently the performance of the infrared image-based applications can decrease. Nevertheless, infrared image dehazing has not received significant interest. In this paper, we analyze the characteristic of infrared images, especially near-infrared (NIR) images, and present an NIR dehazing algorithm using the analyzed characteristics. In particular, a machine learning framework is adopted to obtain an accurate transmission map and several post-processing methods are used for further refinement. Experimental results show that the proposed NIR dehazing algorithm outperforms the conventional color image dehazing method for NIR image dehazing.

Denoising Self-Attention Network for Mixed-type Data Imputation (혼합형 데이터 보간을 위한 디노이징 셀프 어텐션 네트워크)

  • Lee, Do-Hoon;Kim, Han-Joon;Chun, Joonghoon
    • The Journal of the Korea Contents Association
    • /
    • v.21 no.11
    • /
    • pp.135-144
    • /
    • 2021
  • Recently, data-driven decision-making technology has become a key technology leading the data industry, and machine learning technology for this requires high-quality training datasets. However, real-world data contains missing values for various reasons, which degrades the performance of prediction models learned from the poor training data. Therefore, in order to build a high-performance model from real-world datasets, many studies on automatically imputing missing values in initial training data have been actively conducted. Many of conventional machine learning-based imputation techniques for handling missing data involve very time-consuming and cumbersome work because they are applied only to numeric type of columns or create individual predictive models for each columns. Therefore, this paper proposes a new data imputation technique called 'Denoising Self-Attention Network (DSAN)', which can be applied to mixed-type dataset containing both numerical and categorical columns. DSAN can learn robust feature expression vectors by combining self-attention and denoising techniques, and can automatically interpolate multiple missing variables in parallel through multi-task learning. To verify the validity of the proposed technique, data imputation experiments has been performed after arbitrarily generating missing values for several mixed-type training data. Then we show the validity of the proposed technique by comparing the performance of the binary classification models trained on imputed data together with the errors between the original and imputed values.

Combining Radar and Rain Gauge Observations Utilizing Gaussian-Process-Based Regression and Support Vector Learning (가우시안 프로세스 기반 함수근사와 서포트 벡터 학습을 이용한 레이더 및 강우계 관측 데이터의 융합)

  • Yoo, Chul-Sang;Park, Joo-Young
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.18 no.3
    • /
    • pp.297-305
    • /
    • 2008
  • Recently, kernel methods have attracted great interests in the areas of pattern classification, function approximation, and anomaly detection. The role of the kernel is particularly important in the methods such as SVM(support vector machine) and KPCA(kernel principal component analysis), for it can generalize the conventional linear machines to be capable of efficiently handling nonlinearities. This paper considers the problem of combining radar and rain gauge observations utilizing the regression approach based on the kernel-based gaussian process and support vector learning. The data-assimilation results of the considered methods are reported for the radar and rain gauge observations collected over the region covering parts of Gangwon, Kyungbuk, and Chungbuk provinces of Korea, along with performance comparison.

Accuracy of Phishing Websites Detection Algorithms by Using Three Ranking Techniques

  • Mohammed, Badiea Abdulkarem;Al-Mekhlafi, Zeyad Ghaleb
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.2
    • /
    • pp.272-282
    • /
    • 2022
  • Between 2014 and 2019, the US lost more than 2.1 billion USD to phishing attacks, according to the FBI's Internet Crime Complaint Center, and COVID-19 scam complaints totaled more than 1,200. Phishing attacks reflect these awful effects. Phishing websites (PWs) detection appear in the literature. Previous methods included maintaining a centralized blacklist that is manually updated, but newly created pseudonyms cannot be detected. Several recent studies utilized supervised machine learning (SML) algorithms and schemes to manipulate the PWs detection problem. URL extraction-based algorithms and schemes. These studies demonstrate that some classification algorithms are more effective on different data sets. However, for the phishing site detection problem, no widely known classifier has been developed. This study is aimed at identifying the features and schemes of SML that work best in the face of PWs across all publicly available phishing data sets. The Scikit Learn library has eight widely used classification algorithms configured for assessment on the public phishing datasets. Eight was tested. Later, classification algorithms were used to measure accuracy on three different datasets for statistically significant differences, along with the Welch t-test. Assemblies and neural networks outclass classical algorithms in this study. On three publicly accessible phishing datasets, eight traditional SML algorithms were evaluated, and the results were calculated in terms of classification accuracy and classifier ranking as shown in tables 4 and 8. Eventually, on severely unbalanced datasets, classifiers that obtained higher than 99.0 percent classification accuracy. Finally, the results show that this could also be adapted and outperforms conventional techniques with good precision.

Prediction of Material's Formation Energy Using Crystal Graph Convolutional Neural Network (결정그래프 합성곱 인공신경망을 통한 소재의 생성 에너지 예측)

  • Lee, Hyun-Gi;Seo, Dong-Hwa
    • Journal of the Korean Institute of Electrical and Electronic Material Engineers
    • /
    • v.35 no.2
    • /
    • pp.134-142
    • /
    • 2022
  • As industry and technology go through advancement, it is hard to search new materials which satisfy various standards through conventional trial-and-error based research methods. Crystal Graph Convolutional Neural Network(CGCNN) is a neural network which uses material's features as train data, and predicts the material properties(formation energy, bandgap, etc.) much faster than first-principles calculation. This report introduces how to train the CGCNN model which predicts the formation energy using open database. It is anticipated that with a simple programming skill, readers could construct a model using their data and purpose. Developing machine learning model for materials science is going to help researchers who should explore large chemical and structural space to discover materials efficiently.

Reproduction strategy of radiation data with compensation of data loss using a deep learning technique

  • Cho, Woosung;Kim, Hyeonmin;Kim, Duckhyun;Kim, SongHyun;Kwon, Inyong
    • Nuclear Engineering and Technology
    • /
    • v.53 no.7
    • /
    • pp.2229-2236
    • /
    • 2021
  • In nuclear-related facilities, such as nuclear power plants, research reactors, accelerators, and nuclear waste storage sites, radiation detection, and mapping are required to prevent radiation overexposure. Sensor network systems consisting of radiation sensor interfaces and wxireless communication units have become promising tools that can be used for data collection of radiation detection that can in turn be used to draw a radiation map. During data collection, malfunctions in some of the sensors can occasionally occur due to radiation effects, physical damage, network defects, sensor loss, or other reasons. This paper proposes a reproduction strategy for radiation maps using a U-net model to compensate for the loss of radiation detection data. To perform machine learning and verification, 1,561 simulations and 417 measured data of a sensor network were performed. The reproduction results show an accuracy of over 90%. The proposed strategy can offer an effective method that can be used to resolve the data loss problem for conventional sensor network systems and will specifically contribute to making initial responses with preserved data and without the high cost of radiation leak accidents at nuclear facilities.

Case Analysis of Applications of Seismic Data Denoising Methods using Deep-Learning Techniques (심층 학습 기법을 이용한 탄성파 자료 잡음 제거 적용사례 분석)

  • Jo, Jun Hyeon;Ha, Wansoo
    • Geophysics and Geophysical Exploration
    • /
    • v.23 no.2
    • /
    • pp.72-88
    • /
    • 2020
  • Recent rapid advances in computer hardware performance have led to relatively low computational costs, increasing the number of applications of machine-learning techniques to geophysical problems. In particular, deep-learning techniques are gaining in popularity as the number of cases successfully solving complex and nonlinear problems has gradually increased. In this paper, applications of seismic data denoising methods using deep-learning techniques are introduced and investigated. Depending on the type of attenuated noise, these studies are grouped into denoising applications of coherent noise, random noise, and the combination of these two types of noise. Then, we investigate the deep-learning techniques used to remove the corresponding noise. Unlike conventional methods used to attenuate seismic noise, deep neural networks, a typical deep-learning technique, learn the characteristics of the noise independently and then automatically optimize the parameters. Therefore, such methods are less sensitive to generalized problems than conventional methods and can reduce labor costs. Several studies have also demonstrated that deep-learning techniques perform well in terms of computational cost and denoising performance. Based on the results of the applications covered in this paper, the pros and cons of the deep-learning techniques used to remove seismic noise are analyzed and discussed.

Computer vision and deep learning-based post-earthquake intelligent assessment of engineering structures: Technological status and challenges

  • T. Jin;X.W. Ye;W.M. Que;S.Y. Ma
    • Smart Structures and Systems
    • /
    • v.31 no.4
    • /
    • pp.311-323
    • /
    • 2023
  • Ever since ancient times, earthquakes have been a major threat to the civil infrastructures and the safety of human beings. The majority of casualties in earthquake disasters are caused by the damaged civil infrastructures but not by the earthquake itself. Therefore, the efficient and accurate post-earthquake assessment of the conditions of structural damage has been an urgent need for human society. Traditional ways for post-earthquake structural assessment rely heavily on field investigation by experienced experts, yet, it is inevitably subjective and inefficient. Structural response data are also applied to assess the damage; however, it requires mounted sensor networks in advance and it is not intuitional. As many types of damaged states of structures are visible, computer vision-based post-earthquake structural assessment has attracted great attention among the engineers and scholars. With the development of image acquisition sensors, computing resources and deep learning algorithms, deep learning-based post-earthquake structural assessment has gradually shown potential in dealing with image acquisition and processing tasks. This paper comprehensively reviews the state-of-the-art studies of deep learning-based post-earthquake structural assessment in recent years. The conventional way of image processing and machine learning-based structural assessment are presented briefly. The workflow of the methodology for computer vision and deep learning-based post-earthquake structural assessment was introduced. Then, applications of assessment for multiple civil infrastructures are presented in detail. Finally, the challenges of current studies are summarized for reference in future works to improve the efficiency, robustness and accuracy in this field.

Interference Elimination Method of Ultrasonic Sensors Using K-Nearest Neighbor Algorithm (KNN 알고리즘을 활용한 초음파 센서 간 간섭 제거 기법)

  • Im, Hyungchul;Lee, Seongsoo
    • Journal of IKEEE
    • /
    • v.26 no.2
    • /
    • pp.169-175
    • /
    • 2022
  • This paper introduces an interference elimination method using k-nearest neighbor (KNN) algorithm for precise distance estimation by reducing interference between ultrasonic sensors. Conventional methods compare current distance measurement result with previous distance measurement results. If the difference exceeds some thresholds, conventional methods recognize them as interference and exclude them, but they often suffer from imprecise distance prediction. KNN algorithm classifies input values measured by multiple ultrasonic sensors and predicts high accuracy outputs. Experiments of distance measurements are conducted where interference frequently occurs by multiple ultrasound sensors of same type, and the results show that KNN algorithm significantly reduce distance prediction errors. Also the results show that the prediction performance of KNN algorithm is superior to conventional voting methods.

A Fusion Method of Co-training and Label Propagation for Prediction of Bank Telemarketing (은행 텔레마케팅 예측을 위한 레이블 전파와 협동 학습의 결합 방법)

  • Kim, Aleum;Cho, Sung-Bae
    • Journal of KIISE
    • /
    • v.44 no.7
    • /
    • pp.686-691
    • /
    • 2017
  • Telemarketing has become the center of marketing action of the industry in the information society. Recently, machine learning has emerged in many areas, especially, financial prediction. Financial data consists of lots of unlabeled data in most parts, and therefore, it is difficult for humans to perform their labeling. In this paper, we propose a fusion method of semi-supervised learning for automatic labeling of unlabeled data to predict telemarketing. Specifically, we integrate labeling results of label propagation and co-training with a decision tree. The data with lower reliabilities are removed, and the data are extracted that have consistent label from two labeling methods. After adding them to the training set, a decision tree is learned with all of them. To confirm the usefulness of the proposed method, we conduct the experiments with a real telemarketing dataset in a Portugal bank. Accuracy of the proposed method is 83.39%, which is 1.82% higher than that of the conventional method, and precision of the proposed method is 19.37%, which is 2.67% higher than that of the conventional method. As a result, we have shown that the proposed method has a better performance as assessed by the t-test.