• Title/Summary/Keyword: Image machine learning

Search Result 595, Processing Time 0.025 seconds

Sign Language Image Recognition System Using Artificial Neural Network

  • Kim, Hyung-Hoon;Cho, Jeong-Ran
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.2
    • /
    • pp.193-200
    • /
    • 2019
  • Hearing impaired people are living in a voice culture area, but due to the difficulty of communicating with normal people using sign language, many people experience discomfort in daily life and social life and various disadvantages unlike their desires. Therefore, in this paper, we study a sign language translation system for communication between a normal person and a hearing impaired person using sign language and implement a prototype system for this. Previous studies on sign language translation systems for communication between normal people and hearing impaired people using sign language are classified into two types using video image system and shape input device. However, existing sign language translation systems have some problems that they do not recognize various sign language expressions of sign language users and require special devices. In this paper, we use machine learning method of artificial neural network to recognize various sign language expressions of sign language users. By using generalized smart phone and various video equipment for sign language image recognition, we intend to improve the usability of sign language translation system.

Synthetic Image Generation for Military Vehicle Detection (군용물체탐지 연구를 위한 가상 이미지 데이터 생성)

  • Se-Yoon Oh;Hunmin Yang
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.26 no.5
    • /
    • pp.392-399
    • /
    • 2023
  • This research paper investigates the effectiveness of using computer graphics(CG) based synthetic data for deep learning in military vehicle detection. In particular, we explore the use of synthetic image generation techniques to train deep neural networks for object detection tasks. Our approach involves the generation of a large dataset of synthetic images of military vehicles, which is then used to train a deep learning model. The resulting model is then evaluated on real-world images to measure its effectiveness. Our experimental results show that synthetic training data alone can achieve effective results in object detection. Our findings demonstrate the potential of CG-based synthetic data for deep learning and suggest its value as a tool for training models in a variety of applications, including military vehicle detection.

Object Classification Using Point Cloud and True Ortho-image by Applying Random Forest and Support Vector Machine Techniques (랜덤포레스트와 서포트벡터머신 기법을 적용한 포인트 클라우드와 실감정사영상을 이용한 객체분류)

  • Seo, Hong Deok;Kim, Eui Myoung
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.37 no.6
    • /
    • pp.405-416
    • /
    • 2019
  • Due to the development of information and communication technology, the production and processing speed of data is getting faster. To classify objects using machine learning, which is a field of artificial intelligence, data required for training can be easily collected due to the development of internet and geospatial information technology. In the field of geospatial information, machine learning is also being applied to classify or recognize objects using images and point clouds. In this study, the problem of manually constructing training data using existing digital map version 1.0 was improved, and the technique of classifying roads, buildings and vegetation using image and point clouds were proposed. Through experiments, it was possible to classify roads, buildings, and vegetation that could clearly distinguish colors when using true ortho-image with only RGB (Red, Green, Blue) bands. However, if the colors of the objects to be classified are similar, it was possible to identify the limitations of poor classification of the objects. To improve the limitations, random forest and support vector machine techniques were applied after band fusion of true ortho-image and normalized digital surface model, and roads, buildings, and vegetation were classified with more than 85% accuracy.

CNN-based Automatic Machine Fault Diagnosis Method Using Spectrogram Images (스펙트로그램 이미지를 이용한 CNN 기반 자동화 기계 고장 진단 기법)

  • Kang, Kyung-Won;Lee, Kyeong-Min
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.21 no.3
    • /
    • pp.121-126
    • /
    • 2020
  • Sound-based machine fault diagnosis is the automatic detection of abnormal sound in the acoustic emission signals of the machines. Conventional methods of using mathematical models were difficult to diagnose machine failure due to the complexity of the industry machinery system and the existence of nonlinear factors such as noises. Therefore, we want to solve the problem of machine fault diagnosis as a deep learning-based image classification problem. In the paper, we propose a CNN-based automatic machine fault diagnosis method using Spectrogram images. The proposed method uses STFT to effectively extract feature vectors from frequencies generated by machine defects, and the feature vectors detected by STFT were converted into spectrogram images and classified by CNN by machine status. The results show that the proposed method can be effectively used not only to detect defects but also to various automatic diagnosis system based on sound.

Performance Improvement of Classifier by Combining Disjunctive Normal Form features

  • Min, Hyeon-Gyu;Kang, Dong-Joong
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.10 no.4
    • /
    • pp.50-64
    • /
    • 2018
  • This paper describes a visual object detection approach utilizing ensemble based machine learning. Object detection methods employing 1D features have the benefit of fast calculation speed. However, for real image with complex background, detection accuracy and performance are degraded. In this paper, we propose an ensemble learning algorithm that combines a 1D feature classifier and 2D DNF (Disjunctive Normal Form) classifier to improve the object detection performance in a single input image. Also, to improve the computing efficiency and accuracy, we propose a feature selecting method to reduce the computing time and ensemble algorithm by combining the 1D features and 2D DNF features. In the verification experiments, we selected the Haar-like feature as the 1D image descriptor, and demonstrated the performance of the algorithm on a few datasets such as face and vehicle.

Structural Crack Detection Using Deep Learning: An In-depth Review

  • Safran Khan;Abdullah Jan;Suyoung Seo
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.4
    • /
    • pp.371-393
    • /
    • 2023
  • Crack detection in structures plays a vital role in ensuring their safety, durability, and reliability. Traditional crack detection methods sometimes need significant manual inspections, which are laborious, expensive, and prone to error by humans. Deep learning algorithms, which can learn intricate features from large-scale datasets, have emerged as a viable option for automated crack detection recently. This study presents an in-depth review of crack detection methods used till now, like image processing, traditional machine learning, and deep learning methods. Specifically, it will provide a comparative analysis of crack detection methods using deep learning, aiming to provide insights into the advancements, challenges, and future directions in this field. To facilitate comparative analysis, this study surveys publicly available crack detection datasets and benchmarks commonly used in deep learning research. Evaluation metrics employed to check the performance of different models are discussed, with emphasis on accuracy, precision, recall, and F1-score. Moreover, this study provides an in-depth analysis of recent studies and highlights key findings, including state-of-the-art techniques, novel architectures, and innovative approaches to address the shortcomings of the existing methods. Finally, this study provides a summary of the key insights gained from the comparative analysis, highlighting the potential of deep learning in revolutionizing methodologies for crack detection. The findings of this research will serve as a valuable resource for researchers in the field, aiding them in selecting appropriate methods for crack detection and inspiring further advancements in this domain.

Low-Informative Region Detection based on Multi-Layer Perceptron for Automatical Insertion of Virtual Advertisement in Sports Image (스포츠 영상 내에서 자동적인 가상 광고 삽입을 위한 다층퍼셉트론 기반의 저정보 영역 검출)

  • Jung, Jae-Young;Kim, Jong-Ha
    • Journal of Digital Contents Society
    • /
    • v.18 no.1
    • /
    • pp.71-77
    • /
    • 2017
  • Virtual advertisement is an advertising technique that using computer graphic in a media production such as a sports image for inserting product image, logo, advertising slogan, etc. Recently, the image insertion of virtual advertisement is actively spreading due to the satisfaction of technical element for the image insertion of virtual advertisement in sports advertisement by increasing of the image processing technology and the computing performance. In addition, image processing technology for automatic insertion has become an important research field in the virtual advertisement field. In this paper, we propose the method of extracting less-informative region by using image processing technique and machine learning to insert a virtual advertisement automatically in sports image. The proposed method analyzes the brightness level of image through the histogram and extracts the less-informative region using the machine learning method.

Android malicious code Classification using Deep Belief Network

  • Shiqi, Luo;Shengwei, Tian;Long, Yu;Jiong, Yu;Hua, Sun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.1
    • /
    • pp.454-475
    • /
    • 2018
  • This paper presents a novel Android malware classification model planned to classify and categorize Android malicious code at Drebin dataset. The amount of malicious mobile application targeting Android based smartphones has increased rapidly. In this paper, Restricted Boltzmann Machine and Deep Belief Network are used to classify malware into families of Android application. A texture-fingerprint based approach is proposed to extract or detect the feature of malware content. A malware has a unique "image texture" in feature spatial relations. The method uses information on texture image extracted from malicious or benign code, which are mapped to uncompressed gray-scale according to the texture image-based approach. By studying and extracting the implicit features of the API call from a large number of training samples, we get the original dynamic activity features sets. In order to improve the accuracy of classification algorithm on the features selection, on the basis of which, it combines the implicit features of the texture image and API call in malicious code, to train Restricted Boltzmann Machine and Back Propagation. In an evaluation with different malware and benign samples, the experimental results suggest that the usability of this method---using Deep Belief Network to classify Android malware by their texture images and API calls, it detects more than 94% of the malware with few false alarms. Which is higher than shallow machine learning algorithm clearly.

Driver Drowsiness Detection Model using Image and PPG data Based on Multimodal Deep Learning (이미지와 PPG 데이터를 사용한 멀티모달 딥 러닝 기반의 운전자 졸음 감지 모델)

  • Choi, Hyung-Tak;Back, Moon-Ki;Kang, Jae-Sik;Yoon, Seung-Won;Lee, Kyu-Chul
    • Database Research
    • /
    • v.34 no.3
    • /
    • pp.45-57
    • /
    • 2018
  • The drowsiness that occurs in the driving is a very dangerous driver condition that can be directly linked to a major accident. In order to prevent drowsiness, there are traditional drowsiness detection methods to grasp the driver's condition, but there is a limit to the generalized driver's condition recognition that reflects the individual characteristics of drivers. In recent years, deep learning based state recognition studies have been proposed to recognize drivers' condition. Deep learning has the advantage of extracting features from a non-human machine and deriving a more generalized recognition model. In this study, we propose a more accurate state recognition model than the existing deep learning method by learning image and PPG at the same time to grasp driver's condition. This paper confirms the effect of driver's image and PPG data on drowsiness detection and experiment to see if it improves the performance of learning model when used together. We confirmed the accuracy improvement of around 3% when using image and PPG together than using image alone. In addition, the multimodal deep learning based model that classifies the driver's condition into three categories showed a classification accuracy of 96%.

Image-Based Machine Learning Model for Malware Detection on LLVM IR (LLVM IR 대상 악성코드 탐지를 위한 이미지 기반 머신러닝 모델)

  • Kyung-bin Park;Yo-seob Yoon;Baasantogtokh Duulga;Kang-bin Yim
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.34 no.1
    • /
    • pp.31-40
    • /
    • 2024
  • Recently, static analysis-based signature and pattern detection technologies have limitations due to the advanced IT technologies. Moreover, It is a compatibility problem of multiple architectures and an inherent problem of signature and pattern detection. Malicious codes use obfuscation and packing techniques to hide their identity, and they also avoid existing static analysis-based signature and pattern detection techniques such as code rearrangement, register modification, and branching statement addition. In this paper, We propose an LLVM IR image-based automated static analysis of malicious code technology using machine learning to solve the problems mentioned above. Whether binary is obfuscated or packed, it's decompiled into LLVM IR, which is an intermediate representation dedicated to static analysis and optimization. "Therefore, the LLVM IR code is converted into an image before being fed to the CNN-based transfer learning algorithm ResNet50v2 supported by Keras". As a result, we present a model for image-based detection of malicious code.