• Title/Summary/Keyword: object detection and classification

Search Result 296, Processing Time 0.03 seconds

Object Classification and Change Detection in Point Clouds Using Deep Learning (포인트 클라우드에서 딥러닝을 이용한 객체 분류 및 변화 탐지)

  • Seo, Hong-Deok;Kim, Eui-Myoung
    • Journal of Cadastre & Land InformatiX
    • /
    • v.50 no.2
    • /
    • pp.37-51
    • /
    • 2020
  • With the development of machine learning and deep learning technologies, there has been increasing interest and attempt to apply these technologies to the detection of urban changes. However, the traditional methods of detecting changes and constructing spatial information are still often performed manually by humans, which is costly and time-consuming. Besides, a large number of people are needed to efficiently detect changes in buildings in urban areas. Therefore, in this study, a methodology that can detect changes by classifying road, building, and vegetation objects that are highly utilized in the geospatial information field was proposed by applying deep learning technology to point clouds. As a result of the experiment, roads, buildings, and vegetation were classified with an accuracy of 92% or more, and attributes information of the objects could be automatically constructed through this. In addition, if time-series data is constructed, it is thought that changes can be detected and attributes of existing digital maps can be inspected through the proposed methodology.

A BERT-Based Deep Learning Approach for Vulnerability Detection (BERT를 이용한 딥러닝 기반 소스코드 취약점 탐지 방법 연구)

  • Jin, Wenhui;Oh, Heekuck
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.32 no.6
    • /
    • pp.1139-1150
    • /
    • 2022
  • With the rapid development of SW Industry, softwares are everywhere in our daily life. The number of vulnerabilities are also increasing with a large amount of newly developed code. Vulnerabilities can be exploited by hackers, resulting the disclosure of privacy and threats to the safety of property and life. In particular, since the large numbers of increasing code, manually analyzed by expert is not enough anymore. Machine learning has shown high performance in object identification or classification task. Vulnerability detection is also suitable for machine learning, as a reuslt, many studies tried to use RNN-based model to detect vulnerability. However, the RNN model is also has limitation that as the code is longer, the earlier can not be learned well. In this paper, we proposed a novel method which applied BERT to detect vulnerability. The accuracy was 97.5%, which increased by 1.5%, and the efficiency also increased by 69% than Vuldeepecker.

Development of CCTV Cooperation Tracking System for Real-Time Crime Monitoring (실시간 범죄 모니터링을 위한 CCTV 협업 추적시스템 개발 연구)

  • Choi, Woo-Chul;Na, Joon-Yeop
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.20 no.12
    • /
    • pp.546-554
    • /
    • 2019
  • Typically, closed-circuit television (CCTV) monitoring is mainly used for post-processes (i.e. to provide evidence after an incident has occurred), but by using a streaming video feed, machine-based learning, and advanced image recognition techniques, current technology can be extended to respond to crimes or reports of missing persons in real time. The multi-CCTV cooperation technique developed in this study is a program model that delivers similarity information about a suspect (or moving object) extracted via CCTV at one location and sent to a monitoring agent to track the selected suspect or object when he, she, or it moves out of range to another CCTV camera. To improve the operating efficiency of local government CCTV control centers, we describe here the partial automation of a CCTV control system that currently relies upon monitoring by human agents. We envisage an integrated crime prevention service, which incorporates the cooperative CCTV network suggested in this study and that can easily be experienced by citizens in ways such as determining a precise individual location in real time and providing a crime prevention service linked to smartphones and/or crime prevention/safety information.

Dual CNN Structured Sound Event Detection Algorithm Based on Real Life Acoustic Dataset (실생활 음향 데이터 기반 이중 CNN 구조를 특징으로 하는 음향 이벤트 인식 알고리즘)

  • Suh, Sangwon;Lim, Wootaek;Jeong, Youngho;Lee, Taejin;Kim, Hui Yong
    • Journal of Broadcast Engineering
    • /
    • v.23 no.6
    • /
    • pp.855-865
    • /
    • 2018
  • Sound event detection is one of the research areas to model human auditory cognitive characteristics by recognizing events in an environment with multiple acoustic events and determining the onset and offset time for each event. DCASE, a research group on acoustic scene classification and sound event detection, is proceeding challenges to encourage participation of researchers and to activate sound event detection research. However, the size of the dataset provided by the DCASE Challenge is relatively small compared to ImageNet, which is a representative dataset for visual object recognition, and there are not many open sources for the acoustic dataset. In this study, the sound events that can occur in indoor and outdoor are collected on a larger scale and annotated for dataset construction. Furthermore, to improve the performance of the sound event detection task, we developed a dual CNN structured sound event detection system by adding a supplementary neural network to a convolutional neural network to determine the presence of sound events. Finally, we conducted a comparative experiment with both baseline systems of the DCASE 2016 and 2017.

A Feasibility Study on Application of a Deep Convolutional Neural Network for Automatic Rock Type Classification (자동 암종 분류를 위한 딥러닝 영상처리 기법의 적용성 검토 연구)

  • Pham, Chuyen;Shin, Hyu-Soung
    • Tunnel and Underground Space
    • /
    • v.30 no.5
    • /
    • pp.462-472
    • /
    • 2020
  • Rock classification is fundamental discipline of exploring geological and geotechnical features in a site, which, however, may not be easy works because of high diversity of rock shape and color according to its origin, geological history and so on. With the great success of convolutional neural networks (CNN) in many different image-based classification tasks, there has been increasing interest in taking advantage of CNN to classify geological material. In this study, a feasibility of the deep CNN is investigated for automatically and accurately identifying rock types, focusing on the condition of various shapes and colors even in the same rock type. It can be further developed to a mobile application for assisting geologist in classifying rocks in fieldwork. The structure of CNN model used in this study is based on a deep residual neural network (ResNet), which is an ultra-deep CNN using in object detection and classification. The proposed CNN was trained on 10 typical rock types with an overall accuracy of 84% on the test set. The result demonstrates that the proposed approach is not only able to classify rock type using images, but also represents an improvement as taking highly diverse rock image dataset as input.

Object Detection Based on Deep Learning Model for Two Stage Tracking with Pest Behavior Patterns in Soybean (Glycine max (L.) Merr.)

  • Yu-Hyeon Park;Junyong Song;Sang-Gyu Kim ;Tae-Hwan Jun
    • Proceedings of the Korean Society of Crop Science Conference
    • /
    • 2022.10a
    • /
    • pp.89-89
    • /
    • 2022
  • Soybean (Glycine max (L.) Merr.) is a representative food resource. To preserve the integrity of soybean, it is necessary to protect soybean yield and seed quality from threats of various pests and diseases. Riptortus pedestris is a well-known insect pest that causes the greatest loss of soybean yield in South Korea. This pest not only directly reduces yields but also causes disorders and diseases in plant growth. Unfortunately, no resistant soybean resources have been reported. Therefore, it is necessary to identify the distribution and movement of Riptortus pedestris at an early stage to reduce the damage caused by insect pests. Conventionally, the human eye has performed the diagnosis of agronomic traits related to pest outbreaks. However, due to human vision's subjectivity and impermanence, it is time-consuming, requires the assistance of specialists, and is labor-intensive. Therefore, the responses and behavior patterns of Riptortus pedestris to the scent of mixture R were visualized with a 3D model through the perspective of artificial intelligence. The movement patterns of Riptortus pedestris was analyzed by using time-series image data. In addition, classification was performed through visual analysis based on a deep learning model. In the object tracking, implemented using the YOLO series model, the path of the movement of pests shows a negative reaction to a mixture Rina video scene. As a result of 3D modeling using the x, y, and z-axis of the tracked objects, 80% of the subjects showed behavioral patterns consistent with the treatment of mixture R. In addition, these studies are being conducted in the soybean field and it will be possible to preserve the yield of soybeans through the application of a pest control platform to the early stage of soybeans.

  • PDF

Design and Implementation of Radar Signal Processing System for Vehicle Door Collision Prevention (차량 도어 충돌 방지용 레이다 신호처리 시스템 설계 및 구현)

  • Jeongwoo Han;Minsang Kim;Daehong Kim;Yunho Jung
    • Journal of IKEEE
    • /
    • v.28 no.3
    • /
    • pp.397-404
    • /
    • 2024
  • This paper presents the design and implementation results of a Raspberry-Pi-based embedded system with an FPGA accelerator that can detect and classify objects using an FMCW radar sensor for preventing door collision accidents in vehicles. The proposed system performs a radar sensor signal processing and a deep learning processing that classifies objects into bicycles, automobiles, and pedestrians. Since the CNN algorithm requires substantial computation and memory, it is not suitable for embedded systems. To address this, we implemented a lightweight deep learning model, BNN, optimized for embedded systems on an FPGA, and verified the results achieving a classification accuracy of 90.33% and an execution time of 20ms.

Real time instruction classification system

  • Sang-Hoon Lee;Dong-Jin Kwon
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.16 no.3
    • /
    • pp.212-220
    • /
    • 2024
  • A recently the advancement of society, AI technology has made significant strides, especially in the fields of computer vision and voice recognition. This study introduces a system that leverages these technologies to recognize users through a camera and relay commands within a vehicle based on voice commands. The system uses the YOLO (You Only Look Once) machine learning algorithm, widely used for object and entity recognition, to identify specific users. For voice command recognition, a machine learning model based on spectrogram voice analysis is employed to identify specific commands. This design aims to enhance security and convenience by preventing unauthorized access to vehicles and IoT devices by anyone other than registered users. We converts camera input data into YOLO system inputs to determine if it is a person, Additionally, it collects voice data through a microphone embedded in the device or computer, converting it into time-domain spectrogram data to be used as input for the voice recognition machine learning system. The input camera image data and voice data undergo inference tasks through pre-trained models, enabling the recognition of simple commands within a limited space based on the inference results. This study demonstrates the feasibility of constructing a device management system within a confined space that enhances security and user convenience through a simple real-time system model. Finally our work aims to provide practical solutions in various application fields, such as smart homes and autonomous vehicles.

Banner Control Automation System Using YOLO and OpenCV (YOLO와 OpenCV기술을 활용한 현수막 단속 자동화 시스템 방안)

  • Dukwoen Kim;Jihoon Lee
    • Journal of the Semiconductor & Display Technology
    • /
    • v.22 no.4
    • /
    • pp.48-52
    • /
    • 2023
  • From the past to the present, banners are consistently used as effective advertising means. In the case of Korea, there are frequent situations in which hidden advertisements are installed. As a result, such hidden advertisement materials may damage urban aesthetics and moreover, incur unnecessary manpower consumption and waste of money. The proposed method classifies the detected banners into good banner and bad banner. The classification results are based on whether the relevant banners are installed in compliance with legal guidelines. In the process, YOLO and Open Computer Vision library are used to determine from various perspectives whether banners in CCTV images comply with the guidelines. YOLO is used to detect the banner area in CCTV images, and OpenCV is used to detect the color values in the area for color comparison. If a banner is detected in the video, the proposed method calculates the location of the banner and the distance from the designated bulletin to determine whether it was installed within the designated location, and then compares whether the color used in the banner is complied with local government guidelines.

  • PDF

A Study on Establishment Method of Smart Factory Dataset for Artificial Intelligence (인공지능형 스마트공장 데이터셋 구축 방법에 관한 연구)

  • Park, Youn-Soo;Lee, Sang-Deok;Choi, Jeong-Hun
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.21 no.5
    • /
    • pp.203-208
    • /
    • 2021
  • At the manufacturing site, workers have been operating by inputting materials into the manufacturing process and leaving input records according to the work instructions, but product LOT tracking has been not possible due to many omissions. Recently, it is being carried out as a system to automatically input materials using RFID-Tag. In particular, the initial automatic recognition rate was good at 97 percent by automatically generating input information through RACK (TAG) ID and RACK input time analysis, but the automatic recognition rate continues to decrease due to multi-material RACK, TAG loss, and new product input issues. It is expected that it will contribute to increasing speed and yield (normal product ratio) in the overall production process by improving automatic recognition rate and real-time monitoring through the establishment of artificial intelligent smart factory datasets.