• 제목/요약/키워드: Bounding Box Augmentation

검색결과 4건 처리시간 0.023초

Bounding Box CutMix와 표준화 거리 기반의 IoU를 통한 재활용품 탐지 (Recyclable Objects Detection via Bounding Box CutMix and Standardized Distance-based IoU)

  • 이해진;정희철
    • 대한임베디드공학회논문지
    • /
    • 제17권5호
    • /
    • pp.289-296
    • /
    • 2022
  • In this paper, we developed a deep learning-based recyclable object detection model. The model is developed based on YOLOv5 that is a one-stage detector. The deep learning model detects and classifies the recyclable object into 7 categories: paper, carton, can, glass, pet, plastic, and vinyl. We propose two methods for recyclable object detection models to solve problems during training. Bounding Box CutMix solved the no-objects training images problem of Mosaic, a data augmentation used in YOLOv5. Standardized Distance-based IoU replaced DIoU using a normalization factor that is not affected by the center point distance of the bounding boxes. The recyclable object detection model showed a final mAP performance of 0.91978 with Bounding Box CutMix and 0.91149 with Standardized Distance-based IoU.

Object Detection Model 적용성 확대를 위한 BoundingBox 이미지 증강 GUI 프로그램 연구 (Implementation and Design of Bounding Box Image Augmentation GUI Program for expanding Object Detection Models' applicability)

  • 전진영;민연아
    • 한국컴퓨터정보학회:학술대회논문집
    • /
    • 한국컴퓨터정보학회 2022년도 제66차 하계학술대회논문집 30권2호
    • /
    • pp.539-540
    • /
    • 2022
  • 본 논문에서는 Bounding Box가 포함된 증강 이미지 데이터셋을 손쉽게 생성할 수 있는 독립형 GUI 프로그램을 제안한다. 본 논문의 연구를 통하여 직관적인 마우스 클릭 동작만으로 적은 수의 이미지 파일과 annotation 파일로부터 필요한 만큼의 증강 이미지 데이터셋을 짧은 시간 내에 생성하고, 다양한 아키텍처의 학습용 이미지 데이터셋 증강에 적용할 수 있다.

  • PDF

딥 러닝 기반의 영상처리 기법을 이용한 겹침 돼지 분리 (Separation of Occluding Pigs using Deep Learning-based Image Processing Techniques)

  • 이한해솔;사재원;신현준;정용화;박대희;김학재
    • 한국멀티미디어학회논문지
    • /
    • 제22권2호
    • /
    • pp.136-145
    • /
    • 2019
  • The crowded environment of a domestic pig farm is highly vulnerable to the spread of infectious diseases such as foot-and-mouth disease, and studies have been conducted to automatically analyze behavior of pigs in a crowded pig farm through a video surveillance system using a camera. Although it is required to correctly separate occluding pigs for tracking each individual pigs, extracting the boundaries of the occluding pigs fast and accurately is a challenging issue due to the complicated occlusion patterns such as X shape and T shape. In this study, we propose a fast and accurate method to separate occluding pigs not only by exploiting the characteristics (i.e., one of the fast deep learning-based object detectors) of You Only Look Once, YOLO, but also by overcoming the limitation (i.e., the bounding box-based object detector) of YOLO with the test-time data augmentation of rotation. Experimental results with two-pigs occlusion patterns show that the proposed method can provide better accuracy and processing speed than one of the state-of-the-art widely used deep learning-based segmentation techniques such as Mask R-CNN (i.e., the performance improvement over Mask R-CNN was about 11 times, in terms of the accuracy/processing speed performance metrics).

다양한 재료에서 발생되는 연기 및 불꽃에 대한 YOLO 기반 객체 탐지 모델 성능 개선에 관한 연구 (Research on Improving the Performance of YOLO-Based Object Detection Models for Smoke and Flames from Different Materials )

  • 권희준;이보희;정해영
    • 한국전기전자재료학회논문지
    • /
    • 제37권3호
    • /
    • pp.261-273
    • /
    • 2024
  • This paper is an experimental study on the improvement of smoke and flame detection from different materials with YOLO. For the study, images of fires occurring in various materials were collected through an open dataset, and experiments were conducted by changing the main factors affecting the performance of the fire object detection model, such as the bounding box, polygon, and data augmentation of the collected image open dataset during data preprocessing. To evaluate the model performance, we calculated the values of precision, recall, F1Score, mAP, and FPS for each condition, and compared the performance of each model based on these values. We also analyzed the changes in model performance due to the data preprocessing method to derive the conditions that have the greatest impact on improving the performance of the fire object detection model. The experimental results showed that for the fire object detection model using the YOLOv5s6.0 model, data augmentation that can change the color of the flame, such as saturation, brightness, and exposure, is most effective in improving the performance of the fire object detection model. The real-time fire object detection model developed in this study can be applied to equipment such as existing CCTV, and it is believed that it can contribute to minimizing fire damage by enabling early detection of fires occurring in various materials.