• 제목/요약/키워드: YOLO (You Only Look Once)

검색결과 89건 처리시간 0.025초

딥러닝 기반 육상기인 부유쓰레기 탐지 모델 성능 비교 및 현장 적용성 평가 (A Performance Comparison of Land-Based Floating Debris Detection Based on Deep Learning and Its Field Applications)

  • 박수호;장선웅;김흥민;김탁영;예건희
    • 대한원격탐사학회지
    • /
    • 제39권2호
    • /
    • pp.193-205
    • /
    • 2023
  • 집중강우 시 육상으로부터 다량으로 유입된 부유쓰레기는 사회, 경제적 및 환경적으로 부정적인 영향을 주고 있으나 부유쓰레기 집적 구간 및 발생량에 대한 모니터링 체계는 미흡한 실정이다. 최근 인공지능 기술의 발달로 드론 영상과 딥러닝 기반 객체탐지 모델을 활용하여 수계 내 광범위한 지역을 신속하고 효율적인 연구의 필요성이 요구되고 있다. 본 연구에서는 육상기인 부유쓰레기의 효율적인 탐지 기법을 제시하기 위해 드론 영상뿐만 아니라 다양한 이미지를 확보하여 You Only Look Once (YOLO)v5s와 최근에 개발된 YOLO7 및 YOLOv8s로 학습하여 모델별로 성능을 비교하였다. 각 모델의 정성적인 성능 평가 결과, 세 모델 모두 일반적인 상황에서 탐지성능이 우수한 것으로 나타났으나, 이미지의 노출이 심하거나 수면의 태양광 반사가 심한 경우 YOLOv8s 모델에서 대상물을 누락 또는 중복 탐지하는 사례가 나타났다. 정량적인 성능 평가 결과, YOLOv7의 mean Average Precision (intersection over union, IoU 0.5)이 0.940으로 YOLOv5s (0.922)와 YOLOvs8(0.922)보다 좋은 성능을 나타냈다. 데이터 품질에 따른 모델의 성능 비교하기 위해 색상 및 고주파 성분에 왜곡을 발생시킨 결과, YOLOv8s 모델의 성능 저하가 가장 뚜렷하게 나타났으며, YOLOv7 모델이 가장 낮은 성능 저하 폭을 보였다. 이를 통해 수면 위에 존재하는 부유쓰레기 탐지에 있어서 YOLOv7 모델이 YOLOv5s와 YOLOv8s 모델에 비해 강인한 모델임을 확인하였다. 본 연구에서 제안하는 딥러닝 기반 부유쓰레기 탐지 기법은 부유쓰레기의 성상별 분포 현황을 공간적으로 파악할 수 있어 향후 정화작업 계획수립에 기여할 수 있을 것으로 판단된다.

YOLOv5에서 가상 번호판 생성을 통한 차량 번호판 인식 시스템에 관한 연구 (A Study on Vehicle License Plate Recognition System through Fake License Plate Generator in YOLOv5)

  • 하상현;정석찬;전영준;장문석
    • 한국산업융합학회 논문집
    • /
    • 제24권6_2호
    • /
    • pp.699-706
    • /
    • 2021
  • Existing license plate recognition system is used as an optical character recognition method, but a method of using deep learning has been proposed in recent studies because it has problems with image quality and Korean misrecognition. This requires a lot of data collection, but the collection of license plates is not easy to collect due to the problem of the Personal Information Protection Act, and labeling work to designate the location of individual license plates is required, but it also requires a lot of time. Therefore, in this paper, to solve this problem, five types of license plates were created using a virtual Korean license plate generation program according to the notice of the Ministry of Land, Infrastructure and Transport. And the generated license plate is synthesized in the license plate part of collectable vehicle images to construct 10,147 learning data to be used in deep learning. The learning data classifies license plates, Korean, and numbers into individual classes and learn using YOLOv5. Since the proposed method recognizes letters and numbers individually, if the font does not change, it can be recognized even if the license plate standard changes or the number of characters increases. As a result of the experiment, an accuracy of 96.82% was obtained, and it can be applied not only to the learned license plate but also to new types of license plates such as new license plates and eco-friendly license plates.

A deep learning-based approach for feeding behavior recognition of weanling pigs

  • Kim, MinJu;Choi, YoHan;Lee, Jeong-nam;Sa, SooJin;Cho, Hyun-chong
    • Journal of Animal Science and Technology
    • /
    • 제63권6호
    • /
    • pp.1453-1463
    • /
    • 2021
  • Feeding is the most important behavior that represents the health and welfare of weanling pigs. The early detection of feed refusal is crucial for the control of disease in the initial stages and the detection of empty feeders for adding feed in a timely manner. This paper proposes a real-time technique for the detection and recognition of small pigs using a deep-leaning-based method. The proposed model focuses on detecting pigs on a feeder in a feeding position. Conventional methods detect pigs and then classify them into different behavior gestures. In contrast, in the proposed method, these two tasks are combined into a single process to detect only feeding behavior to increase the speed of detection. Considering the significant differences between pig behaviors at different sizes, adaptive adjustments are introduced into a you-only-look-once (YOLO) model, including an angle optimization strategy between the head and body for detecting a head in a feeder. According to experimental results, this method can detect the feeding behavior of pigs and screen non-feeding positions with 95.66%, 94.22%, and 96.56% average precision (AP) at an intersection over union (IoU) threshold of 0.5 for YOLOv3, YOLOv4, and an additional layer and with the proposed activation function, respectively. Drinking behavior was detected with 86.86%, 89.16%, and 86.41% AP at a 0.5 IoU threshold for YOLOv3, YOLOv4, and the proposed activation function, respectively. In terms of detection and classification, the results of our study demonstrate that the proposed method yields higher precision and recall compared to conventional methods.

Automatic identification and analysis of multi-object cattle rumination based on computer vision

  • Yueming Wang;Tiantian Chen;Baoshan Li;Qi Li
    • Journal of Animal Science and Technology
    • /
    • 제65권3호
    • /
    • pp.519-534
    • /
    • 2023
  • Rumination in cattle is closely related to their health, which makes the automatic monitoring of rumination an important part of smart pasture operations. However, manual monitoring of cattle rumination is laborious and wearable sensors are often harmful to animals. Thus, we propose a computer vision-based method to automatically identify multi-object cattle rumination, and to calculate the rumination time and number of chews for each cow. The heads of the cattle in the video were initially tracked with a multi-object tracking algorithm, which combined the You Only Look Once (YOLO) algorithm with the kernelized correlation filter (KCF). Images of the head of each cow were saved at a fixed size, and numbered. Then, a rumination recognition algorithm was constructed with parameters obtained using the frame difference method, and rumination time and number of chews were calculated. The rumination recognition algorithm was used to analyze the head image of each cow to automatically detect multi-object cattle rumination. To verify the feasibility of this method, the algorithm was tested on multi-object cattle rumination videos, and the results were compared with the results produced by human observation. The experimental results showed that the average error in rumination time was 5.902% and the average error in the number of chews was 8.126%. The rumination identification and calculation of rumination information only need to be performed by computers automatically with no manual intervention. It could provide a new contactless rumination identification method for multi-cattle, which provided technical support for smart pasture.

Multi-Class Multi-Object Tracking in Aerial Images Using Uncertainty Estimation

  • Hyeongchan Ham;Junwon Seo;Junhee Kim;Chungsu Jang
    • 대한원격탐사학회지
    • /
    • 제40권1호
    • /
    • pp.115-122
    • /
    • 2024
  • Multi-object tracking (MOT) is a vital component in understanding the surrounding environments. Previous research has demonstrated that MOT can successfully detect and track surrounding objects. Nonetheless, inaccurate classification of the tracking objects remains a challenge that needs to be solved. When an object approaching from a distance is recognized, not only detection and tracking but also classification to determine the level of risk must be performed. However, considering the erroneous classification results obtained from the detection as the track class can lead to performance degradation problems. In this paper, we discuss the limitations of classification in tracking under the classification uncertainty of the detector. To address this problem, a class update module is proposed, which leverages the class uncertainty estimation of the detector to mitigate the classification error of the tracker. We evaluated our approach on the VisDrone-MOT2021 dataset,which includes multi-class and uncertain far-distance object tracking. We show that our method has low certainty at a distant object, and quickly classifies the class as the object approaches and the level of certainty increases.In this manner, our method outperforms previous approaches across different detectors. In particular, the You Only Look Once (YOLO)v8 detector shows a notable enhancement of 4.33 multi-object tracking accuracy (MOTA) in comparison to the previous state-of-the-art method. This intuitive insight improves MOT to track approaching objects from a distance and quickly classify them.

차량 내 영상 센서 기반 고속도로 돌발상황 검지 정밀도 평가 (Precision Evaluation of Expressway Incident Detection Based on Dash Cam)

  • 남상기;정연식
    • 한국ITS학회 논문지
    • /
    • 제22권6호
    • /
    • pp.114-123
    • /
    • 2023
  • 컴퓨터 비전(Computer Vision: CV) 기술 발전으로 폐쇄회로 TV(Closed-Circuit television: CCTV)와 같은 영상 센서로 돌발상황을 검지하고 있다. 그러나 현재 이러한 기술은 대부분 고정식 영상 센서를 기반으로 한다. 따라서 고정식 장비의 영상 범위가 닿지 않는 음영지역의 돌발상황 검지에는 한계가 존재해왔다. 최근 엣지 컴퓨팅(Edge-computing) 기술의 발전으로 이동식 영상정보의 실시간 분석이 가능해졌다. 본 연구는 차량 내 설치된 이동식 영상 센서(dashboard camera 혹은 dash cam)에 컴퓨터 비전 기술을 도입하여 고속도로에서 실시간으로 돌발상황 검지 가능성에 대해 평가하는 것이 목적이다. 이를 위해 한국도로공사 순찰차량에 장착된 dash cam에서 수집된 4,388건의 스틸 프레임 데이터 기반으로 학습데이터를 구축하였으며, YOLO(You Only Look Once) 알고리즘을 활용하여 분석하였다. 분석 결과 객체 모두 예측 정밀도가 70% 이상으로 나타났고, 교통사고는 약 85%의 정밀도를 보였다. 또한 mAP(mean Average Precision)의 경우 0.769로 나타났고, 객체별 AP(Average Precision)를 보면 교통사고가 0.904로 가장 높게 나타났고, 낙하물이 0.629로 가장 낮게 나타났다.

Transfer learning in a deep convolutional neural network for implant fixture classification: A pilot study

  • Kim, Hak-Sun;Ha, Eun-Gyu;Kim, Young Hyun;Jeon, Kug Jin;Lee, Chena;Han, Sang-Sun
    • Imaging Science in Dentistry
    • /
    • 제52권2호
    • /
    • pp.219-224
    • /
    • 2022
  • Purpose: This study aimed to evaluate the performance of transfer learning in a deep convolutional neural network for classifying implant fixtures. Materials and Methods: Periapical radiographs of implant fixtures obtained using the Superline (Dentium Co. Ltd., Seoul, Korea), TS III(Osstem Implant Co. Ltd., Seoul, Korea), and Bone Level Implant(Institut Straumann AG, Basel, Switzerland) systems were selected from patients who underwent dental implant treatment. All 355 implant fixtures comprised the total dataset and were annotated with the name of the system. The total dataset was split into a training dataset and a test dataset at a ratio of 8 to 2, respectively. YOLOv3 (You Only Look Once version 3, available at https://pjreddie.com/darknet/yolo/), a deep convolutional neural network that has been pretrained with a large image dataset of objects, was used to train the model to classify fixtures in periapical images, in a process called transfer learning. This network was trained with the training dataset for 100, 200, and 300 epochs. Using the test dataset, the performance of the network was evaluated in terms of sensitivity, specificity, and accuracy. Results: When YOLOv3 was trained for 200 epochs, the sensitivity, specificity, accuracy, and confidence score were the highest for all systems, with overall results of 94.4%, 97.9%, 96.7%, and 0.75, respectively. The network showed the best performance in classifying Bone Level Implant fixtures, with 100.0% sensitivity, specificity, and accuracy. Conclusion: Through transfer learning, high performance could be achieved with YOLOv3, even using a small amount of data.

Real-Time Comprehensive Assistance for Visually Impaired Navigation

  • Amal Al-Shahrani;Amjad Alghamdi;Areej Alqurashi;Raghad Alzahrani;Nuha imam
    • International Journal of Computer Science & Network Security
    • /
    • 제24권5호
    • /
    • pp.1-10
    • /
    • 2024
  • Individuals with visual impairments face numerous challenges in their daily lives, with navigating streets and public spaces being particularly daunting. The inability to identify safe crossing locations and assess the feasibility of crossing significantly restricts their mobility and independence. Globally, an estimated 285 million people suffer from visual impairment, with 39 million categorized as blind and 246 million as visually impaired, according to the World Health Organization. In Saudi Arabia alone, there are approximately 159 thousand blind individuals, as per unofficial statistics. The profound impact of visual impairments on daily activities underscores the urgent need for solutions to improve mobility and enhance safety. This study aims to address this pressing issue by leveraging computer vision and deep learning techniques to enhance object detection capabilities. Two models were trained to detect objects: one focused on street crossing obstacles, and the other aimed to search for objects. The first model was trained on a dataset comprising 5283 images of road obstacles and traffic signals, annotated to create a labeled dataset. Subsequently, it was trained using the YOLOv8 and YOLOv5 models, with YOLOv5 achieving a satisfactory accuracy of 84%. The second model was trained on the COCO dataset using YOLOv5, yielding an impressive accuracy of 94%. By improving object detection capabilities through advanced technology, this research seeks to empower individuals with visual impairments, enhancing their mobility, independence, and overall quality of life.

딥러닝 객체 탐지 기술을 사용한 스마트 쇼핑카트의 구현 (Implementation of Smart Shopping Cart using Object Detection Method based on Deep Learning)

  • 오진선;천인국
    • 한국산학기술학회논문지
    • /
    • 제21권7호
    • /
    • pp.262-269
    • /
    • 2020
  • 최근 다양한 쇼핑 환경에서 결제에 소요되는 시간을 줄이기 위한 많은 시도들이 이루어지고 있다. 또한 4차 산업혁명시대에 들어서면서 인공지능 기술이 고도화되고 있으며, IoT 장비들은 더욱 소형화되고 저렴해져서 이 두 가지 기술을 융합시킴으로써 사용자의 시간을 절약할, 인간을 대신하는 무인 환경을 구축하는 것에 대한 접근이 용이해졌다. 본 논문에서는 저가 IoT 장비들과 딥러닝 객체 탐지 기술을 기반으로 하는 스마트 쇼핑카트 시스템을 제안한다. 제안된 스마트 카트 시스템은 실시간 상품 인식을 위한 카메라와 라즈베리파이, 트리거 역할을 하는 초음파 센서, 상품이 쇼핑카트에 들어온 것인지 나간 것인지를 판단하기 위한 무게 센서, 가상의 장바구니에 대한 UI를 제공하는 스마트폰 어플리케이션, 학습된 데이터가 저장되는 딥러닝 서버로 구성된다. 각 모듈 간의 통신은 TCP/IP 네트워크 및 HTTP 네트워크로 이루어지며, 서버의 상품 인식을 위해서는 객체탐지 기술이 구현된 YOLO darknet 라이브러리를 사용한다. 사용자는 스마트폰의 앱을 통하여 스마트 카트에 넣은 물건들의 목록을 점검하고 자동으로 결제할 수 있다. 본 논문에서 제안된 스마트 카트 시스템은 가성비가 높은 무인 상점을 구현하는데 응용될 수 있다.