• Title/Summary/Keyword: You Only Look Once

Search Result 112, Processing Time 0.027 seconds

Assessment of the Object Detection Ability of Interproximal Caries on Primary Teeth in Periapical Radiographs Using Deep Learning Algorithms (유치의 치근단 방사선 사진에서 딥 러닝 알고리즘을 이용한 모델의 인접면 우식증 객체 탐지 능력의 평가)

  • Hongju Jeon;Seonmi Kim;Namki Choi
    • Journal of the korean academy of Pediatric Dentistry
    • /
    • v.50 no.3
    • /
    • pp.263-276
    • /
    • 2023
  • The purpose of this study was to evaluate the performance of a model using You Only Look Once (YOLO) for object detection of proximal caries in periapical radiographs of children. A total of 2016 periapical radiographs in primary dentition were selected from the M6 database as a learning material group, of which 1143 were labeled as proximal caries by an experienced dentist using an annotation tool. After converting the annotations into a training dataset, YOLO was trained on the dataset using a single convolutional neural network (CNN) model. Accuracy, recall, specificity, precision, negative predictive value (NPV), F1-score, Precision-Recall curve, and AP (area under curve) were calculated for evaluation of the object detection model's performance in the 187 test datasets. The results showed that the CNN-based object detection model performed well in detecting proximal caries, with a diagnostic accuracy of 0.95, a recall of 0.94, a specificity of 0.97, a precision of 0.82, a NPV of 0.96, and an F1-score of 0.81. The AP was 0.83. This model could be a valuable tool for dentists in detecting carious lesions in periapical radiographs.

YOLO-based Traffic Signal Detection for Identifying the Violation of Motorbike Riders (YOLO 기반의 교통 신호등 인식을 통한 오토바이 운전자의 신호 위반 여부 확인)

  • Wahyutama, Aria Bisma;Hwang, Mintae
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.141-143
    • /
    • 2022
  • This paper presented a new technology to identify traffic violations of motorbike riders by detecting the traffic signal using You Only Look Once (YOLO) object detection. The hardware module that is mounted on the front of the motorbike consists of Raspberry Pi with a camera to run the YOLO object detection, a GPS module to acquire the motorcycle's coordinate, and a LoRa communication module to send the data to a cloud DB. The main goal of the software is to determine whether a motorbike has violated a traffic signal. This paper proposes a function to recognize the red traffic signal colour with its movement inside the camera angle and determine that the traffic signal violation happens if the traffic signal is moving to the right direction (the rider turns left) or moving to the top direction (the riders goes straight). Furthermore, if a motorbike rider is violated the signal, the rider's personal information (name, mobile phone number, etc), the snapshot of the violation situation, rider's location, and date/time will be sent to a cloud DB. The violation information will be delivered to the driver's smartphone as a push notification and the local police station to be used for issuing violation tickets, which is expected to prevent motorbike riders from violating traffic signals.

  • PDF

Drone-mounted fruit recognition algorithm and harvesting mechanism for automatic fruit harvesting (자동 과일 수확을 위한 드론 탑재형 과일 인식 알고리즘 및 수확 메커니즘)

  • Joo, Kiyoung;Hwang, Bohyun;Lee, Sangmin;Kim, Byungkyu;Baek, Joong-Hwan
    • Journal of Aerospace System Engineering
    • /
    • v.16 no.1
    • /
    • pp.49-55
    • /
    • 2022
  • The role of drones has been expanded to various fields such as agriculture, construction, and logistics. In particular, agriculture drones are emerging as an effective alternative to solve the problem of labor shortage and reduce the input cost. In this study therefore, we proposed the fruit recognition algorithm and harvesting mechanism for fruit harvesting drone system that can safely harvest fruits at high positions. In the fruit recognition algorithm, we employ "You-Only-Look-Once" which is a deep learning-based object detection algorithm and verify its feasibility by establishing a virtual simulation environment. In addition, we propose the fruit harvesting mechanism which can be operated by a single driving motor. The rotational motion of the motor is converted into a linear motion by the scotch yoke, and the opened gripper moves forward, grips a fruit and rotates it for harvesting. The feasibility of the proposed mechanism is verified by performing Multi-body dynamics analysis.

Automatic Collection of Production Performance Data Based on Multi-Object Tracking Algorithms (다중 객체 추적 알고리즘을 이용한 가공품 흐름 정보 기반 생산 실적 데이터 자동 수집)

  • Lim, Hyuna;Oh, Seojeong;Son, Hyeongjun;Oh, Yosep
    • The Journal of Society for e-Business Studies
    • /
    • v.27 no.2
    • /
    • pp.205-218
    • /
    • 2022
  • Recently, digital transformation in manufacturing has been accelerating. It results in that the data collection technologies from the shop-floor is becoming important. These approaches focus primarily on obtaining specific manufacturing data using various sensors and communication technologies. In order to expand the channel of field data collection, this study proposes a method to automatically collect manufacturing data based on vision-based artificial intelligence. This is to analyze real-time image information with the object detection and tracking technologies and to obtain manufacturing data. The research team collects object motion information for each frame by applying YOLO (You Only Look Once) and DeepSORT as object detection and tracking algorithms. Thereafter, the motion information is converted into two pieces of manufacturing data (production performance and time) through post-processing. A dynamically moving factory model is created to obtain training data for deep learning. In addition, operating scenarios are proposed to reproduce the shop-floor situation in the real world. The operating scenario assumes a flow-shop consisting of six facilities. As a result of collecting manufacturing data according to the operating scenarios, the accuracy was 96.3%.

Detection and Grading of Compost Heap Using UAV and Deep Learning (UAV와 딥러닝을 활용한 야적퇴비 탐지 및 관리등급 산정)

  • Miso Park;Heung-Min Kim;Youngmin Kim;Suho Bak;Tak-Young Kim;Seon Woong Jang
    • Korean Journal of Remote Sensing
    • /
    • v.40 no.1
    • /
    • pp.33-43
    • /
    • 2024
  • This research assessed the applicability of the You Only Look Once (YOLO)v8 and DeepLabv3+ models for the effective detection of compost heaps, identified as a significant source of non-point source pollution. Utilizing high-resolution imagery acquired through Unmanned Aerial Vehicles(UAVs), the study conducted a comprehensive comparison and analysis of the quantitative and qualitative performances. In the quantitative evaluation, the YOLOv8 model demonstrated superior performance across various metrics, particularly in its ability to accurately distinguish the presence or absence of covers on compost heaps. These outcomes imply that the YOLOv8 model is highly effective in the precise detection and classification of compost heaps, thereby providing a novel approach for assessing the management grades of compost heaps and contributing to non-point source pollution management. This study suggests that utilizing UAVs and deep learning technologies for detecting and managing compost heaps can address the constraints linked to traditional field survey methods, thereby facilitating the establishment of accurate and effective non-point source pollution management strategies, and contributing to the safeguarding of aquatic environments.

Improving the Performance of Deep-Learning-Based Ground-Penetrating Radar Cavity Detection Model using Data Augmentation and Ensemble Techniques (데이터 증강 및 앙상블 기법을 이용한 딥러닝 기반 GPR 공동 탐지 모델 성능 향상 연구)

  • Yonguk Choi;Sangjin Seo;Hangilro Jang;Daeung Yoon
    • Geophysics and Geophysical Exploration
    • /
    • v.26 no.4
    • /
    • pp.211-228
    • /
    • 2023
  • Ground-penetrating radar (GPR) surveys are commonly used to monitor embankments, which is a nondestructive geophysical method. The results of GPR surveys can be complex, depending on the situation, and data processing and interpretation are subject to expert experiences, potentially resulting in false detection. Additionally, this process is time-intensive. Consequently, various studies have been undertaken to detect cavities in GPR survey data using deep learning methods. Deep-learning-based approaches require abundant data for training, but GPR field survey data are often scarce due to cost and other factors constaining field studies. Therefore, in this study, a deep- learning-based model was developed for embankment GPR survey cavity detection using data augmentation strategies. A dataset was constructed by collecting survey data over several years from the same embankment. A you look only once (YOLO) model, commonly used in computer vision for object detection, was employed for this purpose. By comparing and analyzing various strategies, the optimal data augmentation approach was determined. After initial model development, a stepwise process was employed, including box clustering, transfer learning, self-ensemble, and model ensemble techniques, to enhance the final model performance. The model performance was evaluated, with the results demonstrating its effectiveness in detecting cavities in embankment GPR survey data.

Separation of Occluding Pigs using Deep Learning-based Image Processing Techniques (딥 러닝 기반의 영상처리 기법을 이용한 겹침 돼지 분리)

  • Lee, Hanhaesol;Sa, Jaewon;Shin, Hyunjun;Chung, Youngwha;Park, Daihee;Kim, Hakjae
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.2
    • /
    • pp.136-145
    • /
    • 2019
  • The crowded environment of a domestic pig farm is highly vulnerable to the spread of infectious diseases such as foot-and-mouth disease, and studies have been conducted to automatically analyze behavior of pigs in a crowded pig farm through a video surveillance system using a camera. Although it is required to correctly separate occluding pigs for tracking each individual pigs, extracting the boundaries of the occluding pigs fast and accurately is a challenging issue due to the complicated occlusion patterns such as X shape and T shape. In this study, we propose a fast and accurate method to separate occluding pigs not only by exploiting the characteristics (i.e., one of the fast deep learning-based object detectors) of You Only Look Once, YOLO, but also by overcoming the limitation (i.e., the bounding box-based object detector) of YOLO with the test-time data augmentation of rotation. Experimental results with two-pigs occlusion patterns show that the proposed method can provide better accuracy and processing speed than one of the state-of-the-art widely used deep learning-based segmentation techniques such as Mask R-CNN (i.e., the performance improvement over Mask R-CNN was about 11 times, in terms of the accuracy/processing speed performance metrics).

Freeway Bus-Only Lane Enforcement System Using Infrared Image Processing Technique (적외선 영상검지 기술을 활용한 고속도로 버스전용차로 단속시스템 개발)

  • Jang, Jinhwan
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.21 no.5
    • /
    • pp.67-77
    • /
    • 2022
  • An automatic freeway bus-only lane enforcement system was developed and assessed in a real-world environment. Observation of a bus-only lane on the Youngdong freeway, South Korea, revealed that approximately 99% of the vehicles violated the high-occupancy vehicle (HOV) lane regulation. However, the current enforcement by the police not only exhibits a low enforcement rate, but also induces unnecessary safety and delay concerns. Since vehicles with six passengers or higher are permitted to enter freeway bus-only lanes, identifying the number of passengers in a vehicle is a core technology required for a freeway bus-only lane enforcement system. To that end, infrared cameras and the You Only Look Once (YOLOv5) deep learning algorithm were utilized. For assessment of the performance of the developed system, two environments, including a controlled test-bed and a real-world freeway, were used. As a result, the performances under the test-bed and the real-world environments exhibited 7% and 8% errors, respectively, indicating satisfactory outcomes. The developed system would contribute to an efficient freeway bus-only lane operations as well as eliminate safety and delay concerns caused by the current manual enforcement procedures.

Implementation of Drowsy Prevention System Using Arduino and YOLO (아두이노와 YOLO를 이용한 졸음 방지 시스템 구현)

  • Lee, Hyun-Ae;Shin, Seong-Yoon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.7
    • /
    • pp.917-922
    • /
    • 2021
  • In modern society, deaths and property damage due to drowsiness occur every year enormously. Methods to reduce such damage are being studied a lot in all walks of life, and research on preventing drowsy driving is particularly active in automobiles. In this paper, as an Arduino-based water gun firing system that learns open and closed eyes using YOLO, we propose a drowsy prevention system that fires a water gun when the duration of the closed eye exceeds a certain time. This system can be applied and used in various fields, but especially when applied to a car, it is not necessary to purchase expensive specifications and if you pay a little attention, you can reduce accidents caused by drowsy driving by 100% at a very low cost. In addition, it can be said that it is an independent system that overcomes different specifications for each company.

Separation of Touching Pigs using YOLO-based Bounding Box (YOLO 기반 외곽 사각형을 이용한 근접 돼지 분리)

  • Seo, J.;Ju, M.;Choi, Y.;Lee, J.;Chung, Y.;Park, D.
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.2
    • /
    • pp.77-86
    • /
    • 2018
  • Although separation of touching pigs in real-time is an important issue for a 24-h pig monitoring system, it is challenging to separate accurately the touching pigs in a crowded pig room. In this study, we propose a separation method for touching pigs using the information generated from Convolutional Neural Network(CNN). Especially, we apply one of the CNN-based object detection methods(i.e., You Look Only Once, YOLO) to solve the touching objects separation problem in an active manner. First, we evaluate and select the bounding boxes generated from YOLO, and then separate touching pigs by analyzing the relations between the selected bounding boxes. Our experimental results show that the proposed method is more effective than widely-used methods for separating touching pigs, in terms of both accuracy and execution time.