• Title/Summary/Keyword: Real-Time Computer Vision

Search Result 352, Processing Time 0.033 seconds

A Study on the Development of Drowsiness Warning System for a Drowsy Driver (졸음 운전자를 위한 졸음 각성 시스템의 개발에 관한 연구)

  • Chong, K.H.;Kim, H.S.;Lee, J.S.;Kim, B.J.;Kim, D.W.;Kim, N.G.
    • Proceedings of the KOSOMBE Conference
    • /
    • v.1996 no.11
    • /
    • pp.90-94
    • /
    • 1996
  • We studied the problem of driver's low vigilance state which is related to the one reason of traffic accidents. In this paper, we developed the drowsiness warning system for a drowsy driver. To extract the eyes and mouth from the driver's facial image in real time, a computer vision method was used. The eye blink duration and yawning were used as measurement parameters of drowsiness detection. When the drowsy state of a driver was detected, the driver was refreshed by the scent generator and the alarm. Also, the driver's bio-signal was acquired and analyzed to measure the vigilance state.

  • PDF

A Study on Architectural Image Generation using Artificial Intelligence Algorithm - A Fundamental Study on the Generation of Due Diligence Images Based on Architectural Sketch - (인공지능 알고리즘을 활용한 건축 이미지 생성에 관한 연구 - 건축 스케치 기반의 실사 이미지 생성을 위한 기초적 연구 -)

  • Han, Sang-Kook;Shin, Dong-Youn
    • Journal of KIBIM
    • /
    • v.11 no.2
    • /
    • pp.54-59
    • /
    • 2021
  • In the process of designing a building, the process of expressing the designer's ideas through images is essential. However, it is expensive and time consuming for a designer to analyze every individual case image to generate a hypothetical design. This study aims to visualize the basic design draft sketch made by the designer as a real image using the Generative Adversarial Network (GAN) based on the continuously accumulated architectural case images. Through this, we proposed a method to build an automated visualization environment using artificial intelligence and to visualize the architectural idea conceived by the designer in the architectural planning stage faster and cheaper than in the past. This study was conducted using approximately 20,000 images. In our study, the GAN algorithm allowed us to represent primary materials and shades within 2 seconds, but lacked accuracy in material and shading representation. We plan to add image data in the future to address this in a follow-up study.

Light-weight Gender Classification and Age Estimation based on Ensemble Multi-tasking Deep Learning (앙상블 멀티태스킹 딥러닝 기반 경량 성별 분류 및 나이별 추정)

  • Huy Tran, Quoc Bao;Park, JongHyeon;Chung, SunTae
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.1
    • /
    • pp.39-51
    • /
    • 2022
  • Image-based gender classification and age estimation of human are classic problems in computer vision. Most of researches in this field focus just only one task of either gender classification or age estimation and most of the reported methods for each task focus on accuracy performance and are not computationally light. Thus, running both tasks together simultaneously on low cost mobile or embedded systems with limited cpu processing speed and memory capacity are practically prohibited. In this paper, we propose a novel light-weight gender classification and age estimation method based on ensemble multitasking deep learning with light-weight processing neural network architecture, which processes both gender classification and age estimation simultaneously and in real-time even for embedded systems. Through experiments over various well-known datasets, it is shown that the proposed method performs comparably to the state-of-the-art gender classification and/or age estimation methods with respect to accuracy and runs fast enough (average 14fps) on a Jestson Nano embedded board.

A Study on Image Labeling Technique for Deep-Learning-Based Multinational Tanks Detection Model

  • Kim, Taehoon;Lim, Dongkyun
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.14 no.4
    • /
    • pp.58-63
    • /
    • 2022
  • Recently, the improvement of computational processing ability due to the rapid development of computing technology has greatly advanced the field of artificial intelligence, and research to apply it in various domains is active. In particular, in the national defense field, attention is paid to intelligent recognition among machine learning techniques, and efforts are being made to develop object identification and monitoring systems using artificial intelligence. To this end, various image processing technologies and object identification algorithms are applied to create a model that can identify friendly and enemy weapon systems and personnel in real-time. In this paper, we conducted image processing and object identification focused on tanks among various weapon systems. We initially conducted processing the tanks' image using a convolutional neural network, a deep learning technique. The feature map was examined and the important characteristics of the tanks crucial for learning were derived. Then, using YOLOv5 Network, a CNN-based object detection network, a model trained by labeling the entire tank and a model trained by labeling only the turret of the tank were created and the results were compared. The model and labeling technique we proposed in this paper can more accurately identify the type of tank and contribute to the intelligent recognition system to be developed in the future.

YOLOv5 based Anomaly Detection for Subway Safety Management Using Dilated Convolution

  • Nusrat Jahan Tahira;Ju-Ryong Park;Seung-Jin Lim;Jang-Sik Park
    • Journal of the Korean Society of Industry Convergence
    • /
    • v.26 no.2_1
    • /
    • pp.217-223
    • /
    • 2023
  • With the rapid advancement of technologies, need for different research fields where this technology can be used is also increasing. One of the most researched topic in computer vision is object detection, which has widely been implemented in various fields which include healthcare, video surveillance and education. The main goal of object detection is to identify and categorize all the objects in a target environment. Specifically, methods of object detection consist of a variety of significant techniq ues, such as image processing and patterns recognition. Anomaly detection is a part of object detection, anomalies can be found various scenarios for example crowded places such as subway stations. An abnormal event can be assumed as a variation from the conventional scene. Since the abnormal event does not occur frequently, the distribution of normal and abnormal events is thoroughly imbalanced. In terms of public safety, abnormal events should be avoided and therefore immediate action need to be taken. When abnormal events occur in certain places, real time detection is required to prevent and protect the safety of the people. To solve the above problems, we propose a modified YOLOv5 object detection algorithm by implementing dilated convolutional layers which achieved 97% mAP50 compared to other five different models of YOLOv5. In addition to this, we also created a simple mobile application to avail the abnormal event detection on mobile phones.

Leveraging Deep Learning and Farmland Fertility Algorithm for Automated Rice Pest Detection and Classification Model

  • Hussain. A;Balaji Srikaanth. P
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.4
    • /
    • pp.959-979
    • /
    • 2024
  • Rice pest identification is essential in modern agriculture for the health of rice crops. As global rice consumption rises, yields and quality must be maintained. Various methodologies were employed to identify pests, encompassing sensor-based technologies, deep learning, and remote sensing models. Visual inspection by professionals and farmers remains essential, but integrating technology such as satellites, IoT-based sensors, and drones enhances efficiency and accuracy. A computer vision system processes images to detect pests automatically. It gives real-time data for proactive and targeted pest management. With this motive in mind, this research provides a novel farmland fertility algorithm with a deep learning-based automated rice pest detection and classification (FFADL-ARPDC) technique. The FFADL-ARPDC approach classifies rice pests from rice plant images. Before processing, FFADL-ARPDC removes noise and enhances contrast using bilateral filtering (BF). Additionally, rice crop images are processed using the NASNetLarge deep learning architecture to extract image features. The FFA is used for hyperparameter tweaking to optimise the model performance of the NASNetLarge, which aids in enhancing classification performance. Using an Elman recurrent neural network (ERNN), the model accurately categorises 14 types of pests. The FFADL-ARPDC approach is thoroughly evaluated using a benchmark dataset available in the public repository. With an accuracy of 97.58, the FFADL-ARPDC model exceeds existing pest detection methods.

A Study on Adaptive Skin Extraction using a Gradient Map and Saturation Features (경사도 맵과 채도 특징을 이용한 적응적 피부영역 검출에 관한 연구)

  • Hwang, Dae-Dong;Lee, Keun-Soo
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.15 no.7
    • /
    • pp.4508-4515
    • /
    • 2014
  • Real-time body detection has been researched actively. On the other hand, the detection rate of color distorted images is low because most existing detection methods use static skin color model. Therefore, this paper proposes a new method for detecting the skin color region using a gradient map and saturation features. The basic procedure of the proposed method sequentially consists of creating a gradient map, extracting a gradient feature of skin regions, noise removal using the saturation features of skin, creating a cluster for extraction regions, detecting skin regions using cluster information, and verifying the results. This method uses features other than the color to strengthen skin detection not affected by light, race, age, individual features, etc. The results of the detection rate showed that the proposed method is 10% or more higher than the traditional methods.

Deep Neural Networks Learning based on Multiple Loss Functions for Both Person and Vehicles Re-Identification (사람과 자동차 재인식이 가능한 다중 손실함수 기반 심층 신경망 학습)

  • Kim, Kyeong Tae;Choi, Jae Young
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.8
    • /
    • pp.891-902
    • /
    • 2020
  • The Re-Identification(Re-ID) is one of the most popular researches in the field of computer vision due to a variety of applications. To achieve a high-level re-identification performance, recently other methods have developed the deep learning based networks that are specialized for only person or vehicle. However, most of the current methods are difficult to be used in real-world applications that require re-identification of both person and vehicle at the same time. To overcome this limitation, this paper proposes a deep neural network learning method that combines triplet and softmax loss to improve performance and re-identify people and vehicles simultaneously. It's possible to learn the detailed difference between the identities(IDs) by combining the softmax loss with the triplet loss. In addition, weights are devised to avoid bias in one-side loss when combining. We used Market-1501 and DukeMTMC-reID datasets, which are frequently used to evaluate person re-identification experiments. Moreover, the vehicle re-identification experiment was evaluated by using VeRi-776 and VehicleID datasets. Since the proposed method does not designed for a neural network specialized for a specific object, it can re-identify simultaneously both person and vehicle. To demonstrate this, an experiment was performed by using a person and vehicle re-identification dataset together.

A Study on Environmentally Adaptive Real-Time Lane Recognition Using Car Black Box Video Images (차량용 블랙박스 영상을 이용한 환경적응적 실시간 차선인식 연구)

  • Park, Daehyuck;Lee, Jung-hun;Seo, Jeong Goo;Kim, Jihyung;Jin, Seogsig;Yun, Tae-sup;Lee, Hye;Xu, Bin;Lim, Younghwan
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2015.07a
    • /
    • pp.187-190
    • /
    • 2015
  • 주행 중 차선 이탈 경고 시스템은 사고 발생 예방 차원에서 매우 높은 효과가 인정되어서 차선이탈 경고 장치(LDWS) 제품들이 출시되고 있다. 본 논문은 블랙박스의 영상을 이용하여 차선 검출에 정확도를 향상하기 위한 알고리즘을 연구한 것으로 특히 차량에 장착되어 있는 블랙박스 영상을 영상 변환 없이, 실시간 소프트웨어 만 으로 처리할 수 있는 알고리즘을 연구한다. 차선인식을 위한 최적의 영상 ROI를 결정하고, 차선 인식 정확도를 향상하기 위한 전 처리 과정을 적용하고, 동영상의 연속성을 잘못된 차선인식에 대한 보정, 인식이 되지 않는 차선에 대한 후보 차선 추천 알고리즘과 시점 변환에 의한 야간, 곡선 도로에 대한 오인식율을 최소화 하는 방법을 제안한다. 도로주행의 다양한 환경에 대한 실험을 진행했으며, 각각의 방법 적용에 의한 오인식율의 감소와 많은 인식 알고리즘 적용에 의한 처리 속도 저하를 개선하기 위한 연구를 진행했으며, 본 논문은 블랙박스 영상을 이용하여 주행 차선 인식을 위한 최적 알고리즘을 제안한다.

  • PDF

A Study on the Construction of Near-Real Time Drone Image Preprocessing System to use Drone Data in Disaster Monitoring (재난재해 분야 드론 자료 활용을 위한 준 실시간 드론 영상 전처리 시스템 구축에 관한 연구)

  • Joo, Young-Do
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.18 no.3
    • /
    • pp.143-149
    • /
    • 2018
  • Recently, due to the large-scale damage of natural disasters caused by global climate change, a monitoring system applying remote sensing technology is being constructed in disaster areas. Among remote sensing platforms, the drone has been actively used in the private sector due to recent technological developments, and has been applied in the disaster areas owing to advantages such as timeliness and economical efficiency. This paper deals with the development of a preprocessing system that can map the drone image data in a near-real time manner as a basis for constructing the disaster monitoring system using the drones. For the research purpose, our system is based on the SURF algorithm which is one of the computer vision technologies. This system aims to performs the desired correction through the feature point matching technique between reference images and shot images. The study area is selected as the lower part of the Gahwa River and the Daecheong dam basin. The former area has many characteristic points for matching whereas the latter area has a relatively low number of difference, so it is possible to effectively test whether the system can be applied in various environments. The results show that the accuracy of the geometric correction is 0.6m and 1.7m respectively, in both areas, and the processing time is about 30 seconds per 1 scene. This indicates that the applicability of this study may be high in disaster areas requiring timeliness. However, in case of no reference image or low-level accuracy, the results entail the limit of the decreased calibration.