• Title/Summary/Keyword: Camera-based Recognition

Search Result 593, Processing Time 0.03 seconds

Sensitivity Analysis of Excavator Activity Recognition Performance based on Surveillance Camera Locations

  • Yejin SHIN;Seungwon SEO;Choongwan KOO
    • International conference on construction engineering and project management
    • /
    • 2024.07a
    • /
    • pp.1282-1282
    • /
    • 2024
  • Given the widespread use of intelligent surveillance cameras at construction sites, recent studies have introduced vision-based deep learning approaches. These studies have focused on enhancing the performance of vision-based excavator activity recognition to automatically monitor productivity metrics such as activity time and work cycle. However, acquiring a large amount of training data, i.e., videos captured from actual construction sites, is necessary for developing a vision-based excavator activity recognition model. Yet, complexities of dynamic working environments and security concerns at construction sites pose limitations on obtaining such videos from various surveillance camera locations. Consequently, this leads to performance degradation in excavator activity recognition models, reducing the accuracy and efficiency of heavy equipment productivity analysis. To address these limitations, this study aimed to conduct sensitivity analysis of excavator activity recognition performance based on surveillance camera location, utilizing synthetic videos generated from a game-engine-based virtual environment (Unreal Engine). Various scenarios for surveillance camera placement were devised, considering horizontal distance (20m, 30m, and 50m), vertical height (3m, 6m, and 10m), and horizontal angle (0° for front view, 90° for side view, and 180° for backside view). Performance analysis employed a 3D ResNet-18 model with transfer learning, yielding approximately 90.6% accuracy. Main findings revealed that horizontal distance significantly impacted model performance. Overall accuracy decreased with increasing distance (76.8% for 20m, 60.6% for 30m, and 35.3% for 50m). Particularly, videos with a 20m horizontal distance (close distance) exhibited accuracy above 80% in most scenarios. Moreover, accuracy trends in scenarios varied with vertical height and horizontal angle. At 0° (front view), accuracy mostly decreased with increasing height, while accuracy increased at 90° (side view) with increasing height. In addition, limited feature extraction for excavator activity recognition was found at 180° (backside view) due to occlusion of the excavator's bucket and arm. Based on these results, future studies should focus on enhancing the performance of vision-based recognition models by determining optimal surveillance camera locations at construction sites, utilizing deep learning algorithms for video super resolution, and establishing large training datasets using synthetic videos generated from game-engine-based virtual environments.

Adaptive Binarization for Camera-based Document Recognition (카메라 기반 문서 인식을 위한 적응적 이진화)

  • Kim, In-Jung
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.12 no.3
    • /
    • pp.132-140
    • /
    • 2007
  • The quality of the camera image is worse than that of the scanner image because of lighting variation and inaccurate focus. This paper proposes a binarization method for camera-based document recognition, which is tolerant to low-quality camera images. Based on an existing method reported to be effective in previous evaluations, we enhanced the adaptability to the image with a low contrast due to low intensity and inaccurate focus. Furthermore, applying an additional small-size window in the binarization process, it is effective to extract the fine detail of character structure, which is often degraded by conventional methods. In experiments, we applied the proposed method as well as other methods to a document recognizer and compared the performance for many cm images. The result showed the proposed method is effective for recognition of document images captured by the camera.

  • PDF

A Vehicle Recognition Method based on Radar and Camera Fusion in an Autonomous Driving Environment

  • Park, Mun-Yong;Lee, Suk-Ki;Shin, Dong-Jin
    • International journal of advanced smart convergence
    • /
    • v.10 no.4
    • /
    • pp.263-272
    • /
    • 2021
  • At a time when securing driving safety is the most important in the development and commercialization of autonomous vehicles, AI and big data-based algorithms are being studied to enhance and optimize the recognition and detection performance of various static and dynamic vehicles. However, there are many research cases to recognize it as the same vehicle by utilizing the unique advantages of radar and cameras, but they do not use deep learning image processing technology or detect only short distances as the same target due to radar performance problems. Radars can recognize vehicles without errors in situations such as night and fog, but it is not accurate even if the type of object is determined through RCS values, so accurate classification of the object through images such as cameras is required. Therefore, we propose a fusion-based vehicle recognition method that configures data sets that can be collected by radar device and camera device, calculates errors in the data sets, and recognizes them as the same target.

Optimal Camera Arrangement for Automatic Recognition of Steel Material based on Augmented Reality in Outdoor Environment (실외 환경에서의 증강 현실 기반의 자재 인식을 위한 최적의 카메라 배치)

  • Do, Hyun-Min;Kim, Bong-Keun
    • The Journal of Korea Robotics Society
    • /
    • v.5 no.2
    • /
    • pp.143-151
    • /
    • 2010
  • Automation and robotization has been required in construction for several decades and construction industry has become one of the important research areas in the field of service robotics. Especially in the steel construction, automatic recognition of structural steel members in the stockyard is emphasized. However, since the pose of steel frame in the stockyard is site dependent and also the stockyard is usually in the outdoor environment, it is difficult to determine the pose automatically. This paper adopts the recognition method based on the augmented reality to cope with this problem. Particularly focusing on the light condition of the outdoor environment, we formulated the optimization problem with the constraint and suggested the methodology to evaluate the optimal camera arrangement. From simulation results, sub-optimal solution for the position of the camera can be obtained.

Real time instruction classification system

  • Sang-Hoon Lee;Dong-Jin Kwon
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.16 no.3
    • /
    • pp.212-220
    • /
    • 2024
  • A recently the advancement of society, AI technology has made significant strides, especially in the fields of computer vision and voice recognition. This study introduces a system that leverages these technologies to recognize users through a camera and relay commands within a vehicle based on voice commands. The system uses the YOLO (You Only Look Once) machine learning algorithm, widely used for object and entity recognition, to identify specific users. For voice command recognition, a machine learning model based on spectrogram voice analysis is employed to identify specific commands. This design aims to enhance security and convenience by preventing unauthorized access to vehicles and IoT devices by anyone other than registered users. We converts camera input data into YOLO system inputs to determine if it is a person, Additionally, it collects voice data through a microphone embedded in the device or computer, converting it into time-domain spectrogram data to be used as input for the voice recognition machine learning system. The input camera image data and voice data undergo inference tasks through pre-trained models, enabling the recognition of simple commands within a limited space based on the inference results. This study demonstrates the feasibility of constructing a device management system within a confined space that enhances security and user convenience through a simple real-time system model. Finally our work aims to provide practical solutions in various application fields, such as smart homes and autonomous vehicles.

Performance Analysis of Surveillance Camera System Based on Image Recognition Server (화상 인식 서버 기반 감시 카메라 시스템의 성능 분석)

  • Shqairat, Yara;Lee, Goo Yeon
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.42 no.4
    • /
    • pp.816-818
    • /
    • 2017
  • In this paper, we perform a performance analysis of a surveillance camera network system with an image recognition server based on frame discard rate and server utilization. Surveillance camera states are divided into recognition states and silence states to analyze the various parameters such as the optimum number of image frames and the optimum number of cameras based on the processing capacity of the server. The analyzed results will be useful for efficient operation of the evolving surveillance camera network systems.

Camera-based Music Score Recognition Using Inverse Filter

  • Nguyen, Tam;Kim, SooHyung;Yang, HyungJeong;Lee, GueeSang
    • International Journal of Contents
    • /
    • v.10 no.4
    • /
    • pp.11-17
    • /
    • 2014
  • The influence of acquisition environment on music score images captured by a camera has not yet been seriously examined. All existing Optical Music Recognition (OMR) systems attempt to recognize music score images captured by a scanner under ideal conditions. Therefore, when such systems process images under the influence of distortion, different viewpoints or suboptimal illumination effects, the performance, in terms of recognition accuracy and processing time, is unacceptable for deployment in practice. In this paper, a novel, lightweight but effective approach for dealing with the issues caused by camera based music scores is proposed. Based on the staff line information, musical rules, run length code, and projection, all regions of interest are determined. Templates created from inverse filter are then used to recognize the music symbols. Therefore, all fragmentation and deformation problems, as well as missed recognition, can be overcome using the developed method. The system was evaluated on a dataset consisting of real images captured by a smartphone. The achieved recognition rate and processing time were relatively competitive with state of the art works. In addition, the system was designed to be lightweight compared with the other approaches, which mostly adopted machine learning algorithms, to allow further deployment on portable devices with limited computing resources.

Design and Implementation of Smart Self-Learning Aid: Micro Dot Pattern Recognition based Information Embedding Solution (스마트 학습지: 미세 격자 패턴 인식 기반의 지능형 학습 도우미 시스템의 설계와 구현)

  • Shim, Jae-Youen;Kim, Seong-Whan
    • Annual Conference of KIPS
    • /
    • 2011.04a
    • /
    • pp.346-349
    • /
    • 2011
  • In this paper, we design a perceptually invisible dot pattern layout and its recognition scheme, and we apply the recognition scheme into a smart self learning aid for interactive learning aid. To increase maximum information capacity and also increase robustness to the noises, we design a ECC (error correcting code) based dot pattern with directional vector indicator. To make a smart self-learning aid, we embed the micro dot pattern (20 information bit + 15 ECC bits + 9 layout information bit) using K ink (CMYK) and extract the dot pattern using IR (infrared) LED and IR filter based camera, which is embedded in the smart pen. The reason we use K ink is that K ink is a carbon based ink in nature, and carbon is easily recognized with IR even without light. After acquiring IR camera images for the dot patterns, we perform layout adjustment using the 9 layout information bit, and extract 20 information bits from 35 data bits which is composed of 20 information bits and 15 ECC bits. To embed and extract information bits, we use topology based dot pattern recognition scheme which is robust to geometric distortion which is very usual in camera based recognition scheme. Topology based pattern recognition traces next information bit symbols using topological distance measurement from the pivot information bit. We implemented and experimented with sample patterns, and it shows that we can achieve almost 99% recognition for our embedding patterns.

Performance Comparison of Template-based Face Recognition under Robotic Environments (로봇 환경의 템플릿 기반 얼굴인식 알고리즘 성능 비교)

  • Ban, Kyu-Dae;Kwak, Keun-Chang;Chi, Su-Young;Chung, Yun-Koo
    • The Journal of Korea Robotics Society
    • /
    • v.1 no.2
    • /
    • pp.151-157
    • /
    • 2006
  • This paper is concerned with the template-based face recognition from robot camera images with illumination and distance variations. The approaches used in this paper consist of Eigenface, Fisherface, and Icaface which are the most representative recognition techniques frequently used in conjunction with face recognition. These approaches are based on a popular unsupervised and supervised statistical technique that supports finding useful image representations, respectively. Thus we focus on the performance comparison from robot camera images with unwanted variations. The comprehensive experiments are completed for a databases with illumination and distance variations.

  • PDF