• Title/Summary/Keyword: Object recognition system

Search Result 714, Processing Time 0.027 seconds

Recognition of the Center Position of Bolt Hole in the Stand of Insulator Using Multilayer Neural Network (다층 뉴럴네트워크를 이용한 애자 스탠드에서의 볼트 구멍의 중심위치 인식)

  • 안경관;표성만
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.9 no.4
    • /
    • pp.304-309
    • /
    • 2003
  • Uninterrupted power supply has become indispensable during the maintenance task of active electric power lines as a result of today's highly information-oriented society and increasing demand of electric utilities. The maintenance task has the risk of electric shock and the danger of falling from high place. Therefore it is necessary to realize an autonomous robot system. In order to realize these tasks autonomously, the three dimensional position of target object such as electric line and the stand of insulator must be recognized accurately and rapidly. The approaching of an insulator and the wrenching of a nut task is selected as the typical task of the maintenance of active electric power distribution lines in this paper. Image recognition by multilayer neural network and optimal target position calculation method are newly proposed in order to recognize the center 3 dimensional position of the bolt hole in the stand of insulator. By the proposed image recognition method, it is proved that the center 3 dimensional position of the bolt hole can be recognized rapidly and accurately without regard to the pose of the stand of insulator. Finally the approaching and wrenching task is automatically realized using 6-link electro-hydraulic manipulators.

Video Analysis System for Action and Emotion Detection by Object with Hierarchical Clustering based Re-ID (계층적 군집화 기반 Re-ID를 활용한 객체별 행동 및 표정 검출용 영상 분석 시스템)

  • Lee, Sang-Hyun;Yang, Seong-Hun;Oh, Seung-Jin;Kang, Jinbeom
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.1
    • /
    • pp.89-106
    • /
    • 2022
  • Recently, the amount of video data collected from smartphones, CCTVs, black boxes, and high-definition cameras has increased rapidly. According to the increasing video data, the requirements for analysis and utilization are increasing. Due to the lack of skilled manpower to analyze videos in many industries, machine learning and artificial intelligence are actively used to assist manpower. In this situation, the demand for various computer vision technologies such as object detection and tracking, action detection, emotion detection, and Re-ID also increased rapidly. However, the object detection and tracking technology has many difficulties that degrade performance, such as re-appearance after the object's departure from the video recording location, and occlusion. Accordingly, action and emotion detection models based on object detection and tracking models also have difficulties in extracting data for each object. In addition, deep learning architectures consist of various models suffer from performance degradation due to bottlenects and lack of optimization. In this study, we propose an video analysis system consists of YOLOv5 based DeepSORT object tracking model, SlowFast based action recognition model, Torchreid based Re-ID model, and AWS Rekognition which is emotion recognition service. Proposed model uses single-linkage hierarchical clustering based Re-ID and some processing method which maximize hardware throughput. It has higher accuracy than the performance of the re-identification model using simple metrics, near real-time processing performance, and prevents tracking failure due to object departure and re-emergence, occlusion, etc. By continuously linking the action and facial emotion detection results of each object to the same object, it is possible to efficiently analyze videos. The re-identification model extracts a feature vector from the bounding box of object image detected by the object tracking model for each frame, and applies the single-linkage hierarchical clustering from the past frame using the extracted feature vectors to identify the same object that failed to track. Through the above process, it is possible to re-track the same object that has failed to tracking in the case of re-appearance or occlusion after leaving the video location. As a result, action and facial emotion detection results of the newly recognized object due to the tracking fails can be linked to those of the object that appeared in the past. On the other hand, as a way to improve processing performance, we introduce Bounding Box Queue by Object and Feature Queue method that can reduce RAM memory requirements while maximizing GPU memory throughput. Also we introduce the IoF(Intersection over Face) algorithm that allows facial emotion recognized through AWS Rekognition to be linked with object tracking information. The academic significance of this study is that the two-stage re-identification model can have real-time performance even in a high-cost environment that performs action and facial emotion detection according to processing techniques without reducing the accuracy by using simple metrics to achieve real-time performance. The practical implication of this study is that in various industrial fields that require action and facial emotion detection but have many difficulties due to the fails in object tracking can analyze videos effectively through proposed model. Proposed model which has high accuracy of retrace and processing performance can be used in various fields such as intelligent monitoring, observation services and behavioral or psychological analysis services where the integration of tracking information and extracted metadata creates greate industrial and business value. In the future, in order to measure the object tracking performance more precisely, there is a need to conduct an experiment using the MOT Challenge dataset, which is data used by many international conferences. We will investigate the problem that the IoF algorithm cannot solve to develop an additional complementary algorithm. In addition, we plan to conduct additional research to apply this model to various fields' dataset related to intelligent video analysis.

Depth Image Distortion Correction Method according to the Position and Angle of Depth Sensor and Its Hardware Implementation (거리 측정 센서의 위치와 각도에 따른 깊이 영상 왜곡 보정 방법 및 하드웨어 구현)

  • Jang, Kyounghoon;Cho, Hosang;Kim, Geun-Jun;Kang, Bongsoon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.18 no.5
    • /
    • pp.1103-1109
    • /
    • 2014
  • The motion recognition system has been broadly studied in digital image and video processing fields. Recently, method using th depth image is used very useful. However, recognition accuracy of depth image based method will be loss caused by size and shape of object distorted for angle of the depth sensor. Therefore, distortion correction of depth sensor is positively necessary for distinguished performance of the recognition system. In this paper, we propose a pre-processing algorithm to improve the motion recognition system. Depth data from depth sensor converted to real world, performed the corrected angle, and then inverse converted to projective world. The proposed system make progress using the OpenCV and the window program, and we test a system using the Kinect in real time. In addition, designed using Verilog-HDL and verified through the Zynq-7000 FPGA Board of Xilinx.

Development of a Multi-disciplinary Video Identification System for Autonomous Driving (자율주행을 위한 융복합 영상 식별 시스템 개발)

  • Sung-Youn Cho;Jeong-Joon Kim
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.24 no.1
    • /
    • pp.65-74
    • /
    • 2024
  • In recent years, image processing technology has played a critical role in the field of autonomous driving. Among them, image recognition technology is essential for the safety and performance of autonomous vehicles. Therefore, this paper aims to develop a hybrid image recognition system to enhance the safety and performance of autonomous vehicles. In this paper, various image recognition technologies are utilized to construct a system that recognizes and tracks objects in the vehicle's surroundings. Machine learning and deep learning algorithms are employed for this purpose, and objects are identified and classified in real-time through image processing and analysis. Furthermore, this study aims to fuse image processing technology with vehicle control systems to improve the safety and performance of autonomous vehicles. To achieve this, the identified object's information is transmitted to the vehicle control system to enable appropriate autonomous driving responses. The developed hybrid image recognition system in this paper is expected to significantly improve the safety and performance of autonomous vehicles. This is expected to accelerate the commercialization of autonomous vehicles.

Extraction of Workers and Heavy Equipment and Muliti-Object Tracking using Surveillance System in Construction Sites (건설 현장 CCTV 영상을 이용한 작업자와 중장비 추출 및 다중 객체 추적)

  • Cho, Young-Woon;Kang, Kyung-Su;Son, Bo-Sik;Ryu, Han-Guk
    • Journal of the Korea Institute of Building Construction
    • /
    • v.21 no.5
    • /
    • pp.397-408
    • /
    • 2021
  • The construction industry has the highest occupational accidents/injuries and has experienced the most fatalities among entire industries. Korean government installed surveillance camera systems at construction sites to reduce occupational accident rates. Construction safety managers are monitoring potential hazards at the sites through surveillance system; however, the human capability of monitoring surveillance system with their own eyes has critical issues. A long-time monitoring surveillance system causes high physical fatigue and has limitations in grasping all accidents in real-time. Therefore, this study aims to build a deep learning-based safety monitoring system that can obtain information on the recognition, location, identification of workers and heavy equipment in the construction sites by applying multiple object tracking with instance segmentation. To evaluate the system's performance, we utilized the Microsoft common objects in context and the multiple object tracking challenge metrics. These results prove that it is optimal for efficiently automating monitoring surveillance system task at construction sites.

Vehicle Manufacturer Recognition using Deep Learning and Perspective Transformation

  • Ansari, Israfil;Shim, Jaechang
    • Journal of Multimedia Information System
    • /
    • v.6 no.4
    • /
    • pp.235-238
    • /
    • 2019
  • In real world object detection is an active research topic for understanding different objects from images. There are different models presented in past and had significant results. In this paper we are presenting vehicle logo detection using previous object detection models such as You only look once (YOLO) and Faster Region-based CNN (F-RCNN). Both the front and rear view of the vehicles were used for training and testing the proposed method. Along with deep learning an image pre-processing algorithm called perspective transformation is proposed for all the test images. Using perspective transformation, the top view images were transformed into front view images. This algorithm has higher detection rate as compared to raw images. Furthermore, YOLO model has better result as compare to F-RCNN model.

A study of recognition system to the situation reaction (객체 정보에 대한 데이터베이스 구성 연구)

  • Park, Sangjoon;Lee, Jongchan
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2018.10a
    • /
    • pp.161-162
    • /
    • 2018
  • In this paper, we consider the development the database configuration to the search and management of the object information from GPS sensor and video sensor. Also, the design that the object trace of the video sensor to recognized object would be considered.

  • PDF

Estimation of surface reflectance properties and 3D shape recovery using photometric matching (물체의 면 반사특성 추정과 측광정합을 이용한 3차원 형상복구)

  • 김태은;류석현;송호근;최종수
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.21 no.7
    • /
    • pp.1633-1641
    • /
    • 1996
  • In this paper we propose a new method for anlayzing the properties of surface reflectance and reconstructing the shape of object using estimated reflectance parameters. We have investigated the hybrid reflectance surface which has specularreflection and diffuse reflection, which can be explained by Torrance-Sparrow model. Sample sphere made on one maerial is used to estimate the reflectance properties by using LMS algorithm. We can make the reference image which consists of surface normal and brightness value using estimated reflectance parameters, and thenarbitrary shape object made of the same material as sample can be reconstructed by matching with reference image. Photometric matching method proposed in this paper is robust because it mateches object image with the reference imageconsidering its neighbor brightness distribution. Also, in this paper plate diffuse illumination is used to remove intensity disparity with simple scheme. It is expected that the proposed algorithm can be applied to 3D recognition, vision inspection system and other fields.

  • PDF

A New Ergonomic Interface System for the Disabled Person (장애인을 위한 새로운 감성 인터페이스 연구)

  • Heo, Hwan;Lee, Ji-Woo;Lee, Won-Oh;Lee, Eui-Chul;Park, Kang-Ryoung
    • Journal of the Ergonomics Society of Korea
    • /
    • v.30 no.1
    • /
    • pp.229-235
    • /
    • 2011
  • Objective: Making a new ergonomic interface system based on camera vision system, which helps the handicapped in home environment. Background: Enabling the handicapped to manipulate the consumer electronics by the proposed interface system. Method: A wearable device for capturing the eye image using a near-infrared(NIR) camera and illuminators is proposed for tracking eye gaze position(Heo et al., 2011). A frontal viewing camera is attached to the wearable device, which can recognize the consumer electronics to be controlled(Heo et al., 2011). And the amount of user's eye fatigue can be measured based on eye blink rate, and in case that the user's fatigue exceeds in the predetermined level, the proposed system can automatically change the mode of gaze based interface into that of manual selection. Results: The experimental results showed that the gaze estimation error of the proposed method was 1.98 degrees with the successful recognition of the object by the frontal viewing camera(Heo et al., 2011). Conclusion: We made a new ergonomic interface system based on gaze tracking and object recognition Application: The proposed system can be used for helping the handicapped in home environment.

The Design and Implementation of School-Zone Safety Management System Based onContext-Aware (상황인식 기반의 스쿨존 안전 관리 시스템 설계 및 구현)

  • Lee, Jin-Kwan;Lee, Chang-Bok;Park, Sang-Jun;Lee, Jong-Chan;Park, Ki-Hong
    • Convergence Security Journal
    • /
    • v.9 no.1
    • /
    • pp.11-17
    • /
    • 2009
  • The object of this paper is to design a school-zone safety management system based on context-aware, integrated with computing technology. When it occurs to kidnap of elementary school students, the monitoring device creates context information through a combination object extraction and context-aware technology and alarm administrator about an emergency situation. In addition, the proposed system that requires a human perspective, a railroad crossing, statistics research of traffic, and a variety of applications such as factory automation systems can be used to be the best choice.

  • PDF