• Title/Summary/Keyword: Real-time Object Classification

Search Result 74, Processing Time 0.023 seconds

An Optimal Implementation of Object Tracking Algorithm for DaVinci Processor-based Smart Camera (다빈치 프로세서 기반 스마트 카메라에서의 객체 추적 알고리즘의 최적 구현)

  • Lee, Byung-Eun;Nguyen, Thanh Binh;Chung, Sun-Tae
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2009.05a
    • /
    • pp.17-22
    • /
    • 2009
  • DaVinci processors are popular media processors for implementing embedded multimedia applications. They support dual core architecture: ARM9 core for video I/O handling as well as system management and peripheral handling, and DSP C64+ core for effective digital signal processing. In this paper, we propose our efforts for optimal implementation of object tracking algorithm in DaVinci-based smart camera which is being designed and implemented by our laboratory. The smart camera in this paper is supposed to support object detection, object tracking, object classification and detection of intrusion into surveillance regions and sending the detection event to remote clients using IP protocol. Object tracking algorithm is computationally expensive since it needs to process several procedures such as foreground mask extraction, foreground mask correction, connected component labeling, blob region calculation, object prediction, and etc. which require large amount of computation times. Thus, if it is not implemented optimally in Davinci-based processors, one cannot expect real-time performance of the smart camera.

  • PDF

Development of Automated Surface Inspection System using the Computer V (컴퓨터 비젼을 이용한 표면결함검사장치 개발)

  • Lee, Jong-Hak;Jung, Jin-Yang
    • Proceedings of the KIEE Conference
    • /
    • 1999.07b
    • /
    • pp.668-670
    • /
    • 1999
  • We have developed a automatic surface inspection system for cold Rolled strips in steel making process for several years. We have experienced the various kinds of surface inspection systems, including linear CCD camera type and the laser type inspection system which was installed in cold rolled strips production lines. But, we did not satisfied with these inspection systems owing to insufficient detection and classification rate, real time processing performance and limited line speed of real production lines. In order to increase detection and computing power, we have used the Dark Field illumination with Infra_Red LED, Bright Field illumination with Xenon Lamp, Parallel Computing Processor with Area typed CCD camera and full software based image processing technique for the ease up_grading and maintenance. In this paper, we introduced the automatic inspection system and real time image processing technique using the Object Detection, Defect Detection, Classification algorithms. As a result of experiment, under the situation of the high speed processed line(max 1000 meter per minute) defect detection is above 90% for all occurred defects in real line, defect name classification rate is about 80% for most frequently occurred 8 defect, and defect grade classification rate is 84% for name classified defect.

  • PDF

Real-Time Object Recognition for Children Education Applications based on Augmented Reality (증강현실 기반 아동 학습 어플리케이션을 위한 실시간 영상 인식)

  • Park, Kang-Kyu;Yi, Kang
    • Journal of Korea Multimedia Society
    • /
    • v.20 no.1
    • /
    • pp.17-31
    • /
    • 2017
  • The aim of the paper is to present an object recognition method toward augmented reality system that utilizes existing education instruments that was designed without any consideration on image processing and recognition. The light reflection, sizes, shapes, and color range of the existing target education instruments are major hurdles to our object recognition. In addition, the real-time performance requirements on embedded devices and user experience constraints for children users are quite challenging issues to be solved for our image processing and object recognition approach. In order to meet these requirements we employed a method cascading light-weight weak classification methods that are complimentary each other to make a resultant complicated and highly accurate object classifier toward practically reasonable precision ratio. We implemented the proposed method and tested the performance by video with more than 11,700 frames of actual playing scenario. The experimental result showed 0.54% miss ratio and 1.35% false hit ratio.

Sparse Feature Convolutional Neural Network with Cluster Max Extraction for Fast Object Classification

  • Kim, Sung Hee;Pae, Dong Sung;Kang, Tae-Koo;Kim, Dong W.;Lim, Myo Taeg
    • Journal of Electrical Engineering and Technology
    • /
    • v.13 no.6
    • /
    • pp.2468-2478
    • /
    • 2018
  • We propose the Sparse Feature Convolutional Neural Network (SFCNN) to reduce the volume of convolutional neural networks (CNNs). Despite the superior classification performance of CNNs, their enormous network volume requires high computational cost and long processing time, making real-time applications such as online-training difficult. We propose an advanced network that reduces the volume of conventional CNNs by producing a region-based sparse feature map. To produce the sparse feature map, two complementary region-based value extraction methods, cluster max extraction and local value extraction, are proposed. Cluster max is selected as the main function based on experimental results. To evaluate SFCNN, we conduct an experiment with two conventional CNNs. The network trains 59 times faster and tests 81 times faster than the VGG network, with a 1.2% loss of accuracy in multi-class classification using the Caltech101 dataset. In vehicle classification using the GTI Vehicle Image Database, the network trains 88 times faster and tests 94 times faster than the conventional CNNs, with a 0.1% loss of accuracy.

A New Object Region Detection and Classification Method using Multiple Sensors on the Driving Environment (다중 센서를 사용한 주행 환경에서의 객체 검출 및 분류 방법)

  • Kim, Jung-Un;Kang, Hang-Bong
    • Journal of Korea Multimedia Society
    • /
    • v.20 no.8
    • /
    • pp.1271-1281
    • /
    • 2017
  • It is essential to collect and analyze target information around the vehicle for autonomous driving of the vehicle. Based on the analysis, environmental information such as location and direction should be analyzed in real time to control the vehicle. In particular, obstruction or cutting of objects in the image must be handled to provide accurate information about the vehicle environment and to facilitate safe operation. In this paper, we propose a method to simultaneously generate 2D and 3D bounding box proposals using LiDAR Edge generated by filtering LiDAR sensor information. We classify the classes of each proposal by connecting them with Region-based Fully-Covolutional Networks (R-FCN), which is an object classifier based on Deep Learning, which uses two-dimensional images as inputs. Each 3D box is rearranged by using the class label and the subcategory information of each class to finally complete the 3D bounding box corresponding to the object. Because 3D bounding boxes are created in 3D space, object information such as space coordinates and object size can be obtained at once, and 2D bounding boxes associated with 3D boxes do not have problems such as occlusion.

Classification of Objects using CNN-Based Vision and Lidar Fusion in Autonomous Vehicle Environment

  • G.komali ;A.Sri Nagesh
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.11
    • /
    • pp.67-72
    • /
    • 2023
  • In the past decade, Autonomous Vehicle Systems (AVS) have advanced at an exponential rate, particularly due to improvements in artificial intelligence, which have had a significant impact on social as well as road safety and the future of transportation systems. The fusion of light detection and ranging (LiDAR) and camera data in real-time is known to be a crucial process in many applications, such as in autonomous driving, industrial automation and robotics. Especially in the case of autonomous vehicles, the efficient fusion of data from these two types of sensors is important to enabling the depth of objects as well as the classification of objects at short and long distances. This paper presents classification of objects using CNN based vision and Light Detection and Ranging (LIDAR) fusion in autonomous vehicles in the environment. This method is based on convolutional neural network (CNN) and image up sampling theory. By creating a point cloud of LIDAR data up sampling and converting into pixel-level depth information, depth information is connected with Red Green Blue data and fed into a deep CNN. The proposed method can obtain informative feature representation for object classification in autonomous vehicle environment using the integrated vision and LIDAR data. This method is adopted to guarantee both object classification accuracy and minimal loss. Experimental results show the effectiveness and efficiency of presented approach for objects classification.

Forest Fire Detection System using Drone Streaming Images (드론 스트리밍 영상 이미지 분석을 통한 실시간 산불 탐지 시스템)

  • Yoosin Kim
    • Journal of Advanced Navigation Technology
    • /
    • v.27 no.5
    • /
    • pp.685-689
    • /
    • 2023
  • The proposed system in the study aims to detect forest fires in real-time stream data received from the drone-camera. Recently, the number of wildfires has been increasing, and also the large scaled wildfires are frequent more and more. In order to prevent forest fire damage, many experiments using the drone camera and vision analysis are actively conducted, however there were many challenges, such as network speed, pre-processing, and model performance, to detect forest fires from real-time streaming data of the flying drone. Therefore, this study applied image data processing works to capture five good image frames for vision analysis from whole streaming data and then developed the object detection model based on YOLO_v2. As the result, the classification model performance of forest fire images reached upto 93% of accuracy, and the field test for the model verification detected the forest fire with about 70% accuracy.

Real-Time Surveillance of People on an Embedded DSP-Platform

  • Qiao, Qifeng;Peng, Yu;Zhang, Dali
    • Journal of Ubiquitous Convergence Technology
    • /
    • v.1 no.1
    • /
    • pp.3-8
    • /
    • 2007
  • This paper presents a set of techniques used in a real-time visual surveillance system. The system is implemented on a low-cost embedded DSP platform that is designed to work with stationary video sources. It consists of detection, a tracking and a classification module. The detector uses a statistical method to establish the background model and extract the foreground pixels. These pixels are grouped into blobs which are classified into single person, people in a group and other objects by the dynamic periodicity analysis. The tracking module uses mean shift algorithm to locate the target position. The system aims to control the human density in the surveilled scene and detect what happens abnormally. The major advantage of this system is the real-time capability and it only requires a video stream without other additional sensors. We evaluate the system in the real application, for example monitoring the subway entrance and the building hall, and the results prove the system's superior performance.

  • PDF

(Real Time Classification System for Lead Pin Images) (실시간 Lead Pin 영상 분류 시스템)

  • 장용훈
    • Journal of the Korea Computer Industry Society
    • /
    • v.3 no.9
    • /
    • pp.1177-1188
    • /
    • 2002
  • To classify real time Lead pin images in this paper, The image acquisition system was composed to C.C.D, image frame grabber(DT3153), P.C(PentiumIII). I proposed image processing algorithms. This algorithms were composed to real time monitoring, Lead Pin image acquisition, image noise deletion, object area detection, point detection and pattern classification algorithm. The raw images were acquired from Lead pin images using the system. The result images were obtained from raw images by image processing algorithms. In implemental result, The right recognition was 97 of 100 acceptable products, 95 of 100 defective products. The recognition rate was 96% for total 200 Lead Pins.

  • PDF

Design and Implementation of a Real-Time Face Detection System (실시간 얼굴 검출 시스템 설계 및 구현)

  • Jung Sung-Tae;Lee Ho-Geun
    • Journal of Korea Multimedia Society
    • /
    • v.8 no.8
    • /
    • pp.1057-1068
    • /
    • 2005
  • This paper proposes a real-time face detection system which detects multiple faces from low resolution video such as web-camera video. First, It finds face region candidates by using AdaBoost based object detection method which selects a small number of critical features from a larger set. Next, it generates reduced feature vector for each face region candidate by using principle component analysis. Finally, it classifies if the candidate is a face or non-face by using SVM(Support Vector Machine) based binary classification. According to experiment results, the proposed method achieves real-time face detection from low resolution video. Also, it reduces the false detection rate than existing methods by using PCA and SVM based face classification step.

  • PDF