• Title/Summary/Keyword: Intelligent Video Analysis Method

Search Result 54, Processing Time 0.037 seconds

Object Tracking & PTZ camera Control for Intelligent Surveillance System (지능형 감시 시스템을 위한 객체 추적 및 PTZ 카메라 제어)

  • Lee, Young-Sik;Kim, Tae-Woo;Nam, Kee-Hwan;Park, Ho-Sik;Bae, Cheol-Soo
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.1 no.2
    • /
    • pp.65-70
    • /
    • 2008
  • Smart surveillance, is the use of automatic video analysis technologies in video surveillance applications. We present a robust object tracking method using pan-tilt-zoom camera for intelligent surveillance System, As the result of the experiment using 78 vehicle, the success rate of the tracking for moving object & non-moving object werw 97.4% and 91%. and 84.6%. the success rate o PTZ control for license plate image.

  • PDF

Low Resolution Rate Face Recognition Based on Multi-scale CNN

  • Wang, Ji-Yuan;Lee, Eung-Joo
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.12
    • /
    • pp.1467-1472
    • /
    • 2018
  • For the problem that the face image of surveillance video cannot be accurately identified due to the low resolution, this paper proposes a low resolution face recognition solution based on convolutional neural network model. Convolutional Neural Networks (CNN) model for multi-scale input The CNN model for multi-scale input is an improvement over the existing "two-step method" in which low-resolution images are up-sampled using a simple bi-cubic interpolation method. Then, the up sampled image and the high-resolution image are mixed as a model training sample. The CNN model learns the common feature space of the high- and low-resolution images, and then measures the feature similarity through the cosine distance. Finally, the recognition result is given. The experiments on the CMU PIE and Extended Yale B datasets show that the accuracy of the model is better than other comparison methods. Compared with the CMDA_BGE algorithm with the highest recognition rate, the accuracy rate is 2.5%~9.9%.

Specified Object Tracking Problem in an Environment of Multiple Moving Objects

  • Park, Seung-Min;Park, Jun-Heong;Kim, Hyung-Bok;Sim, Kwee-Bo
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.11 no.2
    • /
    • pp.118-123
    • /
    • 2011
  • Video based object tracking normally deals with non-stationary image streams that change over time. Robust and real time moving object tracking is considered to be a problematic issue in computer vision. Multiple object tracking has many practical applications in scene analysis for automated surveillance. In this paper, we introduce a specified object tracking based particle filter used in an environment of multiple moving objects. A differential image region based tracking method for the detection of multiple moving objects is used. In order to ensure accurate object detection in an unconstrained environment, a background image update method is used. In addition, there exist problems in tracking a particular object through a video sequence, which cannot rely only on image processing techniques. For this, a probabilistic framework is used. Our proposed particle filter has been proved to be robust in dealing with nonlinear and non-Gaussian problems. The particle filter provides a robust object tracking framework under ambiguity conditions and greatly improves the estimation accuracy for complicated tracking problems.

Detection of Abnormal Behavior by Scene Analysis in Surveillance Video (감시 영상에서의 장면 분석을 통한 이상행위 검출)

  • Bae, Gun-Tae;Uh, Young-Jung;Kwak, Soo-Yeong;Byun, Hye-Ran
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.36 no.12C
    • /
    • pp.744-752
    • /
    • 2011
  • In intelligent surveillance system, various methods for detecting abnormal behavior were proposed recently. However, most researches are not robust enough to be utilized for actual reality which often has occlusions because of assumption the researches have that individual objects can be tracked. This paper presents a novel method to detect abnormal behavior by analysing major motion of the scene for complex environment in which object tracking cannot work. First, we generate Visual Word and Visual Document from motion information extracted from input video and process them through LDA(Latent Dirichlet Allocation) algorithm which is one of document analysis technique to obtain major motion information(location, magnitude, direction, distribution) of the scene. Using acquired information, we compare similarity between motion appeared in input video and analysed major motion in order to detect motions which does not match to major motions as abnormal behavior.

A Deep Learning Algorithm for Fusing Action Recognition and Psychological Characteristics of Wrestlers

  • Yuan Yuan;Yuan Yuan;Jun Liu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.3
    • /
    • pp.754-774
    • /
    • 2023
  • Wrestling is one of the popular events for modern sports. It is difficult to quantitatively describe a wrestling game between athletes. And deep learning can help wrestling training by human recognition techniques. Based on the characteristics of latest wrestling competition rules and human recognition technologies, a set of wrestling competition video analysis and retrieval system is proposed. This system uses a combination of literature method, observation method, interview method and mathematical statistics to conduct statistics, analysis, research and discussion on the application of technology. Combined the system application in targeted movement technology. A deep learning-based facial recognition psychological feature analysis method for the training and competition of classical wrestling after the implementation of the new rules is proposed. The experimental results of this paper showed that the proportion of natural emotions of male and female wrestlers was about 50%, indicating that the wrestler's mentality was relatively stable before the intense physical confrontation, and the test of the system also proved the stability of the system.

Video Analysis System for Action and Emotion Detection by Object with Hierarchical Clustering based Re-ID (계층적 군집화 기반 Re-ID를 활용한 객체별 행동 및 표정 검출용 영상 분석 시스템)

  • Lee, Sang-Hyun;Yang, Seong-Hun;Oh, Seung-Jin;Kang, Jinbeom
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.1
    • /
    • pp.89-106
    • /
    • 2022
  • Recently, the amount of video data collected from smartphones, CCTVs, black boxes, and high-definition cameras has increased rapidly. According to the increasing video data, the requirements for analysis and utilization are increasing. Due to the lack of skilled manpower to analyze videos in many industries, machine learning and artificial intelligence are actively used to assist manpower. In this situation, the demand for various computer vision technologies such as object detection and tracking, action detection, emotion detection, and Re-ID also increased rapidly. However, the object detection and tracking technology has many difficulties that degrade performance, such as re-appearance after the object's departure from the video recording location, and occlusion. Accordingly, action and emotion detection models based on object detection and tracking models also have difficulties in extracting data for each object. In addition, deep learning architectures consist of various models suffer from performance degradation due to bottlenects and lack of optimization. In this study, we propose an video analysis system consists of YOLOv5 based DeepSORT object tracking model, SlowFast based action recognition model, Torchreid based Re-ID model, and AWS Rekognition which is emotion recognition service. Proposed model uses single-linkage hierarchical clustering based Re-ID and some processing method which maximize hardware throughput. It has higher accuracy than the performance of the re-identification model using simple metrics, near real-time processing performance, and prevents tracking failure due to object departure and re-emergence, occlusion, etc. By continuously linking the action and facial emotion detection results of each object to the same object, it is possible to efficiently analyze videos. The re-identification model extracts a feature vector from the bounding box of object image detected by the object tracking model for each frame, and applies the single-linkage hierarchical clustering from the past frame using the extracted feature vectors to identify the same object that failed to track. Through the above process, it is possible to re-track the same object that has failed to tracking in the case of re-appearance or occlusion after leaving the video location. As a result, action and facial emotion detection results of the newly recognized object due to the tracking fails can be linked to those of the object that appeared in the past. On the other hand, as a way to improve processing performance, we introduce Bounding Box Queue by Object and Feature Queue method that can reduce RAM memory requirements while maximizing GPU memory throughput. Also we introduce the IoF(Intersection over Face) algorithm that allows facial emotion recognized through AWS Rekognition to be linked with object tracking information. The academic significance of this study is that the two-stage re-identification model can have real-time performance even in a high-cost environment that performs action and facial emotion detection according to processing techniques without reducing the accuracy by using simple metrics to achieve real-time performance. The practical implication of this study is that in various industrial fields that require action and facial emotion detection but have many difficulties due to the fails in object tracking can analyze videos effectively through proposed model. Proposed model which has high accuracy of retrace and processing performance can be used in various fields such as intelligent monitoring, observation services and behavioral or psychological analysis services where the integration of tracking information and extracted metadata creates greate industrial and business value. In the future, in order to measure the object tracking performance more precisely, there is a need to conduct an experiment using the MOT Challenge dataset, which is data used by many international conferences. We will investigate the problem that the IoF algorithm cannot solve to develop an additional complementary algorithm. In addition, we plan to conduct additional research to apply this model to various fields' dataset related to intelligent video analysis.

Knowledge-based Video Retrieval System Using Korean Closed-caption (한국어 폐쇄자막을 이용한 지식기반 비디오 검색 시스템)

  • 조정원;정승도;최병욱
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.41 no.3
    • /
    • pp.115-124
    • /
    • 2004
  • The content-based retrieval using low-level features can hardly provide the retrieval result that corresponds with conceptual demand of user for intelligent retrieval. Video includes not only moving picture data, but also audio or closed-caption data. Knowledge-based video retrieval is able to provide the retrieval result that corresponds with conceptual demand of user because of performing automatic indexing with such a variety data. In this paper, we present the knowledge-based video retrieval system using Korean closed-caption. The closed-caption is indexed by Korean keyword extraction system including the morphological analysis process. As a result, we are able to retrieve the video by using keyword from the indexing database. In the experiment, we have applied the proposed method to news video with closed-caption generated by Korean stenographic system, and have empirically confirmed that the proposed method provides the retrieval result that corresponds with more meaningful conceptual demand of user.

The Development of Vehicle Counting System at Intersection Using Mean Shift (Mean Shift를 이용한 교차로 교통량 측정 시스템 개발)

  • Chun, In-Gook
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.7 no.3
    • /
    • pp.38-47
    • /
    • 2008
  • A vehicle counting system at intersection is designed and implemented using analyzing a video stream from a camera. To separate foreground image from background, we compare three different methods, among which Li's method is chosen. Blobs are extracted from the foreground image using connected component analysis and the blobs are tracked by a blob tracker, frame by frame. The primary tracker use only the size and location of blob in foreground image. If there is a collision between blobs, the mean-shift tracking algorithm based on color distribution of blob is used. The proposed system is tested using real video data at intersection. If some huristics is applied, the system shows a good detection rate and a low error rate.

  • PDF

Unusual Motion Detection for Vision-Based Driver Assistance

  • Fu, Li-Hua;Wu, Wei-Dong;Zhang, Yu;Klette, Reinhard
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.15 no.1
    • /
    • pp.27-34
    • /
    • 2015
  • For a vision-based driver assistance system, unusual motion detection is one of the important means of preventing accidents. In this paper, we propose a real-time unusual-motion-detection model, which contains two stages: salient region detection and unusual motion detection. In the salient-region-detection stage, we present an improved temporal attention model. In the unusual-motion-detection stage, three kinds of factors, the speed, the motion direction, and the distance, are extracted for detecting unusual motion. A series of experimental results demonstrates the proposed method and shows the feasibility of the proposed model.

Quantitative Analysis Method for Encrypted Video (암호화된 동영상 비식별화율의 정량적 분석 방법)

  • Deok-Han Kim;Young-Gab Kim
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.11a
    • /
    • pp.193-195
    • /
    • 2023
  • 최근 다양한 동영상 미디어 콘텐츠가 늘어나면서 민감한 개인정보 유출의 위험성 또한 높아졌다. 이에 따라 다양한 동영상 비식별화 기법이 연구되었고, 그중에서 동영상 암호화 기술은 별도로 원본 동영상을 보관하지 않아도 복호화를 통해 원본 동영상을 얻을 수 있다는 장점 때문에 계속해서 연구가 진행되고 있다. 많은 동영상 암호화 연구에서는 암호화된 동영상의 비식별성을 입증하기 위해 기존의 이미지 암호화 연구에서 사용되던 평가 지표를 사용한다. 그러나 이러한 지표들은 암호화된 동영상의 비식별성을 입증하기에는 적합하지 않다. 따라서 본 논문에서는 암호화된 동영상이 전체 구간에서 비식별화되었는지 확인하는 방법을 제안한다. 본 논문에서는 기존의 지표들을 가중 합산하여 동영상의 모든 프레임에 대해 측정하고 이를 그래프로 표현하여 분석한다. 이 방법을 통해 암호화된 동영상에서 비식별화가 정상적으로 적용되지 않은 부분을 쉽게 파악할 수 있다.