• Title/Summary/Keyword: Video Data Classification

Search Result 133, Processing Time 0.025 seconds

Video Data Classification based on a Video Feature Profile (특성정보 프로파일에 기반한 동영상 데이터 분류)

  • Son Jeong-Sik;Chang Joong-Hyuk;Lee Won-Suk
    • The KIPS Transactions:PartD
    • /
    • v.12D no.1 s.97
    • /
    • pp.31-42
    • /
    • 2005
  • Generally, conventional video searching or classification methods are based on its meta-data. However, it is almost Impossible to represent the precise information of a video data by its meta-data. Therefore, a processing method of video data that is based on its meta-data has a limitation to be efficiently applied in application fields. In this paper, for efficient classification of video data, a classification method of video data that is based on its low-level data is proposed. The proposed method extracts the characteristics of video data from the given video data by clustering process, and makes the profile of the video data. Subsequently. the similarity between the profile and video data to be classified is computed by a comparing process of the profile and the video data. Based on the similarity. the video data is classified properly. Furthermore, in order to improve the performance of the comparing process, generating and comparing techniques of integrated profile are presented. A comparing technique based on a differentiated weight to improve a result of a comparing Process Is also Presented. Finally, the performance of the proposed method is verified through a series of experiments using various video data.

Using the fusion of spatial and temporal features for malicious video classification (공간과 시간적 특징 융합 기반 유해 비디오 분류에 관한 연구)

  • Jeon, Jae-Hyun;Kim, Se-Min;Han, Seung-Wan;Ro, Yong-Man
    • The KIPS Transactions:PartB
    • /
    • v.18B no.6
    • /
    • pp.365-374
    • /
    • 2011
  • Recently, malicious video classification and filtering techniques are of practical interest as ones can easily access to malicious multimedia contents through the Internet, IPTV, online social network, and etc. Considerable research efforts have been made to developing malicious video classification and filtering systems. However, the malicious video classification and filtering is not still being from mature in terms of reliable classification/filtering performance. In particular, the most of conventional approaches have been limited to using only the spatial features (such as a ratio of skin regions and bag of visual words) for the purpose of malicious image classification. Hence, previous approaches have been restricted to achieving acceptable classification and filtering performance. In order to overcome the aforementioned limitation, we propose new malicious video classification framework that takes advantage of using both the spatial and temporal features that are readily extracted from a sequence of video frames. In particular, we develop the effective temporal features based on the motion periodicity feature and temporal correlation. In addition, to exploit the best data fusion approach aiming to combine the spatial and temporal features, the representative data fusion approaches are applied to the proposed framework. To demonstrate the effectiveness of our method, we collect 200 sexual intercourse videos and 200 non-sexual intercourse videos. Experimental results show that the proposed method increases 3.75% (from 92.25% to 96%) for classification of sexual intercourse video in terms of accuracy. Further, based on our experimental results, feature-level fusion approach (for fusing spatial and temporal features) is found to achieve the best classification accuracy.

Video Classification System Based on Similarity Representation Among Sequential Data (순차 데이터간의 유사도 표현에 의한 동영상 분류)

  • Lee, Hosuk;Yang, Jihoon
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.7 no.1
    • /
    • pp.1-8
    • /
    • 2018
  • It is not easy to learn simple expressions of moving picture data since it contains noise and a lot of information in addition to time-based information. In this study, we propose a similarity representation method and a deep learning method between sequential data which can express such video data abstractly and simpler. This is to learn and obtain a function that allow them to have maximum information when interpreting the degree of similarity between image data vectors constituting a moving picture. Through the actual data, it is confirmed that the proposed method shows better classification performance than the existing moving image classification methods.

Classification of Phornographic Videos Based on the Audio Information (오디오 신호에 기반한 음란 동영상 판별)

  • Kim, Bong-Wan;Choi, Dae-Lim;Lee, Yong-Ju
    • MALSORI
    • /
    • no.63
    • /
    • pp.139-151
    • /
    • 2007
  • As the Internet becomes prevalent in our lives, harmful contents, such as phornographic videos, have been increasing on the Internet, which has become a very serious problem. To prevent such an event, there are many filtering systems mainly based on the keyword-or image-based methods. The main purpose of this paper is to devise a system that classifies pornographic videos based on the audio information. We use the mel-cepstrum modulation energy (MCME) which is a modulation energy calculated on the time trajectory of the mel-frequency cepstral coefficients (MFCC) as well as the MFCC as the feature vector. For the classifier, we use the well-known Gaussian mixture model (GMM). The experimental results showed that the proposed system effectively classified 98.3% of pornographic data and 99.8% of non-pornographic data. We expect the proposed method can be applied to the more accurate classification system which uses both video and audio information.

  • PDF

Automatic Video Genre Classification Method in MPEG compressed domain (MPEG 부호화 영역에서 Video Genre 자동 분류 방법)

  • Kim, Tae-Hee;Lee, Woong-Hee;Jeong, Dong-Seok
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.27 no.8A
    • /
    • pp.836-845
    • /
    • 2002
  • Video summary is one of the tools which can provide the fast and effective browsing for a lengthy video. Video summary consists of many key-frames that could be defined differently depending on the video genre it belongs to. Consequently, the video summary constructed by the uniform manner might lead into inadequate result. Therefore, identifying the video genre is the important first step in generating the meaningful video summary. We propose a new method that can classify the genre of the video data in MPEC compressed bit-stream domain. Since the proposed method operates directly on the compressed bit-stream without decoding the frame, it has merits such as simple calculation and short processing time. In the proposed method, only the visual information is utilized through the spatial-temporal analysis to classify the video genre. Experiments are done for 6 genres of video: Cartoon, commercial, Music Video, News, Sports, and Talk Show. Experimental result shows more than 90% of accuracy in genre classification for the well -structured video data such as Talk Show and Sports.

Modeling and Classification of MPEG VBR Video Data using Gradient-based Fuzzy c_means with Divergence Measure (분산 기반의 Gradient Based Fuzzy c-means 에 의한 MPEG VBR 비디오 데이터의 모델링과 분류)

  • 박동철;김봉주
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.7C
    • /
    • pp.931-936
    • /
    • 2004
  • GBFCM(DM), Gradient-based Fuzzy c-means with Divergence Measure, for efficient clustering of GPDF(Gaussian Probability Density Function) in MPEG VBR video data modeling is proposed in this paper. The proposed GBFCM(DM) is based on GBFCM( Gradient-based Fuzzy c-means) with the Divergence for its distance measure. In this paper, sets of real-time MPEG VBR Video traffic data are considered. Each of 12 frames MPEG VBR Video data are first transformed to 12-dimensional data for modeling and the transformed 12-dimensional data are Pass through the proposed GBFCM(DM) for classification. The GBFCM(DM) is compared with conventional FCM and GBFCM algorithms. The results show that the GBFCM(DM) gives 5∼15% improvement in False Alarm Rate over conventional algorithms such as FCM and GBFCM.

Video augmentation technique for human action recognition using genetic algorithm

  • Nida, Nudrat;Yousaf, Muhammad Haroon;Irtaza, Aun;Velastin, Sergio A.
    • ETRI Journal
    • /
    • v.44 no.2
    • /
    • pp.327-338
    • /
    • 2022
  • Classification models for human action recognition require robust features and large training sets for good generalization. However, data augmentation methods are employed for imbalanced training sets to achieve higher accuracy. These samples generated using data augmentation only reflect existing samples within the training set, their feature representations are less diverse and hence, contribute to less precise classification. This paper presents new data augmentation and action representation approaches to grow training sets. The proposed approach is based on two fundamental concepts: virtual video generation for augmentation and representation of the action videos through robust features. Virtual videos are generated from the motion history templates of action videos, which are convolved using a convolutional neural network, to generate deep features. Furthermore, by observing an objective function of the genetic algorithm, the spatiotemporal features of different samples are combined, to generate the representations of the virtual videos and then classified through an extreme learning machine classifier on MuHAVi-Uncut, iXMAS, and IAVID-1 datasets.

Implementation of Smart Video Surveillance System Based on Safety Map (안전지도와 연계한 지능형 영상보안 시스템 구현)

  • Park, Jang-Sik
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.13 no.1
    • /
    • pp.169-174
    • /
    • 2018
  • There are many CCTV cameras connected to the video surveillance and monitoring center for the safety of citizens, and it is difficult for a few monitoring agents to monitor many channels of videos. In this paper, we propose an intelligent video surveillance system utilizing a safety map to efficiently monitor many channels of CCTV camera videos. The safety map establishes the frequency of crime occurrence as a database, expresses the degree of crime risk and makes it possible for agents of the video surveillance center to pay attention when a woman enters the crime risk area. The proposed gender classification method is processed in the order of pedestrian detection, tracking and classification with deep training. The pedestrian detection and tracking uses Adaboost algorithm and probabilistic data association filter, respectively. In order to classify the gender of the pedestrian, relatively simple AlexNet is applied to determine gender. Experimental results show that the proposed gender classification method is more effective than the conventional algorithm. In addition, the results of implementation of intelligent video security system combined with safety map are introduced.

Image Mood Classification Using Deep CNN and Its Application to Automatic Video Generation (심층 CNN을 활용한 영상 분위기 분류 및 이를 활용한 동영상 자동 생성)

  • Cho, Dong-Hee;Nam, Yong-Wook;Lee, Hyun-Chang;Kim, Yong-Hyuk
    • Journal of the Korea Convergence Society
    • /
    • v.10 no.9
    • /
    • pp.23-29
    • /
    • 2019
  • In this paper, the mood of images was classified into eight categories through a deep convolutional neural network and video was automatically generated using proper background music. Based on the collected image data, the classification model is learned using a multilayer perceptron (MLP). Using the MLP, a video is generated by using multi-class classification to predict image mood to be used for video generation, and by matching pre-classified music. As a result of 10-fold cross-validation and result of experiments on actual images, each 72.4% of accuracy and 64% of confusion matrix accuracy was achieved. In the case of misclassification, by classifying video into a similar mood, it was confirmed that the music from the video had no great mismatch with images.

CCTV Based Gender Classification Using a Convolutional Neural Networks (컨볼루션 신경망을 이용한 CCTV 영상 기반의 성별구분)

  • Kang, Hyun Gon;Park, Jang Sik;Song, Jong Kwan;Yoon, Byung Woo
    • Journal of Korea Multimedia Society
    • /
    • v.19 no.12
    • /
    • pp.1943-1950
    • /
    • 2016
  • Recently, gender classification has attracted a great deal of attention in the field of video surveillance system. It can be useful in many applications such as detecting crimes for women and business intelligence. In this paper, we proposed a method which can detect pedestrians from CCTV video and classify the gender of the detected objects. So far, many algorithms have been proposed to classify people according the their gender. This paper presents a gender classification using convolutional neural network. The detection phase is performed by AdaBoost algorithm based on Haar-like features and LBP features. Classifier and detector is trained with data-sets generated form CCTV images. The experimental results of the proposed method is male matching rate of 89.9% and the results shows 90.7% of female videos. As results of simulations, it is shown that the proposed gender classification is better than conventional classification algorithm.