• Title/Summary/Keyword: Video Data Classification

Search Result 136, Processing Time 0.024 seconds

Classification and Recommendation of Scene Templates for PR Video Making Service based on Strategic Meta Information (홍보동영상 제작 서비스를 위한 전략메타정보 기반 장면템플릿 분류 및 추천)

  • Park, Jongbin;Lee, Han-Duck;Kim, Kyung-Won;Jung, Jong-Jin;Lim, Tae-Beom
    • Journal of Broadcast Engineering
    • /
    • v.20 no.6
    • /
    • pp.848-861
    • /
    • 2015
  • In this paper, we introduce a new web-based PR video making service system. Many video editing tools have required tough editing skill or scenario planning stage for a just simple PR video making. Some users may prefer a simple and fast way than sophisticated and complex functionality. To solve this problem, it is important to provide easy user interface and intelligent classification and recommendation scheme. Therefore, we propose a new template classification and recommendation scheme using a topic modeling method. The proposed scheme has the big advantage of being able to handle the unstructured meta data as well as structured one.

Ontology and Sequential Rule Based Streaming Media Event Recognition (온톨로지 및 순서 규칙 기반 대용량 스트리밍 미디어 이벤트 인지)

  • Soh, Chi-Seung;Park, Hyun-Kyu;Park, Young-Tack
    • Journal of KIISE
    • /
    • v.43 no.4
    • /
    • pp.470-479
    • /
    • 2016
  • As the number of various types of media data such as UCC (User Created Contents) increases, research is actively being carried out in many different fields so as to provide meaningful media services. Amidst these studies, a semantic web-based media classification approach has been proposed; however, it encounters some limitations in video classification because of its underlying ontology derived from meta-information such as video tag and title. In this paper, we define recognized objects in a video and activity that is composed of video objects in a shot, and introduce a reasoning approach based on description logic. We define sequential rules for a sequence of shots in a video and describe how to classify it. For processing the large amount of increasing media data, we utilize Spark streaming, and a distributed in-memory big data processing framework, and describe how to classify media data in parallel. To evaluate the efficiency of the proposed approach, we conducted an experiment using a large amount of media ontology extracted from Youtube videos.

Two person Interaction Recognition Based on Effective Hybrid Learning

  • Ahmed, Minhaz Uddin;Kim, Yeong Hyeon;Kim, Jin Woo;Bashar, Md Rezaul;Rhee, Phill Kyu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.2
    • /
    • pp.751-770
    • /
    • 2019
  • Action recognition is an essential task in computer vision due to the variety of prospective applications, such as security surveillance, machine learning, and human-computer interaction. The availability of more video data than ever before and the lofty performance of deep convolutional neural networks also make it essential for action recognition in video. Unfortunately, limited crafted video features and the scarcity of benchmark datasets make it challenging to address the multi-person action recognition task in video data. In this work, we propose a deep convolutional neural network-based Effective Hybrid Learning (EHL) framework for two-person interaction classification in video data. Our approach exploits a pre-trained network model (the VGG16 from the University of Oxford Visual Geometry Group) and extends the Faster R-CNN (region-based convolutional neural network a state-of-the-art detector for image classification). We broaden a semi-supervised learning method combined with an active learning method to improve overall performance. Numerous types of two-person interactions exist in the real world, which makes this a challenging task. In our experiment, we consider a limited number of actions, such as hugging, fighting, linking arms, talking, and kidnapping in two environment such simple and complex. We show that our trained model with an active semi-supervised learning architecture gradually improves the performance. In a simple environment using an Intelligent Technology Laboratory (ITLab) dataset from Inha University, performance increased to 95.6% accuracy, and in a complex environment, performance reached 81% accuracy. Our method reduces data-labeling time, compared to supervised learning methods, for the ITLab dataset. We also conduct extensive experiment on Human Action Recognition benchmarks such as UT-Interaction dataset, HMDB51 dataset and obtain better performance than state-of-the-art approaches.

User Perception of Olfactory Information for Video Reality and Video Classification (영상실감을 위한 후각정보에 대한 사용자 지각과 영상분류)

  • Lee, Guk-Hee;Li, Hyung-Chul O.;Ahn, Chung Hyun;Choi, Ji Hoon;Kim, Shin Woo
    • Journal of the HCI Society of Korea
    • /
    • v.8 no.2
    • /
    • pp.9-19
    • /
    • 2013
  • There has been much advancement in reality enhancement using audio-visual information. On the other hand, there is little research on provision of olfactory information because smell is difficult to implement and control. In order to obtain necessary basic data when intend to provide smell for video reality, in this research, we investigated user perception of smell in diverse videos and then classified the videos based on the collected user perception data. To do so, we chose five main questions which were 'whether smell is present in the video'(smell presence), 'whether one desire to experience the smell with the video'(preference for smell presence with the video), 'whether one likes the smell itself'(preference for the smell itself), 'desired smell intensity if it is presented with the video'(smell intensity), and 'the degree of smell concreteness'(smell concreteness). After sampling video clips of various genre which are likely to receive either high and low ratings in the questions, we had participants watch each video after which they provided ratings on 7-point scale for the above five questions. Using the rating data for each video clips, we constructed scatter plots by pairing the five questions and representing the rating scale of each paired questions as X-Y axes in 2 dimensional spaces. The video clusters and distributional shape in the scatter plots would provide important insight into characteristics of each video clusters and about how to present olfactory information for video reality.

  • PDF

Frontal Face Video Analysis for Detecting Fatigue States

  • Cha, Simyeong;Ha, Jongwoo;Yoon, Soungwoong;Ahn, Chang-Won
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.6
    • /
    • pp.43-52
    • /
    • 2022
  • We can sense somebody's feeling fatigue, which means that fatigue can be detected through sensing human biometric signals. Numerous researches for assessing fatigue are mostly focused on diagnosing the edge of disease-level fatigue. In this study, we adapt quantitative analysis approaches for estimating qualitative data, and propose video analysis models for measuring fatigue state. Proposed three deep-learning based classification models selectively include stages of video analysis: object detection, feature extraction and time-series frame analysis algorithms to evaluate each stage's effect toward dividing the state of fatigue. Using frontal face videos collected from various fatigue situations, our CNN model shows 0.67 accuracy, which means that we empirically show the video analysis models can meaningfully detect fatigue state. Also we suggest the way of model adaptation when training and validating video data for classifying fatigue.

Land Cover Classification and Accuracy Assessment Using Aerial Videography and Landsat-TM Satellite Image -A Case Study of Taean Seashore National Park- (항공비디오와 Landsat-TM 자료를 이용한 지피의 분류와 평가 - 태안 해안국립공원을 사례로 -)

  • 서동조;박종화;조용현
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.27 no.4
    • /
    • pp.131-136
    • /
    • 1999
  • Aerial videography techniques have been used to inventory conditions associated with grassland, forests, and agricultural crop production. Most recently, aerial videography has been used to verity satellite image classifications as part of the natural ecosystem survey. The objectives of this study were: (1) to use aerial video images of the study area, one part of Taean Seashore National Park, for the accuracy assessment, and (2) to determine the suitability of aerial videography as an accuracy assessment, of the land cover classification with Landsat-TM data. Video images were collected twice, summer and winter seasons, and divided into two kinds of images, wide angle and narrow angle images. Accuracy assessment methods include the calculation of the error matrix, the overall accuracy and kappa coefficient of agreement. This study indicates that aerial videography is an effective tool for accuracy assessment of the satellite image classifications of which features are relatively large and continuous. And it would be possible to overcome the limits of the present natural ecosystem survey method.

  • PDF

Design and Implementation of ONVIF Video Analytics Service for a Smart IP Network camera (Smart IP 네트워크 카메라의 비디오 내용 분석 서비스 설계 및 구현)

  • Nguyen, Vo Thanh Phu;Nguyen, Thanh Binh;Chung, Sun-Tae;Kang, Ho-Seok
    • Proceedings of the Korea Multimedia Society Conference
    • /
    • 2012.05a
    • /
    • pp.102-105
    • /
    • 2012
  • ONVIF is becoming a de factor standard specification for supporting interoperability among network video products, which also supports a specification for video analytics service. A smart IP network camera is an IP network supporting video analytics. In this paper, we present our efforts in integrating ONVIF Video Analytics Service into our currently developing smart IP network camera(SS IPNC; Soongsil Smart IP Network Camera). SSIPNC supports object detection, tracking, classification, and event detection with proprietary configuration protocol and meta data formats. SSIPNC is based on TI' IPNC ONVIF implementation which supports ONVI Core specification, and several ONVIF services such as device service, imaging service and media service, but not video analytics service.

  • PDF

Classification of Degradation Types Based on Distribution of Blocky Blocks for IP-Based Video Services

  • Min, Kyung-Yeon;Lee, Seon-Oh;Sim, Dong-Gyu;Lee, Hyun-Woo;Ryu, Won;Lee, Kyoung-Hee
    • ETRI Journal
    • /
    • v.33 no.3
    • /
    • pp.454-457
    • /
    • 2011
  • In this letter, we propose a new quality measurement method to identify the causes of video quality degradation for IP-based video services. This degradation mainly results from network performance issues and video compression. The proposed algorithm identifies the causes based on statistical feature values from blocky block distribution in degraded IP-based videos. We found that the sensitivity and specificity of the proposed algorithm are 93.63% and 91.99%, respectively, in comparison with real error types and subjective test data.

Support Vector Machines-based classification of video file fragments (서포트 벡터 머신 기반 비디오 조각파일 분류)

  • Kang, Hyun-Suk;Lee, Young-Seok
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.16 no.1
    • /
    • pp.652-657
    • /
    • 2015
  • BitTorrent is an innovative protocol related to file-sharing and file-transferring, which allows users to receive pieces of files from multiple sharer on the Internet to make the pieces into complete files. In reality, however, free distribution of illegal or copyright related video data is counted for crime. Difficulty of regulation on the copyright of data on BitTorrent is caused by the fact that data is transferred with the pieces of files instead of the complete file formats. Therefore, the classification process of file formats of the digital contents should take precedence in order to restore digital contents from the pieces of files received from BitTorrent, and to check the violation of copyright. This study has suggested SVM classifier for the classification of digital files, which has the feature vector of histogram differential on the pieces of files. The suggested classifier has evaluated the performance with the division factor by applying the classifier to three different formats of video files.

Classification of Phornographic Video with using the Features of Multiple Audio (다중 오디오 특징을 이용한 유해 동영상의 판별)

  • Kim, Jung-Soo;Chung, Myung-Bum;Sung, Bo-Kyung;Kwon, Jin-Man;Koo, Kwang-Hyo;Ko, Il-Ju
    • 한국HCI학회:학술대회논문집
    • /
    • 2009.02a
    • /
    • pp.522-525
    • /
    • 2009
  • This paper proposed the content-based method of classifying filthy Phornographic video, which causes a big problem of modern society as the reverse function of internet. Audio data was used to extract the features from Phornographic video. There are frequency spectrum, autocorrelation, and MFCC as the feature of audio used in this paper. The sound that could be filthy contents was extracted, and the Phornographic was classified by measuring how much percentage of relevant sound was corresponding with the whole audio of video. For the experiment on the proposed method, The efficiency of classifying Phornographic was measured on each feature, and the measured result and comparison with using multi features were performed. I can obtain the better result than when only one feature of audio was extracted, and used.

  • PDF