• Title/Summary/Keyword: Video Retrieval

Search Result 372, Processing Time 0.022 seconds

Video Shot Retrieval in H.264/AVC compression domain (H.264/AVC 압축 영역에서의 동영상 검색)

  • Byun Ju-Wan;Kim Sung-Min;Won Chee-Sun
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.5 s.311
    • /
    • pp.72-78
    • /
    • 2006
  • In this paper, we present a video shot retrieval algorithm in H.264/AVC compression domain. Unlike previous standards such as MPEG-2 and 4, H.264/AVC supports a variable block size for motion compensation. Therefore, existing video retrieval algorithms exploiting the motion vectors in MPEG-2 and 4 domains are not appropriate for H.264/AVC. So, we devise a method to project motion vectors with larger than $4{\times}4$ block sizes into those for the smallest $4{\times}4$ blocks. It also uses correlations among features for the measure of similarity. Experimental results with standard videos of 10558 frames and commercial videos of 48161 frames show that the proposed method yields ANMRR less than 0.2.

Pattern Similarity Retrieval of Data Sequences for Video Retrieval System (비디오 검색 시스템을 위한 데이터 시퀀스 패턴 유사성 검색)

  • Lee Seok-Lyong
    • The KIPS Transactions:PartD
    • /
    • v.13D no.3 s.106
    • /
    • pp.347-356
    • /
    • 2006
  • A video stream can be represented by a sequence of data points in a multidimensional space. In this paper, we introduce a trend vector that approximates values of data points in a sequence and represents the moving trend of points in the sequence, and present a pattern similarity matching method for data sequences using the trend vector. A sequence is partitioned into multiple segments, each of which is represented by a trend vector. The query processing is based on the comparison of these vectors instead of scanning data elements of entire sequences. Using the trend vector, our method is designed to filter out irrelevant sequences from a database and to find similar sequences with respect to a query. We have performed an extensive experiment on synthetic sequences as well as video streams. Experimental results show that the precision of our method is up to 2.1 times higher and the processing time is up to 45% reduced, compared with an existing method.

Video Event Analysis and Retrieval System for the KFD Web Database System (KFD 웹 데이터베이스 시스템을 위한 동영상 이벤트 분석 및 검색 시스템)

  • Oh, Seung-Geun;Im, Young-Hee;Chung, Yong-Wha;Chang, Jin-Kyung;Park, Dai-Hee
    • The Journal of the Korea Contents Association
    • /
    • v.10 no.11
    • /
    • pp.20-29
    • /
    • 2010
  • The typical Kinetic Family Drawing (KFD) Web database system, a form of prototype system, has been developed, relying on the suggestions from family art therapists, with an aim to handle large amounts of assessment data and to facilitate effective implement of assessment activities. However, unfortunately such a system has an intrinsic problem that it fails to collect clients' behaviors, attitudes, facial expressions, voices, and other critical information observed while they are drawing. Accordingly we propose the ontology based video event analysis and video retrieval system in this paper, in order to enhance the function of a KFD Web database system by using a web camera and drawing tool. More specifically, a newly proposed system is designed to deliver two kinds of services: the client video retrieval service and the sketch video retrieval service, accompanied by a summary report of occurred events and dynamic behaviors relative to each family member object, respectively. The proposed system can support the reinforced KFD assessments by providing quantitative and subjective information on clients' working attitudes and behaviors, and KFD preparation processes.

Extraction of Superimposed-Caption Frame Scopes and Its Regions for Analyzing Digital Video (비디오 분석을 위한 자막프레임구간과 자막영역 추출)

  • Lim, Moon-Cheol;Kim, Woo-Saeng
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.11
    • /
    • pp.3333-3340
    • /
    • 2000
  • Recently, Requnremeni for video data have been increased rapidly by high progress of both hardware and cornpression technique. Because digital video data are unformed and mass capacity, it needs various retrieval techniquesjust as contednt-based rehieval Superimposed-caption ina digital video can help us to analyze the video story easier and be used as indexing information for many retrieval techniques In this research we propose a new method that segments the caption as analyzing texture eature of caption regions in each video frame, and that extracts the accurate scope of superimposed-caption frame and its key regions and color by measunng cominuity of caption regions between frames

  • PDF

Design of Moving Picture Retrieval System using Scene Change Technique (장면 전환 기법을 이용한 동영상 검색 시스템 설계)

  • Kim, Jang-Hui;Kang, Dae-Seong
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.44 no.3
    • /
    • pp.8-15
    • /
    • 2007
  • Recently, it is important to process multimedia data efficiently. Especially, in case of retrieval of multimedia information, technique of user interface and retrieval technique are necessary. This paper proposes a new technique which detects cuts effectively in compressed image information by MPEG. A cut is a turning point of scenes. The cut-detection is the basic work and the first-step for video indexing and retrieval. Existing methods have a weak point that they detect wrong cuts according to change of a screen such as fast motion of an object, movement of a camera and a flash. Because they compare between previous frame and present frame. The proposed technique detects shots at first using DC(Direct Current) coefficient of DCT(Discrete Cosine Transform). The database is composed of these detected shots. Features are extracted by HMMD color model and edge histogram descriptor(EHD) among the MPEG-7 visual descriptors. And detections are performed in sequence by the proposed matching technique. Through this experiments, an improved video segmentation system is implemented that it performs more quickly and precisely than existing techniques have.

Creation of Soccer Video Highlight Using The Structural Features of Caption (자막의 구조적 특징을 이용한 축구 비디오 하이라이트 생성)

  • Huh, Moon-Haeng;Shin, Seong-Yoon;Lee, Yang-Weon;Ryu, Keun-Ho
    • The KIPS Transactions:PartD
    • /
    • v.10D no.4
    • /
    • pp.671-678
    • /
    • 2003
  • A digital video is usually very long temporally. requiring large storage capacity. Therefore, users want to watch pre-summarized video before they watch a large long video. Especially in the field of sports video, they want to watch a highlight video. Consequently, highlight video is used that the viewers decide whether it is valuable for them to watch the video or not. This paper proposes how to create soccer video highlight using the structural features of the caption such as temporal and spatial features. Caption frame intervals and caption key frames are extracted by using those structural features. And then, highlight video is created by using scene relocation, logical indexing and highlight creation rule. Finally. retrieval and browsing of highlight and video segment is performed by selection of item on browser.

Segmentation of Objects of Interest for Video Content Analysis (동영상 내용 분석을 위한 관심 객체 추출)

  • Park, So-Jung;Kim, Min-Hwan
    • Journal of Korea Multimedia Society
    • /
    • v.10 no.8
    • /
    • pp.967-980
    • /
    • 2007
  • Video objects of interest play an important role in representing the video content and are useful for improving the performance of video retrieval and compression. The objects of interest may be a main object in describing contents of a video shot or a core object that a video producer wants to represent in the video shot. We know that any object attracting one's eye much in the video shot may not be an object of interest and a non-moving object may be an object of interest as well as a moving one. However it is not easy to define an object of interest clearly, because procedural description of human interest is difficult. In this paper, a set of four filtering conditions for extracting moving objects of interest is suggested, which is defined by considering variation of location, size, and moving pattern of moving objects in a video shot. Non-moving objects of interest are also defined as another set of four extracting conditions that are related to saliency of color/texture, location, size, and occurrence frequency of static objects in a video shot. On a test with 50 video shots, the segmentation method based on the two sets of conditions could extract the moving and non-moving objects of interest chosen manually on accuracy of 84%.

  • PDF

The Development of Efficient Multimedia Retrieval System of the Object-Based using the Hippocampal Neural Network (해마신경망을 이용한 관심 객체 기반의 효율적인 멀티미디어 검색 시스템의 개발)

  • Jeong Seok-Hoon;Kang Dae-Seong
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.2 s.308
    • /
    • pp.57-64
    • /
    • 2006
  • Tn this paper, We propose a user friendly object-based multimedia retrieval system using the HCNN(HippoCampus Neural Network. Most existing approaches to content-based retrieval rely on query by example or user based low-level features such as color, shape, texture. In this paper we perform a scene change detection and key frame extraction for the compressed video stream that is video compression standard such as MPEG. We propose a method for automatic color object extraction and ACE(Adaptive Circular filter and Edge) of content-based multimedia retrieval system. And we compose multimedia retrieval system after learned by the HCNN such extracted features. Proposed HCNN makes an adaptive real-time content-based multimedia retrieval system using excitatory teaming method that forwards important features to long-term memories and inhibitory learning method that forwards unimportant features to short-term memories controlled by impression.

Caption Detection and Recognition for Video Image Information Retrieval (비디오 영상 정보 검색을 위한 문자 추출 및 인식)

  • 구건서
    • Journal of the Korea Computer Industry Society
    • /
    • v.3 no.7
    • /
    • pp.901-914
    • /
    • 2002
  • In this paper, We propose an efficient automatic caption detection and location method, caption recognition using FE-MCBP(Feature Extraction based Multichained BackPropagation) neural network for content based retrieval of video. Frames are selected at fixed time interval from video and key frames are selected by gray scale histogram method. for each key frames, segmentation is performed and caption lines are detected using line scan method. lastly each characters are separated. This research improves speed and efficiency by color segmentation using local maximum analysis method before line scanning. Caption detection is a first stage of multimedia database organization and detected captions are used as input of text recognition system. Recognized captions can be searched by content based retrieval method.

  • PDF

XMARS : XML-based Multimedia Annotation and Retrieval System (XMARS : XML 기반 멀티미디어 주석 및 검색 시스템)

  • Nam, Yun-Young;Hwang, Een-Jun
    • The KIPS Transactions:PartB
    • /
    • v.9B no.5
    • /
    • pp.541-548
    • /
    • 2002
  • This paper proposes an XML based Multimedia Annotation and Retrieval System, which can represent and retrieve video data efficiently using XML. The system provides a graphical user interface for annotating, searching, and browsing multimedia data. It is Implemented based on the hierarchical metadata model to represent multimedia information. The metadata about video is organized based on multimedia description schema using XML Schema that basically conforms to the MPEG-7 standard. Also, for the effective indexing and retrieval of multimedia data, video segments are annotated and categorized using the closed caption.