• Title/Summary/Keyword: Content-based video retrieval

Search Result 131, Processing Time 0.033 seconds

A Signature-based Video Indexing Scheme using Spatio-Temporal Modeling for Content-based and Concept-based Retrieval on Moving Objects (이동 객체의 내용 및 개념 기반 검색을 위한 시공간 모델링에 근거한 시그니쳐 기반 비디오 색인 기법)

  • Sim, Chun-Bo;Jang, Jae-U
    • The KIPS Transactions:PartD
    • /
    • v.9D no.1
    • /
    • pp.31-42
    • /
    • 2002
  • In this paper, we propose a new spatio-temporal representation scheme which can model moving objets trajectories effectively in video data and a new signature-based access method for moving objects trajectories which can support efficient retrieval on user query based on moving objects trajectories. The proposed spatio-temporal representation scheme supports content-based retrieval based on moving objects trajectories and concept-based retrieval based on concepts(semantics) which are acquired through the location information of moving objects trajectories. Also, compared with the sequential search, our signature-based access method can improve retrieval performance by reducing a large number of disk accesses because it access disk using only retrieved candidate signatures after it first scans all signatures and performs filtering before accessing the data file. Finally, we show the experimental results that proposed scheme is superior to the Li and Shan's scheme in terns of both retrieval effectiveness and efficiency.

Car Frame Extraction using Background Frame in Video (동영상에서 배경프레임을 이용한 차량 프레임 검출)

  • Nam, Seok-Woo;Oh, Hea-Seok
    • The KIPS Transactions:PartB
    • /
    • v.10B no.6
    • /
    • pp.705-710
    • /
    • 2003
  • Recent years, as a rapid development of multimedia technology, video database system to retrieve video data efficiently seems to core technology in the oriented society. This thesis describes an efficient automatic frame detection and location method for content based retrieval of video. Frame extraction part is consist of incoming / outgoing car frame extraction and car number frame extraction stage. We gain star/end time of car video also car number frames. Frames are selected at fixed time interval from video and key frames are selected by color scale histogram and edge operation method. Car frame recognized can be searched by content based retrieval method.

Similar Sub-Trajectory Retrieval based on k-warping Algorithm for Moving Objects in Video Databases (비디오 데이타베이스에서 이동 객체를 위한 k-워핑 알고리즘 기반 유사 부분궤적 검색)

  • 심춘보;장재우
    • Journal of KIISE:Databases
    • /
    • v.30 no.1
    • /
    • pp.14-26
    • /
    • 2003
  • Moving objects' trajectories play an important role in indexing video data on their content and semantics for content-based video retrieval. In this paper, we propose new similar sub-trajectory retrieval schemes based on k-warping algorithm for efficient retrieval on moving objects' trajectories in video data. The proposed schemes are fixed-replication similar sub-trajectory retrieval(FRSR) and variable-replication similar sub-trajectory retrieval(VRSR). The former can replicate motions with a fixed number for all motions being composed of the trajectory. The latter can replicate motions with a variable number. Our schemes support multiple properties including direction, distance, and time interval as well as a single property of direction, which is mainly used for modeling moving objects' trajectories. Finally, we show from our experiment that our schemes outperform Li's scheme(no-warping) and Shan's scheme(infinite-warping) in terns of precision and recall measures.

The Design Interface and Mobile Internet Contents Type Analysis (모바일 인터넷 컨텐츠 유형 분석 및 인터페이스 설계)

  • Cho, Hyun-Seob;Ryu, In-Ho
    • Proceedings of the KAIS Fall Conference
    • /
    • 2011.05a
    • /
    • pp.371-374
    • /
    • 2011
  • Recently, retrieval of various video data has become an important issue as more and more multimedia content services are being provided. To effectively deal with video data, a semantic-based retrieval scheme that allows for processing diverse user queries and saving them on the database is required. In this regard, this paper proposes a semantic-based video retrieval system that allows the user to search diverse meanings of video data for electrical safety-related educational purposes by means of automatic annotation processing. If the user inputs a keyword to search video data for electrical safety-related educational purposes, the mobile agent of the proposed system extracts the features of the video data that are afterwards learned in a continuous manner, and detailed information on electrical safety education is saved on the database. The proposed system is designed to enhance video data retrieval efficiency for electrical safety-related educational purposes.

  • PDF

BADA-$IV/I^2R$: Design & Implementation of an Efficient Content-based Image Retrieval System using a High-Dimensional Image Index Structure (바다-$IV/I^2R$: 고차원 이미지 색인 구조를 이용한 효율적인 내용 기반 이미지 검색 시스템의 설계와 구현)

  • Kim, Yeong-Gyun;Lee, Jang-Seon;Lee, Hun-Sun;Kim, Wan-Seok;Kim, Myeong-Jun
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.2S
    • /
    • pp.678-691
    • /
    • 2000
  • A variety of multimedia applications require multimedia database management systems to manage multimedia data, such as text, image, and video, as well as t support content-based image or video retrieval. In this paper we design and implement a content-based image retrieval system, BADA-IV/I$^2$R(Image Information Retrieval), which is developed based on BADA-IV multimedia database management system. In this system image databases can be efficiently constructed and retrieved with the visual features, such as color, shape, and texture, of image. we extend SQL statements to define image query based on both annotations and visual features of image together. A high-dimensional index structure, called CIR-tree, is also employed in the system to provide an efficient access method to image databases. We show that BADA-IV/I$^2$R provides a flexible way to define query for image retrieval and retrieves image data fast and effectively: the effectiveness and performance of image retrieval are shown by BEP(Bull's Eye Performance) that is used to measure the retrieval effectiveness in MPEG-7 and comparing the performance of CIR-tree with those of X-tree and TV-tree, respectively.

  • PDF

Novel Intent based Dimension Reduction and Visual Features Semi-Supervised Learning for Automatic Visual Media Retrieval

  • kunisetti, Subramanyam;Ravichandran, Suban
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.6
    • /
    • pp.230-240
    • /
    • 2022
  • Sharing of online videos via internet is an emerging and important concept in different types of applications like surveillance and video mobile search in different web related applications. So there is need to manage personalized web video retrieval system necessary to explore relevant videos and it helps to peoples who are searching for efficient video relates to specific big data content. To evaluate this process, attributes/features with reduction of dimensionality are computed from videos to explore discriminative aspects of scene in video based on shape, histogram, and texture, annotation of object, co-ordination, color and contour data. Dimensionality reduction is mainly depends on extraction of feature and selection of feature in multi labeled data retrieval from multimedia related data. Many of the researchers are implemented different techniques/approaches to reduce dimensionality based on visual features of video data. But all the techniques have disadvantages and advantages in reduction of dimensionality with advanced features in video retrieval. In this research, we present a Novel Intent based Dimension Reduction Semi-Supervised Learning Approach (NIDRSLA) that examine the reduction of dimensionality with explore exact and fast video retrieval based on different visual features. For dimensionality reduction, NIDRSLA learns the matrix of projection by increasing the dependence between enlarged data and projected space features. Proposed approach also addressed the aforementioned issue (i.e. Segmentation of video with frame selection using low level features and high level features) with efficient object annotation for video representation. Experiments performed on synthetic data set, it demonstrate the efficiency of proposed approach with traditional state-of-the-art video retrieval methodologies.

The Development of Efficient Multimedia Retrieval System of the Object-Based using the Hippocampal Neural Network (해마신경망을 이용한 관심 객체 기반의 효율적인 멀티미디어 검색 시스템의 개발)

  • Jeong Seok-Hoon;Kang Dae-Seong
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.2 s.308
    • /
    • pp.57-64
    • /
    • 2006
  • Tn this paper, We propose a user friendly object-based multimedia retrieval system using the HCNN(HippoCampus Neural Network. Most existing approaches to content-based retrieval rely on query by example or user based low-level features such as color, shape, texture. In this paper we perform a scene change detection and key frame extraction for the compressed video stream that is video compression standard such as MPEG. We propose a method for automatic color object extraction and ACE(Adaptive Circular filter and Edge) of content-based multimedia retrieval system. And we compose multimedia retrieval system after learned by the HCNN such extracted features. Proposed HCNN makes an adaptive real-time content-based multimedia retrieval system using excitatory teaming method that forwards important features to long-term memories and inhibitory learning method that forwards unimportant features to short-term memories controlled by impression.

Story-based Information Retrieval (스토리 기반의 정보 검색 연구)

  • You, Eun-Soon;Park, Seung-Bo
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.4
    • /
    • pp.81-96
    • /
    • 2013
  • Video information retrieval has become a very important issue because of the explosive increase in video data from Web content development. Meanwhile, content-based video analysis using visual features has been the main source for video information retrieval and browsing. Content in video can be represented with content-based analysis techniques, which can extract various features from audio-visual data such as frames, shots, colors, texture, or shape. Moreover, similarity between videos can be measured through content-based analysis. However, a movie that is one of typical types of video data is organized by story as well as audio-visual data. This causes a semantic gap between significant information recognized by people and information resulting from content-based analysis, when content-based video analysis using only audio-visual data of low level is applied to information retrieval of movie. The reason for this semantic gap is that the story line for a movie is high level information, with relationships in the content that changes as the movie progresses. Information retrieval related to the story line of a movie cannot be executed by only content-based analysis techniques. A formal model is needed, which can determine relationships among movie contents, or track meaning changes, in order to accurately retrieve the story information. Recently, story-based video analysis techniques have emerged using a social network concept for story information retrieval. These approaches represent a story by using the relationships between characters in a movie, but these approaches have problems. First, they do not express dynamic changes in relationships between characters according to story development. Second, they miss profound information, such as emotions indicating the identities and psychological states of the characters. Emotion is essential to understanding a character's motivation, conflict, and resolution. Third, they do not take account of events and background that contribute to the story. As a result, this paper reviews the importance and weaknesses of previous video analysis methods ranging from content-based approaches to story analysis based on social network. Also, we suggest necessary elements, such as character, background, and events, based on narrative structures introduced in the literature. We extract characters' emotional words from the script of the movie Pretty Woman by using the hierarchical attribute of WordNet, which is an extensive English thesaurus. WordNet offers relationships between words (e.g., synonyms, hypernyms, hyponyms, antonyms). We present a method to visualize the emotional pattern of a character over time. Second, a character's inner nature must be predetermined in order to model a character arc that can depict the character's growth and development. To this end, we analyze the amount of the character's dialogue in the script and track the character's inner nature using social network concepts, such as in-degree (incoming links) and out-degree (outgoing links). Additionally, we propose a method that can track a character's inner nature by tracing indices such as degree, in-degree, and out-degree of the character network in a movie through its progression. Finally, the spatial background where characters meet and where events take place is an important element in the story. We take advantage of the movie script to extracting significant spatial background and suggest a scene map describing spatial arrangements and distances in the movie. Important places where main characters first meet or where they stay during long periods of time can be extracted through this scene map. In view of the aforementioned three elements (character, event, background), we extract a variety of information related to the story and evaluate the performance of the proposed method. We can track story information extracted over time and detect a change in the character's emotion or inner nature, spatial movement, and conflicts and resolutions in the story.

Application of Speech Recognition with Closed Caption for Content-Based Video Segmentations

  • Son, Jong-Mok;Bae, Keun-Sung
    • Speech Sciences
    • /
    • v.12 no.1
    • /
    • pp.135-142
    • /
    • 2005
  • An important aspect of video indexing is the ability to segment video into meaningful segments, i.e., content-based video segmentation. Since the audio signal in the sound track is synchronized with image sequences in the video program, a speech signal in the sound track can be used to segment video into meaningful segments. In this paper, we propose a new approach to content-based video segmentation. This approach uses closed caption to construct a recognition network for speech recognition. Accurate time information for video segmentation is then obtained from the speech recognition process. For the video segmentation experiment for TV news programs, we made 56 video summaries successfully from 57 TV news stories. It demonstrates that the proposed scheme is very promising for content-based video segmentation.

  • PDF

Knowledge-based Video Retrieval System Using Korean Closed-caption (한국어 폐쇄자막을 이용한 지식기반 비디오 검색 시스템)

  • 조정원;정승도;최병욱
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.41 no.3
    • /
    • pp.115-124
    • /
    • 2004
  • The content-based retrieval using low-level features can hardly provide the retrieval result that corresponds with conceptual demand of user for intelligent retrieval. Video includes not only moving picture data, but also audio or closed-caption data. Knowledge-based video retrieval is able to provide the retrieval result that corresponds with conceptual demand of user because of performing automatic indexing with such a variety data. In this paper, we present the knowledge-based video retrieval system using Korean closed-caption. The closed-caption is indexed by Korean keyword extraction system including the morphological analysis process. As a result, we are able to retrieve the video by using keyword from the indexing database. In the experiment, we have applied the proposed method to news video with closed-caption generated by Korean stenographic system, and have empirically confirmed that the proposed method provides the retrieval result that corresponds with more meaningful conceptual demand of user.