• Title/Summary/Keyword: video retrieval

Search Result 372, Processing Time 0.022 seconds

Novel Intent based Dimension Reduction and Visual Features Semi-Supervised Learning for Automatic Visual Media Retrieval

  • kunisetti, Subramanyam;Ravichandran, Suban
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.6
    • /
    • pp.230-240
    • /
    • 2022
  • Sharing of online videos via internet is an emerging and important concept in different types of applications like surveillance and video mobile search in different web related applications. So there is need to manage personalized web video retrieval system necessary to explore relevant videos and it helps to peoples who are searching for efficient video relates to specific big data content. To evaluate this process, attributes/features with reduction of dimensionality are computed from videos to explore discriminative aspects of scene in video based on shape, histogram, and texture, annotation of object, co-ordination, color and contour data. Dimensionality reduction is mainly depends on extraction of feature and selection of feature in multi labeled data retrieval from multimedia related data. Many of the researchers are implemented different techniques/approaches to reduce dimensionality based on visual features of video data. But all the techniques have disadvantages and advantages in reduction of dimensionality with advanced features in video retrieval. In this research, we present a Novel Intent based Dimension Reduction Semi-Supervised Learning Approach (NIDRSLA) that examine the reduction of dimensionality with explore exact and fast video retrieval based on different visual features. For dimensionality reduction, NIDRSLA learns the matrix of projection by increasing the dependence between enlarged data and projected space features. Proposed approach also addressed the aforementioned issue (i.e. Segmentation of video with frame selection using low level features and high level features) with efficient object annotation for video representation. Experiments performed on synthetic data set, it demonstrate the efficiency of proposed approach with traditional state-of-the-art video retrieval methodologies.

Knowledge-based Video Retrieval System Using Korean Closed-caption (한국어 폐쇄자막을 이용한 지식기반 비디오 검색 시스템)

  • 조정원;정승도;최병욱
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.41 no.3
    • /
    • pp.115-124
    • /
    • 2004
  • The content-based retrieval using low-level features can hardly provide the retrieval result that corresponds with conceptual demand of user for intelligent retrieval. Video includes not only moving picture data, but also audio or closed-caption data. Knowledge-based video retrieval is able to provide the retrieval result that corresponds with conceptual demand of user because of performing automatic indexing with such a variety data. In this paper, we present the knowledge-based video retrieval system using Korean closed-caption. The closed-caption is indexed by Korean keyword extraction system including the morphological analysis process. As a result, we are able to retrieve the video by using keyword from the indexing database. In the experiment, we have applied the proposed method to news video with closed-caption generated by Korean stenographic system, and have empirically confirmed that the proposed method provides the retrieval result that corresponds with more meaningful conceptual demand of user.

Content-Based Video Retrieval Algorithms using Spatio-Temporal Information about Moving Objects (객체의 시공간적 움직임 정보를 이용한 내용 기반 비디오 검색 알고리즘)

  • Jeong, Jong-Myeon;Moon, Young-Shik
    • Journal of KIISE:Software and Applications
    • /
    • v.29 no.9
    • /
    • pp.631-644
    • /
    • 2002
  • In this paper efficient algorithms for content-based video retrieval using motion information are proposed, including temporal scale-invariant retrieval and temporal scale-absolute retrieval. In temporal scale-invariant video retrieval, the distance transformation is performed on each trail image in database. Then, from a given que교 trail the pixel values along the query trail are added in each distance image to compute the average distance between the trails of query image and database image, since the intensity of each pixel in distance image represents the distance from that pixel to the nearest edge pixel. For temporal scale-absolute retrieval, a new coding scheme referred to as Motion Retrieval Code is proposed. This code is designed to represent object motions in the human visual sense so that the retrieval performance can be improved. The proposed coding scheme can also achieve a fast matching, since the similarity between two motion vectors can be computed by simple bit operations. The efficiencies of the proposed methods are shown by experimental results.

The Design Interface and Mobile Internet Contents Type Analysis (모바일 인터넷 컨텐츠 유형 분석 및 인터페이스 설계)

  • Cho, Hyun-Seob;Ryu, In-Ho
    • Proceedings of the KAIS Fall Conference
    • /
    • 2011.05a
    • /
    • pp.371-374
    • /
    • 2011
  • Recently, retrieval of various video data has become an important issue as more and more multimedia content services are being provided. To effectively deal with video data, a semantic-based retrieval scheme that allows for processing diverse user queries and saving them on the database is required. In this regard, this paper proposes a semantic-based video retrieval system that allows the user to search diverse meanings of video data for electrical safety-related educational purposes by means of automatic annotation processing. If the user inputs a keyword to search video data for electrical safety-related educational purposes, the mobile agent of the proposed system extracts the features of the video data that are afterwards learned in a continuous manner, and detailed information on electrical safety education is saved on the database. The proposed system is designed to enhance video data retrieval efficiency for electrical safety-related educational purposes.

  • PDF

MPEG Video Retrieval Using U-Trees Construction (KD-Trees구조를 이용한MPEG 비디오 검색)

  • Kim, Daeil;Hong, Jong-Sun;Jang, Hye-Kyoung;Kim, Young-Ho;Kang, Dae-Seong
    • Proceedings of the IEEK Conference
    • /
    • 2003.07e
    • /
    • pp.1855-1858
    • /
    • 2003
  • In this paper, we propose image retrieval method more accurate and efficient than the conventional one. First of ail, we perform a shot detection and key frame extraction from the DC image constructed by DCT DC coefficients in the compressed video stream that is video compression standard such as MPEG[I][2]. We get principal axis applying PCA(Principal Component Analysis) to key frames for obtaining indexing information, and divide a domain. Video retrieval uses indexing information of high dimension. We apply KD-Trees(K Dimensional-Trees)[3] which shows efficient retrieval in data set of high dimension to video retrieval method. The proposed method can represent property of images more efficiently and property of domains more accurately using KD-Trees.

  • PDF

Design of video ontology for semantic web service (시맨틱 웹 서비스를 위한 동영상 온톨로지 설계)

  • Lee, Young-seok;Youn, Sung-dae
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2009.05a
    • /
    • pp.195-198
    • /
    • 2009
  • Recently, research in building up semantic web for exchanging information and knowledge is active. To make use of video contents as knowledge on semantic web, semantic-based retrieval should be preceded. At present, retrieval based on consentaneity between metadata and keyword is common used. In this paper, I propose ontolgy establishment which enlarge user participation and add usefulness value and history information. This will facilitate semantic retrieval as well as use of video contents by using collective Intelligence. The proposed ontology schema will allow semantic-based retrieval of video contents on semantic web get higher recall compared to current way of retrieval. Moreover it enables you to make use of various video contents as knowledge.

  • PDF

A Study on the Performance Analysis of Content-based Image & Video Retrieval Systems (내용기반 이미지 및 비디오 검색 시스템 성능분석에 관한 연구)

  • Kim, Seong-Hee
    • Journal of the Korean BIBLIA Society for library and Information Science
    • /
    • v.15 no.2
    • /
    • pp.97-115
    • /
    • 2004
  • The paper examined the concepts and features of content-based Image and Video retrieval systems. It then analyzed the retrieval performance of on five content_based retrieval systems in terms of usability and retrieval features. The results showed that the combination of content_based retrieval techniques and meta-data based retrieval will be able to improve the retrieval effectiveness.

  • PDF

Video Data Modeling for Supporting Structural and Semantic Retrieval (구조 및 의미 검색을 지원하는 비디오 데이타의 모델링)

  • 복경수;유재수;조기형
    • Journal of KIISE:Databases
    • /
    • v.30 no.3
    • /
    • pp.237-251
    • /
    • 2003
  • In this paper, we propose a video retrieval system to search logical structure and semantic contents of video data efficiently. The proposed system employs a layered modelling method that orBanifes video data in raw data layer, content layer and key frame layer. The layered modelling of the proposed system represents logical structures and semantic contents of video data in content layer. Also, the proposed system supports various types of searches such as text search, visual feature based similarity search, spatio-temporal relationship based similarity search and semantic contents search.

Natural Language based Video Retrieval System with Event Analysis of Multi-camera Image Sequence in Office Environment (사무실 환경 내 다중카메라 영상의 이벤트분석을 통한 자연어 기반 동영상 검색시스템)

  • Lim, Soo-Jung;Hong, Jin-Hyuk;Cho, Sung-Bae
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02a
    • /
    • pp.384-389
    • /
    • 2008
  • Recently, the necessity of systems which effectively store and retrieve video data has increased. Conventional video retrieval systems retrieve data using menus or text based keywords. Due to the lack of information, many video clips are simultaneously searched, and the user must have a certain level of knowledge to utilize the system. In this paper, we suggest a natural language based conversational video retrieval system that reflects users' intentions and includes more information than keyword based queries. This system can also retrieve from events or people to their movements. First, an event database is constructed based on meta-data which are generated by domain analysis for collected video in an office environment. Then, a script database is also constructed based on the query pre-processing and analysis. From that, a method to retrieve a video through a matching technique between natural language queries and answers is suggested and validated through performance and process evaluation for 10 users The natural language based retrieval system has shown its better efficiency in performance and user satisfaction than the menu based retrieval system.

  • PDF

A Study on the Alternative Method of Video Characteristics Using Captioning in Text-Video Retrieval Model (텍스트-비디오 검색 모델에서의 캡션을 활용한 비디오 특성 대체 방안 연구)

  • Dong-hun, Lee;Chan, Hur;Hyeyoung, Park;Sang-hyo, Park
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.17 no.6
    • /
    • pp.347-353
    • /
    • 2022
  • In this paper, we propose a method that performs a text-video retrieval model by replacing video properties using captions. In general, the exisiting embedding-based models consist of both joint embedding space construction and the CNN-based video encoding process, which requires a lot of computation in the training as well as the inference process. To overcome this problem, we introduce a video-captioning module to replace the visual property of video with captions generated by the video-captioning module. To be specific, we adopt the caption generator that converts candidate videos into captions in the inference process, thereby enabling direct comparison between the text given as a query and candidate videos without joint embedding space. Through the experiment, the proposed model successfully reduces the amount of computation and inference time by skipping the visual processing process and joint embedding space construction on two benchmark dataset, MSR-VTT and VATEX.