• Title/Summary/Keyword: Image Retrieval and Extraction

Search Result 159, Processing Time 0.025 seconds

FE-CBIRS Using Color Distribution for Cut Retrieval in IPTV (IPTV에서 컷 검색을 위한 색 분포정보를 이용한 FE-CBIRS)

  • Koo, Gun-Seo
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.1
    • /
    • pp.91-97
    • /
    • 2009
  • This paper proposes novel FE-CBIRS that finds best position of a cut to be retrieved based on color feature distribution in digital contents of IPTV. Conventional CBIRS have used a method that utilizes both color and shape information together to classify images, as well as a method that utilizes both feature information of the entire region and feature information of a partial region that is extracted by segmentation for searching. Also, in the algorithm, average, standard deviation and skewness values are used in case of color features for each hue, saturation and intensity values respectively. Furthermore, in case of using partial regions, only a few major colors are used and in case of shape features, the invariant moment is mainly used on the extracted partial regions. Due to these reasons, some problems have been issued in CBIRS in processing time and accuracy so far. Therefore, in order to tackle these problems, this paper proposes the FE-CBIRS that makes searching speed faster by classifying and indexing the extracted color information by each class and by using several cuts that are restricted in range as comparative images.

Keyword Spotting on Hangul Document Images Using Character Feature Models (문자 별 특징 모델을 이용한 한글 문서 영상에서 키워드 검색)

  • Park, Sang-Cheol;Kim, Soo-Hyung;Choi, Deok-Jai
    • The KIPS Transactions:PartB
    • /
    • v.12B no.5 s.101
    • /
    • pp.521-526
    • /
    • 2005
  • In this Paper, we propose a keyword spotting system as an alternative to searching system for poor quality Korean document images and compare the Proposed system with an OCR-based document retrieval system. The system is composed of character segmentation, feature extraction for the query keyword, and word-to-word matching. In the character segmentation step, we propose an effective method to remove the connectivity between adjacent characters and a character segmentation method by making the variance of character widths minimum. In the query creation step, feature vector for the query is constructed by a combination of a character model by typeface. In the matching step, word-to-word matching is applied base on a character-to-character matching. We demonstrated that the proposed keyword spotting system is more efficient than the OCR-based one to search a keyword on the Korean document images, especially when the quality of documents is quite poor and point size is small.

A Reduction Method of Over-Segmented Regions at Image Segmentation based on Homogeneity Threshold (동질성 문턱 값 기반 영상분할에서 과분할 영역 축소 방법)

  • Han, Gi-Tae
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.1 no.1
    • /
    • pp.55-68
    • /
    • 2012
  • In this paper, we propose a novel method to solve the problem of excessive segmentation out of the method of segmenting regions from an image using Homogeneity Threshold($H_T$). The algorithm of the previous image segmentation based on $H_T$ was carried out region growth by using only the center pixel of selected window. Therefore it was caused resulting in excessive segmented regions. However, before carrying region growth, the proposed method first of all finds out whether the selected window is homogeneity or not. Subsequently, if the selected window is homogeneity it carries out region growth using the total pixels of selected window. But if the selected window is not homogeneity, it carries out region growth using only the center pixel of selected window. So, the method can reduce remarkably the number of excessive segmented regions of image segmentation based on $H_T$. In order to show the validity of the proposed method, we carried out multiple experiments to compare the proposed method with previous method in same environment and conditions. As the results, the proposed method can reduce the number of segmented regions above 40% and doesn't make any difference in the quality of visual image when we compare with previous method. Especially, when we compare the image united with regions of descending order by size of segmented regions in experimentation with the previous method, even though the united image has regions more than 1,000, we can't recognize what the image means. However, in the proposed method, even though image is united by segmented regions less than 10, we can recognize what the image is. For these reason, we expect that the proposed method will be utilized in various fields, such as the extraction of objects, the retrieval of informations from the image, research for anatomy, biology, image visualization, and animation and so on.

Similar Movie Contents Retrieval Using Peak Features from Audio (오디오의 Peak 특징을 이용한 동일 영화 콘텐츠 검색)

  • Chung, Myoung-Bum;Sung, Bo-Kyung;Ko, Il-Ju
    • Journal of Korea Multimedia Society
    • /
    • v.12 no.11
    • /
    • pp.1572-1580
    • /
    • 2009
  • Combing through entire video files for the purpose of recognizing and retrieving matching movies requires much time and memory space. Instead, most current similar movie-matching methods choose to analyze only a part of each movie's video-image information. Yet, these methods still share a critical problem of erroneously recognizing as being different matching videos that have been altered only in resolution or converted merely with a different codecs. This paper proposes an audio-information-based search algorithm by which similar movies can be identified. The proposed method prepares and searches through a database of movie's spectral peak information that remains relatively steady even with changes in the bit-rate, codecs, or sample-rate. The method showed a 92.1% search success rate, given a set of 1,000 video files whose audio-bit-rate had been altered or were purposefully written in a different codec.

  • PDF

Toward a Key-frame Automatic Extraction Method for Video Storyboard Surrogates Based on Users' EEG Signals and Discriminant Analysis (뇌파측정기술(EEG)과 판별분석을 이용한 영상물의 키프레임 자동 분류 방안 연구)

  • Kim, Hyun-Hee
    • Journal of the Korean Society for information Management
    • /
    • v.32 no.3
    • /
    • pp.377-396
    • /
    • 2015
  • This study proposed a key-frame automatic extraction method for video storyboard surrogates based on users' cognitive responses, EEG signals and discriminant analysis. Using twenty participants, we examined which ERP pattern is suitable for each step, assuming that there are five image recognition and process steps (stimuli attention, stimuli perception, memory retrieval, stimuli/memory comparison, relevance judgement). As a result, we found that each step has a suitable ERP pattern, such as N100, P200, N400, P3b, and P600. Moreover, we also found that the peak amplitude of left parietal lobe (P7) and the latency of FP2 are important variables in distinguishing among relevant, partial, and non-relevant frames. Using these variables, we conducted a discriminant analysis to classify between relevant and non-relevant frames.

Real-time Face Extraction for Content-based Image Retrieval (내용기반 영상 검색을 위한 실시간 얼굴 영역 추출)

  • 이미숙;이성환
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 1996.06a
    • /
    • pp.169-174
    • /
    • 1996
  • 객체 인식은 대용량의 영상 데이터를 분석, 탐색하고 재구성하기 위한 내용기반 영상 검색의 매우 중요한 분야이며, 특히 인간의 얼굴은 검색 영상 내에서 대부분 주요한 장면에 위치하고 있기 때문에 그 비중이 매우 크다. 본 논문에서는 내용기반 영상 검색을 위한 실시간 얼굴 영역 추출 방법을 제안한다. 제안된 방법에서는 다층 피라미드 구조와 간단한 형태의 머리 형판을 사용하여 얼굴의 후보 영역을 추출한 후, 보다 정확한 얼굴 영역을 추출하기 위하여 후보 영역 내에서 눈의 위치를 탐색하고, 두 눈의 위치를 기준으로 최종적인 얼굴 영역을 추출하였다. 얼굴 후보 영역 추출 단계에서는 얼굴의 형태 정보를 포함하고 있는 모자이크 형판을 사용하여 머리와 턱을 포함한 얼굴 영역을 추출하였으며, 눈 위치 추출 단계에서는 눈의 위치 정보를 사용하여 눈의 탐색 영역을 결정하고, 탐색 영역 내에서 이진 영상 형판을 사용하여 눈의 위치를 추출한 후, 눈 영역의 무게 중심을 눈의 중심 위치로 설정하였다. 마지막 얼굴 영역 추출단계에서는 두 눈의 위치를 기준으로 사각형의 영역을 얼굴 영역으로 추출하였다. 제안된 방법의 성능을 검증하기 위하여 1700장의 다양한 영상에 대하여 실험하였으며, 실험 결과 한 장의 영상에서 얼굴 영역을 추출하는데 있어서, Pentium 166Mz의 PC상에서 평균 3.2초의 처리 속도와 91.7%의 추출률을 보임으로써, 실시간 얼굴 영역 추출에 매우 효과적임을 알 수 있었다.

  • PDF

Improved Image Retrieval Method using Color and Texture Feature Extraction (컬러 및 질감 특징 추출을 이용한 향상된 이미지 검색 기법)

  • Park, Sunghyun;Shin, In-Kyoung;Ahn, Hyochang;Lee, Yong-Hwan;Cho, Han-Jin;Lee, June-Hwan
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2013.11a
    • /
    • pp.1563-1566
    • /
    • 2013
  • 최근 네트워크와 멀티미디어 관련 기술의 발달로 이미지 및 동영상과 같은 대용량 멀티미디어 데이터가 증가하고 있다. 이에 따라 대용량의 데이터에서 영상 정보의 효율적인 검색 방법이 요구 되고 있다. 하지만 기존의 전통적인 색인기술은 관리자가 영상을 직접 보면서 적절한 텍스트 내용을 입력하는 방법으로 시간이 많이 소요되며, 관리자의 성향에 따라 색인어의 입력이 다를 수 있어 검색시 오류를 발생시킬 수 있다. 따라서 본 논문에서는 영상으로부터 컬러 특징과 질감 특징을 추출하여 보다 효율적으로 내용 기반 영상 검색을 수행하는 방법을 제안한다. 실험을 통하여 다른 기존의 영상 검색 방법보다 검색 효율성에서 안정적이며 보다 나은 결과를 얻음을 확인한다.

Hangul Component Decomposition in Outline Fonts (한글 외곽선 폰트의 자소 분할)

  • Koo, Sang-Ok;Jung, Soon-Ki
    • Journal of the Korea Computer Graphics Society
    • /
    • v.17 no.4
    • /
    • pp.11-21
    • /
    • 2011
  • This paper proposes a method for decomposing a Hangul glyph of outline fonts into its initial, medial and final components using statistical-structural information. In a font family, the positions of components are statistically consistent and the stroke relationships of a Hangul character reflect its structure. First, we create the component histograms that accumulate the shapes and positions of the same components. Second, we make pixel clusters from character image based on pixel direction probabilities and extract the candidate strokes using position, direction, size of clusters and adjacencies between clusters. Finally, we find the best structural match between candidate strokes and predefined character model by relaxation labeling. The proposed method in this paper can be used for a study on formative characteristics of Hangul font, and for a font classification/retrieval system.

Video Scene Detection using Shot Clustering based on Visual Features (시각적 특징을 기반한 샷 클러스터링을 통한 비디오 씬 탐지 기법)

  • Shin, Dong-Wook;Kim, Tae-Hwan;Choi, Joong-Min
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.2
    • /
    • pp.47-60
    • /
    • 2012
  • Video data comes in the form of the unstructured and the complex structure. As the importance of efficient management and retrieval for video data increases, studies on the video parsing based on the visual features contained in the video contents are researched to reconstruct video data as the meaningful structure. The early studies on video parsing are focused on splitting video data into shots, but detecting the shot boundary defined with the physical boundary does not cosider the semantic association of video data. Recently, studies on structuralizing video shots having the semantic association to the video scene defined with the semantic boundary by utilizing clustering methods are actively progressed. Previous studies on detecting the video scene try to detect video scenes by utilizing clustering algorithms based on the similarity measure between video shots mainly depended on color features. However, the correct identification of a video shot or scene and the detection of the gradual transitions such as dissolve, fade and wipe are difficult because color features of video data contain a noise and are abruptly changed due to the intervention of an unexpected object. In this paper, to solve these problems, we propose the Scene Detector by using Color histogram, corner Edge and Object color histogram (SDCEO) that clusters similar shots organizing same event based on visual features including the color histogram, the corner edge and the object color histogram to detect video scenes. The SDCEO is worthy of notice in a sense that it uses the edge feature with the color feature, and as a result, it effectively detects the gradual transitions as well as the abrupt transitions. The SDCEO consists of the Shot Bound Identifier and the Video Scene Detector. The Shot Bound Identifier is comprised of the Color Histogram Analysis step and the Corner Edge Analysis step. In the Color Histogram Analysis step, SDCEO uses the color histogram feature to organizing shot boundaries. The color histogram, recording the percentage of each quantized color among all pixels in a frame, are chosen for their good performance, as also reported in other work of content-based image and video analysis. To organize shot boundaries, SDCEO joins associated sequential frames into shot boundaries by measuring the similarity of the color histogram between frames. In the Corner Edge Analysis step, SDCEO identifies the final shot boundaries by using the corner edge feature. SDCEO detect associated shot boundaries comparing the corner edge feature between the last frame of previous shot boundary and the first frame of next shot boundary. In the Key-frame Extraction step, SDCEO compares each frame with all frames and measures the similarity by using histogram euclidean distance, and then select the frame the most similar with all frames contained in same shot boundary as the key-frame. Video Scene Detector clusters associated shots organizing same event by utilizing the hierarchical agglomerative clustering method based on the visual features including the color histogram and the object color histogram. After detecting video scenes, SDCEO organizes final video scene by repetitive clustering until the simiarity distance between shot boundaries less than the threshold h. In this paper, we construct the prototype of SDCEO and experiments are carried out with the baseline data that are manually constructed, and the experimental results that the precision of shot boundary detection is 93.3% and the precision of video scene detection is 83.3% are satisfactory.