• Title/Summary/Keyword: Image Caption

Search Result 51, Processing Time 0.032 seconds

A Method for Recovering Text Regions in Video using Extended Block Matching and Region Compensation (확장적 블록 정합 방법과 영역 보상법을 이용한 비디오 문자 영역 복원 방법)

  • 전병태;배영래
    • Journal of KIISE:Software and Applications
    • /
    • v.29 no.11
    • /
    • pp.767-774
    • /
    • 2002
  • Conventional research on image restoration has focused on restoring degraded images resulting from image formation, storage and communication, mainly in the signal processing field. Related research on recovering original image information of caption regions includes a method using BMA(block matching algorithm). The method has problem with frequent incorrect matching and propagating the errors by incorrect matching. Moreover, it is impossible to recover the frames between two scene changes when scene changes occur more than twice. In this paper, we propose a method for recovering original images using EBMA(Extended Block Matching Algorithm) and a region compensation method. To use it in original image recovery, the method extracts a priori knowledge such as information about scene changes, camera motion and caption regions. The method decides the direction of recovery using the extracted caption information(the start and end frames of a caption) and scene change information. According to the direction of recovery, the recovery is performed in units of character components using EBMA and the region compensation method. Experimental results show that EBMA results in good recovery regardless of the speed of moving object and complexity of background in video. The region compensation method recovered original images successfully, when there is no information about the original image to refer to.

A Method for Recovering Image Data for Caption Regions and Replacing Caption Text (비디오 자막 영역 원영상 복원 후 자막 교환 방법)

  • Chun, Byung-Tae;Han, Kyu-Seo;Bae, Young-Lae
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2001.10a
    • /
    • pp.743-746
    • /
    • 2001
  • 멀티미디어 정보 중 비디오 데이터는 정보량이 많기 때문에 자동화된 비디오 영상 처리 기술이 필요하다. 시청자의 이해와 시청의 편의성을 제공하기 위하여 대부분 비디오에 자막을 삽입하게 된다. 외국 방송물 및 영화에 삽입된 다른 언어 자막을 편집 과정에서 자막으로 교환 할 필요성이 종종 있게 된다. 기존의 방법들은 자막 부분을 충분히 포함하는 일정 영역에 특정 색상을 채운 후 새로운 자막을 삽입하게 된다. 기존 방법의 문제점은 많은 영역에 대해 비디오 영상 정보를 손실 시킴으로써 시청자에게 시청의 불편을 초래하고 자막 교환의 비 효율성과 부 자연스러움을 발생시킨다. 본 논문에서는 기존 방법의 문제점을 극복하기 위하여 자막 영역을 원영상으로 복원한 후 다른 자막으로 교환하는 방법을 제안하고자 한다. 원영상 복원을 위하여 비디오 정보와 BMA(Block Matching Algorithm)를 이용한 원영상 복원 방법을 제안하고, 복원된 영역에 다른 자막으로 교환함으로써 효과적인 자막 교환 방법을 제안하고자 한다. 실험 결과 원영상 복원을 이용한 자막 교환 방법은 기존 방법에 비해 자연스럽고 효과적인 교환 방법임을 볼 수 있었다.

  • PDF

The Signification of Words and Photography in Photojournalism (포토저널리즘 사진과 캡션의 의미작용에 대한 연구)

  • Chung, Hong-Gi
    • Korean journal of communication and information
    • /
    • v.18
    • /
    • pp.231-268
    • /
    • 2002
  • This is a study about the audience's aspect of decoding the photography and words comprising photojournalism. For this study, the way of audience's decoding five pictures depicting a foreign worker's death and the captions for each were analyzed. Ethnographic method was used to find distinctive features of the audiences in decoding the pictures. Picture images without any title and caption shown to audiences bring quite different decodings from those of picture images titled and captioned. This led to the following three conclusions. First, picture image added by caption transforms the way the audience decode. Second, the cultural background of an audience works as a major variable, which means how the audience decode the picture image depends on what sort of cultural background the audience has. Third, it is also learned that picture image not captioned would not be able to represent the reality faithfully. The author made an experimental trial in this thesis how the communication process works successfully in a reality that photojournalists tried to represent and the way the audience decode it for the successful communication in photojournalism.

  • PDF

Extraction of Features in key frames of News Video for Content-based Retrieval (내용 기반 검색을 위한 뉴스 비디오 키 프레임의 특징 정보 추출)

  • Jung, Yung-Eun;Lee, Dong-Seop;Jeon, Keun-Hwan;Lee, Yang-Weon
    • The Transactions of the Korea Information Processing Society
    • /
    • v.5 no.9
    • /
    • pp.2294-2301
    • /
    • 1998
  • The aim of this paper is to extract features from each news scenes for example, symbol icon which can be distinct each broadcasting corp, icon and caption which are has feature and important information for the scene in respectively, In this paper, we propose extraction methods of caption that has important prohlem of news videos and it can be classified in three steps, First of al!, we converted that input images from video frame to YIQ color vector in first stage. And then, we divide input image into regions in clear hy using equalized color histogram of input image, In last, we extracts caption using edge histogram based on vertical and horizontal line, We also propose the method which can extract news icon in selected key frames by the difference of inter-histogram and can divide each scene by the extracted icon. In this paper, we used comparison method of edge histogram instead of complex methcxls based on color histogram or wavelet or moving objects, so we shorten computation through using simpler algorithm. and we shown good result of feature's extraction.

  • PDF

Study on News Video Character Extraction and Recognition (뉴스 비디오 자막 추출 및 인식 기법에 관한 연구)

  • 김종열;김성섭;문영식
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.40 no.1
    • /
    • pp.10-19
    • /
    • 2003
  • Caption information in news videos can be useful for video indexing and retrieval since it usually suggests or implies the contents of the video very well. In this paper, a new algorithm for extracting and recognizing characters from news video is proposed, without a priori knowledge such as font type, color, size of character. In the process of text region extraction, in order to improve the recognition rate for videos with complex background at low resolution, continuous frames with identical text regions are automatically detected to compose an average frame. The image of the averaged frame is projected to horizontal and vertical direction, and we apply region filling to remove backgrounds to produce the character. Then, K-means color clustering is applied to remove remaining backgrounds to produce the final text image. In the process of character recognition, simple features such as white run and zero-one transition from the center, are extracted from unknown characters. These feature are compared with the pre-composed character feature set to recognize the characters. Experimental results tested on various news videos show that the proposed method is superior in terms of caption extraction ability and character recognition rate.

A Study on Analyzing Caption Characteristic for Recovering Original Images of Caption Region in TV Scene (원 영상 복원을 위한 TV 자막 특성 분석에 관한 연구)

  • Chun, Byung-Tae
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.10 no.4
    • /
    • pp.177-182
    • /
    • 2010
  • Research on recovering original images from captions has been widely done in a reusability point of view. In usual, dynamic images imported from foreign countries often have captions of foreign languages, so it is necessary to translate them into one's language. For the natural exchange of captions without loss of original images, recovering the images corresponding to captions is necessary. However, though recovering original images is very important, systematic analysis on the characteristics of captions has not been done yet. Therefore, in this paper, we first survey the classification methods of TV programs at academic worlds, broadcasting stations, and broadcasting organizations, and then analyses the frequency of captions, importance of caption contents, and necessity of recovering according to their types. Also, we analyze the characteristics of captions which are significantly recognized to be necessary to recover, and use them as recovering information.

Analysis of the Reading Materials in Elementary School Science Textbooks developed under the 2009 Revised National Science Curriculum (2009 개정 과학교육과정에 따른 초등학교 과학 교과서의 읽기자료 분석)

  • Koh, Hanjoong;Seok, Jongim;Kang, Sukjin
    • Journal of Korean Elementary Science Education
    • /
    • v.36 no.2
    • /
    • pp.129-142
    • /
    • 2017
  • In this study, the characteristics of the reading materials in elementary school science textbooks developed under the 2009 revised National Science Curriculum were investigated. The criteria for classifying the reading materials were the types of topic, purpose, students' activity, and presentation. The visual images in the reading materials were also analyzed from the viewpoint of type, role, caption type, and proximity type. The results indicated that the number of the reading materials in the 2009 revised science textbooks decreased compared to that of the 2007 revised science textbooks. It was also found that the frequencies of the reading materials expanding concepts of the text and/or requiring corresponding students' inquiry increased. More visual images were used in the reading materials of the 2009 revised science textbooks. However, several limitations were still found to exist; most visual images were illustration and/or picture; many visual images were presented without a caption; there was a problem in the proximity of visual image to text.

A Method for Character Segmentation using MST(Minimum Spanning Tree) (MST를 이용한 문자 영역 분할 방법)

  • Chun, Byung-Tae;Kim, Young-In
    • Journal of the Korea Society of Computer and Information
    • /
    • v.11 no.3
    • /
    • pp.73-78
    • /
    • 2006
  • Conventional caption extraction methods use the difference between frames or color segmentation methods from the whole image. Because these methods depend heavily on heuristics, we should have a priori knowledge of the captions to be extracted. Also they are difficult to implement. In this paper, we propose a method that uses little heuristic and simplified algorithm. We use topographical features of characters to extract the character points and use MST(Minimum Spanning Tree) to extract the candidate regions for captions. Character regions are determined by testing several conditions and verifying those candidate regions. Experimental results show that the candidate region extraction rate is 100%, and the character region extraction rate is 98.2%. And then we can see the results that caption area in complex images is well extracted.

  • PDF

패션디자인 DB 개발

  • 김정회
    • Proceedings of the Korea Database Society Conference
    • /
    • 1997.10a
    • /
    • pp.358-375
    • /
    • 1997
  • 가. 패션 디자인 기초 정보 수집/분석 - 국내외에 산재하는 패션디자인 정보의 기초자료를 입수 - 디자이너별/ 컬렉션별/주제별로 분류 - 가공 나- 패션디자인정보의 멀티미디어 DATA BASE개발 - 화상(IMAGE)/해설(TEXT)/ SOUND의 복합 DATA BASE SYSTEM - PC통신망 서비스를 위한 DATA개발 다. 패션디자인 관련자료의 DB화 - 패션디자인 이론서 - 패션디자인 컨테스트 / 이벤트 정보 - 패션디자인 교육기관 정보 - 패션브랜드 정보 (내셔널 / 디자이너 / 수입) 라. DATA BASE 공급 서비스 - PC통신망을 통한 서비스(DOWN LOAD 가능) - 디자인작품 IMAGE 및 CONCEPT/ DETAILS/ CAPTION - PC통신을 이용 디자인 인력 구인/구직정보 활용 - 패션디자인 해외유학 정보 마. Inter-NET 서비스 - Inter-NET을 이용 국내디자이너작품 해외 소개(중략)

  • PDF

Caption Extraction in News Video Sequence using Frequency Characteristic

  • Youglae Bae;Chun, Byung-Tae;Seyoon Jeong
    • Proceedings of the IEEK Conference
    • /
    • 2000.07b
    • /
    • pp.835-838
    • /
    • 2000
  • Popular methods for extracting a text region in video images are in general based on analysis of a whole image such as merge and split method, and comparison of two frames. Thus, they take long computing time due to the use of a whole image. Therefore, this paper suggests the faster method of extracting a text region without processing a whole image. The proposed method uses line sampling methods, FFT and neural networks in order to extract texts in real time. In general, text areas are found in the higher frequency domain, thus, can be characterized using FFT The candidate text areas can be thus found by applying the higher frequency characteristics to neural network. Therefore, the final text area is extracted by verifying the candidate areas. Experimental results show a perfect candidate extraction rate and about 92% text extraction rate. The strength of the proposed algorithm is its simplicity, real-time processing by not processing the entire image, and fast skipping of the images that do not contain a text.

  • PDF