• Title/Summary/Keyword: Captions

Search Result 64, Processing Time 0.039 seconds

Size-Independent Caption Extraction for Korean Captions with Edge Connected Components

  • Jung, Je-Hee;Kim, Jaekwang;Lee, Jee-Hyong
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.12 no.4
    • /
    • pp.308-318
    • /
    • 2012
  • Captions include information which relates to the images. In order to obtain the information in the captions, text extraction methods from images have been developed. However, most existing methods can be applied to captions with a fixed height or stroke width using fixed pixel-size or block-size operators which are derived from morphological supposition. We propose an edge connected components based method that can extract Korean captions that are composed of various sizes and fonts. We analyze the properties of edge connected components embedding captions and build a decision tree which discriminates edge connected components which include captions from ones which do not. The images for the experiment are collected from broadcast programs such as documentaries and news programs which include captions with various heights and fonts. We evaluate our proposed method by comparing the performance of the latent caption area extraction. The experiment shows that the proposed method can efficiently extract various sizes of Korean captions.

Connected Component-Based and Size-Independent Caption Extraction with Neural Networks (신경망을 이용한 자막 크기에 무관한 연결 객체 기반의 자막 추출)

  • Jung, Je-Hee;Yoon, Tae-Bok;Kim, Dong-Moon;Lee, Jee-Hyong
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.17 no.7
    • /
    • pp.924-929
    • /
    • 2007
  • Captions which appear in images include information that relates to the images. In order to obtain the information carried by captions, the methods for text extraction from images have been developed. However, most existing methods can be applied to captions with fixed height of stroke's width. We propose a method which can be applied to various caption size. Our method is based on connected components. And then the edge pixels are detected and grouped into connected components. We analyze the properties of connected components and build a neural network which discriminates connected components which include captions from ones which do not. Experimental data is collected from broadcast programs such as news, documentaries, and show programs which include various height caption. Experimental result is evaluated by two criteria : recall and precision. Recall is the ratio of the identified captions in all the captions in images and the precision is the ratio of the captions in the objects identified as captions. The experiment shows that the proposed method can efficiently extract captions various in size.

Control of Time-varying and Nonstationary Stochastic Systems using a Neural Network Controller and Dynamic Bayesian Network Modeling (신경회로망 제어기와 동적 베이시안 네트워크를 이용한 시변 및 비정치 확률시스템의 제어)

  • Cho, Hyun-Cheol;Lee, Jin-Woo;Lee, Young-Jin;Lee, Kwon-Soon
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.17 no.7
    • /
    • pp.930-938
    • /
    • 2007
  • Captions which appear in images include information that relates to the images. In order to obtain the information carried by captions, the methods for text extraction from images have been developed. However, most existing methods can be applied to captions with fixed height of stroke's width. We propose a method which can be applied to various caption size. Our method is based on connected components. And then the edge pixels are detected and grouped into connected components. We analyze the properties of connected components and build a neural network which discriminates connected components which include captions from ones which do not. Experimental data is collected from broadcast programs such as news, documentaries, and show programs which include various height caption. Experimental result is evaluated by two criteria : recall and precision. Recall is the ratio of the identified captions in all the captions in images and the precision is the ratio of the captions in the objects identified as captions. The experiment shows that the proposed method can efficiently extract captions various in size.

A Study on Analyzing Caption Characteristic for Recovering Original Images of Caption Region in TV Scene (원 영상 복원을 위한 TV 자막 특성 분석에 관한 연구)

  • Chun, Byung-Tae
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.10 no.4
    • /
    • pp.177-182
    • /
    • 2010
  • Research on recovering original images from captions has been widely done in a reusability point of view. In usual, dynamic images imported from foreign countries often have captions of foreign languages, so it is necessary to translate them into one's language. For the natural exchange of captions without loss of original images, recovering the images corresponding to captions is necessary. However, though recovering original images is very important, systematic analysis on the characteristics of captions has not been done yet. Therefore, in this paper, we first survey the classification methods of TV programs at academic worlds, broadcasting stations, and broadcasting organizations, and then analyses the frequency of captions, importance of caption contents, and necessity of recovering according to their types. Also, we analyze the characteristics of captions which are significantly recognized to be necessary to recover, and use them as recovering information.

A Novel Approach for Key Caption Detection in Golf Videos Using Color Patterns

  • Jung, Cheol-Kon;Kim, Joong-Kyu
    • ETRI Journal
    • /
    • v.30 no.5
    • /
    • pp.750-752
    • /
    • 2008
  • This paper provides a novel method of detecting key captions containing player information in golf videos. We use the color pattern of captions and its repetition property to determine the key captions. The experimental results show that the proposed method achieves a much higher accuracy than existing methods.

  • PDF

Localization of captions in MPEG compression images based on I frame (I 프레임에 기반한 MPEG 압축영상에서의 자막 탐지)

  • 유태웅
    • Journal of the Korea Computer Industry Society
    • /
    • v.2 no.11
    • /
    • pp.1465-1476
    • /
    • 2001
  • For the applications like video indexing, text understanding, and automatic captions localization system, real-time localization of captions is an essential task. This paper presents a algorithm for localization of captions in MPEG compression images based on I frame. In this algorithm, caption text regions are segmented from background images using their distinguishing texture characteristics and chrominance information. Unlike previously published algorithms which fully decompress the video sequence before extracting the text regions, this algorithm locates candidate caption text region directly in the DCT compressed domain.

  • PDF

The Ordering Effects of Captions on Korean Aviation Students' Self-confidence, Interest, and Motivation in English Learning (영어 자막 활용의 순서가 항공 대학 학생들의 영어 학습에 대한 자신감, 흥미, 및 동기에 미치는 영향)

  • Kim, Na-Young
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.4
    • /
    • pp.362-373
    • /
    • 2020
  • This study examined the ordering effects of caption use on the English learning of aviation students. This study focused on the use of captions and the students' self-confidence, interest, and motivation for English learning. Sixty participants were divided randomly into the Caption 1 group (n = 30) and Caption 2 group (n = 30). During 16 weeks, the Caption 1 group was exposed to captioned media first, and then the same media with no captions was provided. The Caption 2 group was exposed to the non-captioned media first and then provided with captions on the second exposure. A survey was given to assess the attitude of each group to English learning. The survey results showed that all participants revealed moderately positive satisfaction on the use of captions in EFL classes. Regarding the ordering effects of caption use, the findings showed that the first exposure to the captions results in the students' self-confidence and interest. The second exposure to the captions, however, was beneficial for improving the students' motivation. From these findings, several pedagogical implications can be suggested.

Detection of Artificial Caption using Temporal and Spatial Information in Video (시·공간 정보를 이용한 동영상의 인공 캡션 검출)

  • Joo, SungIl;Weon, SunHee;Choi, HyungIl
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.1 no.2
    • /
    • pp.115-126
    • /
    • 2012
  • The artificial captions appearing in videos include information that relates to the videos. In order to obtain the information carried by captions, many methods for caption extraction from videos have been studied. Most traditional methods of detecting caption region have used one frame. However video include not only spatial information but also temporal information. So we propose a method of detection caption region using temporal and spatial information. First, we make improved Text-Appearance-Map and detect continuous candidate regions through matching between candidate-regions. Second, we detect disappearing captions using disappearance test in candidate regions. In case of captions disappear, the caption regions are decided by a merging process which use temporal and spatial information. Final, we decide final caption regions through ANNs using edge direction histograms for verification. Our proposed method was experienced on many kinds of captions with a variety of sizes, shapes, positions and the experiment result was evaluated through Recall and Precision.

A Case Study on Closed Captions: Focusing on on Netflix (넷플릭스 <오징어 게임> 폐쇄자막 연구)

  • Jeong, Sua;Lee, Jimin
    • The Journal of the Convergence on Culture Technology
    • /
    • v.10 no.2
    • /
    • pp.279-285
    • /
    • 2024
  • This study aims to evaluate the accuracy and completeness of Korean and English closed captions for Netflix's "Squid Game" and to present implications based on the findings. To achieve this, the closed captioning guidelines of the U.S. Federal Communications Commission, DCMP, and the Korea Communications Commission were identified and analyzed. The analysis of the subtitle of the entire "Squid Game" series reveals that, while Korean closed captions accurately present slangs and titles, they present non-existent information in speaker identification. In English closed captions, speaker identification guidelines are well followed, but omissions of slangs and title mistranslations are observed. In terms of completeness, both Korean and English closed captions are found to omit certain audio parts. To address these issues, the study suggests strengthening the QA process, establishing a system to communicate original text problems during translation, and utilizing general English subtitles.

A Hangeul Recognition Method Using Directional Edges in Open Captions

  • Jun, Seung-Chul;Kang, Myeong-Gyu;Park, Sung-Han
    • Proceedings of the IEEK Conference
    • /
    • 2002.07b
    • /
    • pp.1157-1160
    • /
    • 2002
  • This paper proposes an efficient method to recognize Hangeul in video open captions. The open captions in news video can play an important role in the video indexing. The strokes of Korean character have a very strong horizontal and vertical directionality and some strokes appear repeatedly in each character. Based on this characteristics, in this paper, we propose an efficient algorithm to extract the character regions in open caption and recognize the characters based on these characteristics of Korean character. The simulation results demonstrate the efficiency of our algorithm in terms of computation time and recognition accuracy.

  • PDF