• Title/Summary/Keyword: 자막

Search Result 541, Processing Time 0.029 seconds

Closed Caption Synchronization Using Dynamic Programming (동적계획법을 이용한 장애인방송 폐쇄자막 동기화)

  • Oh, Juhyun
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2020.07a
    • /
    • pp.461-464
    • /
    • 2020
  • 지상파 방송에서는 청각장애인을 위해 폐쇄자막(closed caption) 서비스가 제공되고 있다. 현재의 폐쇄자막 방송은 속기사가 실시간으로 방송을 보면서 입력하기 때문에 지연이 있다. 또한 이렇게 입력된 폐쇄자막은 TV 프로그램 영상과 별도로 저장되기 때문에 영상과 그 시작점이 맞지 않는 경우가 대부분이다. 폐쇄자막을 온라인 서비스 등에 제공하고자 할 때 이러한 문제로 인해 영상과의 동기가 맞지 않아 사용이 어렵다. 본 논문에서는 TV 프로그램의 음성을 인식하여 동기화된 텍스트를 추출하고, 이를 기 저장된 폐쇄자막과 정렬하여 동기화하는 방법을 제안한다. 실제 TV 프로그램과 자막에 적용하였을 때 대부분의 음절과 라인에서 동기화가 정확히 이루어짐을 확인하였다.

  • PDF

Subtitle generation using Speech recognition (음성인식기술을 이용한 자막생성 연구)

  • AHN, Chung Hyun;Jang, In Sun
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2016.06a
    • /
    • pp.48-49
    • /
    • 2016
  • 본 논문에서는 동영상, 팟캐스트 오로부터 자막을 생성하여 청각장애인의 미디어 접근권을 향상시키는 음성인식기술을 적용한 자막생성에 대하여 제안한다. 또한 레퍼런스 음성 DB 와 드라마, 팟캐스트 오디오로부터 생성된 자막의 정확도에 대해 평가하였다. 오디오를 이용하여 생성된 자막은 사극의 경우에는 다소 정확도가 낮게 평가되었으나, 전체적으로는 약 80%이상의 정확도를 갖는 것으로 파악되었다.

  • PDF

An Efficient Scheme for Detecting Caption Regions in DCT-Compressed Images (DCT 기반 압축 영상에서 자막 영역을 검출하기 위한 효율적인 방법)

  • 장현성;강경옥
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2002.11a
    • /
    • pp.127-130
    • /
    • 2002
  • 동영상 장면에서 자막은 흔히 중요한 의미 정보를 나타내기 때문에 영상으로부터 자막 영역을 검출하는 것은 동영상에 대한 의미적인 분석 및 색인 등 다양한 범위에서 널리 응용될 수 있다. 본 논문에서는 DCT 기반으로 압축된 영상에서 자막 후보 영역을 검출하기 위한 고속의 방법을 제안한다. 제안하는 방법은 자막 영역에서 나타나는 수평 밝기 값의 교대 패턴을 검출하기 위하여 각 블록 별로 1회의 1-D IDCT 과정을 필요로 한다

  • PDF

Methods for Video Caption Extraction and Extracted Caption Image Enhancement (영화 비디오 자막 추출 및 추출된 자막 이미지 향상 방법)

  • Kim, So-Myung;Kwak, Sang-Shin;Choi, Yeong-Woo;Chung, Kyu-Sik
    • Journal of KIISE:Software and Applications
    • /
    • v.29 no.4
    • /
    • pp.235-247
    • /
    • 2002
  • For an efficient indexing and retrieval of digital video data, research on video caption extraction and recognition is required. This paper proposes methods for extracting artificial captions from video data and enhancing their image quality for an accurate Hangul and English character recognition. In the proposed methods, we first find locations of beginning and ending frames of the same caption contents and combine those multiple frames in each group by logical operation to remove background noises. During this process an evaluation is performed for detecting the integrated results with different caption images. After the multiple video frames are integrated, four different image enhancement techniques are applied to the image: resolution enhancement, contrast enhancement, stroke-based binarization, and morphological smoothing operations. By applying these operations to the video frames we can even improve the image quality of phonemes with complex strokes. Finding the beginning and ending locations of the frames with the same caption contents can be effectively used for the digital video indexing and browsing. We have tested the proposed methods with the video caption images containing both Hangul and English characters from cinema, and obtained the improved results of the character recognition.

Caption Region Extraction of Sports Video Using Multiple Frame Merge (다중 프레임 병합을 이용한 스포츠 비디오 자막 영역 추출)

  • 강오형;황대훈;이양원
    • Journal of Korea Multimedia Society
    • /
    • v.7 no.4
    • /
    • pp.467-473
    • /
    • 2004
  • Caption in video plays an important role that delivers video content. Existing caption region extraction methods are difficult to extract caption region from background because they are sensitive to noise. This paper proposes the method to extract caption region in sports video using multiple frame merge and MBR(Minimum Bounding Rectangles). As preprocessing, adaptive threshold can be extracted using contrast stretching and Othu Method. Caption frame interval is extracted by multiple frame merge and caption region is efficiently extracted by median filtering, morphological dilation, region labeling, candidate character region filtering, and MBR extraction.

  • PDF

Extraction of Superimposed-Caption Frame Scopes and Its Regions for Analyzing Digital Video (비디오 분석을 위한 자막프레임구간과 자막영역 추출)

  • Lim, Moon-Cheol;Kim, Woo-Saeng
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.11
    • /
    • pp.3333-3340
    • /
    • 2000
  • Recently, Requnremeni for video data have been increased rapidly by high progress of both hardware and cornpression technique. Because digital video data are unformed and mass capacity, it needs various retrieval techniquesjust as contednt-based rehieval Superimposed-caption ina digital video can help us to analyze the video story easier and be used as indexing information for many retrieval techniques In this research we propose a new method that segments the caption as analyzing texture eature of caption regions in each video frame, and that extracts the accurate scope of superimposed-caption frame and its key regions and color by measunng cominuity of caption regions between frames

  • PDF

A Method for Text Detection and Enhancement using Spatio-Temporal Information (시공간 정보를 이용한 자막 탐지 및 향상 기법)

  • Jeong, Jong-Myeon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.8
    • /
    • pp.43-50
    • /
    • 2009
  • Text information in a digital video provides crucial information to acquire semantic information of the video. In the proposed method. text candidate regions are extracted from input sequence by using characteristics of stroke and text candidate regions are localized by using projection to produce text bounding boxes. Bounding boxes containing text regions are verified geometrically and each bounding box existing same location is tracked by calculating matching measure. which is defined as the mean of absolute difference between bounding boxes in the current frame and previous frames. Finally. text regions are enhanced using temporal redundancy of bounding boxes to produce final results. Experimental results for various videos show the validity of the proposed method.

A Study on Efficient Positioning of Subtitles in 360 VR (360 VR 영상에서 효율적인 자막 위치 선정에 관한 연구)

  • Kim, Hyeong-Gyun
    • Journal of Digital Convergence
    • /
    • v.18 no.6
    • /
    • pp.93-98
    • /
    • 2020
  • In this paper, we proposed a technique in which subtitles are followed according to changes in the user's viewpoint in 360 VR. Create a Sphere object in Unity's Scene and insert a 360-degree image on the surface of the Sphere object. At this time, the ReverseNormals script is used to convert the viewpoint to the inside. The SightOrbitproved script is used to modify the camera view. Use this script to set the environment in which subtitles can move depending on the viewpoint. Next, add the 3D text (subtitle) that the user wants to the lower layer of the main camera and build a 360 VR object. The 3D text subtitles implemented through this study were compared according to the change of the user's viewpoint. As a result, as the viewpoint changes, normal subtitles flow out of line of sight according to the user's point of view, but 3D Text subtitles move according to the user's point of view, and it can be seen that the user can always view the subtitles.

Creation of Soccer Video Highlight Using The Structural Features of Caption (자막의 구조적 특징을 이용한 축구 비디오 하이라이트 생성)

  • Huh, Moon-Haeng;Shin, Seong-Yoon;Lee, Yang-Weon;Ryu, Keun-Ho
    • The KIPS Transactions:PartD
    • /
    • v.10D no.4
    • /
    • pp.671-678
    • /
    • 2003
  • A digital video is usually very long temporally. requiring large storage capacity. Therefore, users want to watch pre-summarized video before they watch a large long video. Especially in the field of sports video, they want to watch a highlight video. Consequently, highlight video is used that the viewers decide whether it is valuable for them to watch the video or not. This paper proposes how to create soccer video highlight using the structural features of the caption such as temporal and spatial features. Caption frame intervals and caption key frames are extracted by using those structural features. And then, highlight video is created by using scene relocation, logical indexing and highlight creation rule. Finally. retrieval and browsing of highlight and video segment is performed by selection of item on browser.

Characteristics of Entertainment Program Subtitles and Effects on the Audience's Perception : Text Analysis of JTBC (예능프로그램 자막의 특성과 수용자 인식에 미치는 영향 : JTBC <비정상회담> 텍스트 분석)

  • Kim, Ho-Kyung;Kwon, Ki-Seok;Seo, Sang-Ho
    • The Journal of the Korea Contents Association
    • /
    • v.16 no.3
    • /
    • pp.232-246
    • /
    • 2016
  • Subtitles in entertainment programs have extended the roles from providing additional explanation and complementary sound effect to maximizing amusement and impression as the type which triggers the audience's interest. This study examined the characteristics of the subtitles and the effects on the audiences, mainly focused on of JTBC. Based on the result of the content analysis, the interested subtitles are continuously presented in each episode (10~20%) compared to the proportion of the total captions. Producers have repeatedly used the nickname of the cast members to construct and reinforce the character of them. The text analysis shows that the audience's perception about the cast is mainly influenced by the subtitles. The producers intentionally use the subtitles and subjectively intervene in the entertainment programs. The subtitles have a major effect on the audience's perception and image formation. Producers should be more prudent about their use of subtitles and viewers, on the other hand, must interpret the meaning of subtitles, as active information consumers.