• Title/Summary/Keyword: video character recognition

Search Result 42, Processing Time 0.033 seconds

Study on video character extraction and recognition (비디오 자막 추출 및 인식 기법에 관한 연구)

  • 김종렬;김성섭;문영식
    • Proceedings of the IEEK Conference
    • /
    • 2001.06c
    • /
    • pp.141-144
    • /
    • 2001
  • In this paper, a new algorithm for extracting and recognizing characters from video, without pre-knowledge such as font, color, size of character, is proposed. To improve the recognition rate for videos with complex background at low resolution, continuous frames with identical text region are automatically detected to compose an average frame. Using boundary pixels of a text region as seeds, we apply region filling to remove background from the character Then color clustering is applied to remove remaining backgrounds according to the verification of region filling process. Features such as white run and zero-one transition from the center, are extracted from unknown characters. These feature are compared with a pre-composed character feature set to recognize the characters.

  • PDF

An Implementation of a System for Video Translation on Window Platform Using OCR (윈도우 기반의 광학문자인식을 이용한 영상 번역 시스템 구현)

  • Hwang, Sun-Myung;Yeom, Hee-Gyun
    • Journal of Internet of Things and Convergence
    • /
    • v.5 no.2
    • /
    • pp.15-20
    • /
    • 2019
  • As the machine learning research has developed, the field of translation and image analysis such as optical character recognition has made great progress. However, video translation that combines these two is slower than previous developments. In this paper, we develop an image translator that combines existing OCR technology and translation technology and verify its effectiveness. Before developing, we presented what functions are needed to implement this system and how to implement them, and then tested their performance. With the application program developed through this paper, users can access translation more conveniently, and also can contribute to ensuring the convenience provided in any environment.

Frame Rearrangement Method by Time Information Remarked on Recovered Image (복원된 영상에 표기된 시간 정보에 의한 프레임 재정렬 기법)

  • Kim, Yong Jin;Lee, Jung Hwan;Byun, Jun Seok;Park, Nam In
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.12
    • /
    • pp.1641-1652
    • /
    • 2021
  • To analyze the crime scene, the role of digital evidence such as CCTV and black box is very important. Such digital evidence is often damaged due to device defects or intentional deletion. In this case, the deleted video can be restored by well-known techniques like the frame-based recovery method. Especially, the data such as the video can be generally fragmented and saved in the case of the memory used almost fully. If the fragmented video were recovered in units of images, the sequence of the recovered images may not be continuous. In this paper, we proposed a new video restoration method to match the sequence of recovered images. First, the images are recovered through a frame-based recovery technique. Then, after analyzing the time information marked on the images, the time information was extracted and recognized via optical character recognition (OCR). Finally, the recovered images are rearranged based on the time information obtained by OCR. For performance evaluation, we evaluate the recovery rate of our proposed video restoration method. As a result, it was shown that the recovery rate for the fragmented video was recovered from a minimum of about 47% to a maximum of 98%.

Color Recognition and Phoneme Pattern Segmentation of Hangeul Using Augmented Reality (증강현실을 이용한 한글의 색상 인식과 자소 패턴 분리)

  • Shin, Seong-Yoon;Choi, Byung-Seok;Rhee, Yang-Won
    • Journal of the Korea Society of Computer and Information
    • /
    • v.15 no.6
    • /
    • pp.29-35
    • /
    • 2010
  • While diversification of the use of video in the prevalence of cheap video equipment, augmented reality can print additional real-world images and video image. Although many recent advent augmented reality techniques, currently attempting to correct the character recognition is performed. In this paper characters marked with a visual marker recognition, and the color to match the marker color of the characters finds. And, it was shown on the screen by the character recognition. In this paper, by applying the phoneme pattern segmentation algorithm by the horizontal projection, we propose to segment the phoneme to match the six types of Hangul representation. Throughout the experiment sample of phoneme segmentation using augmented reality showed proceeding result at each step, and the experimental results was found to be that detection rate was above 90%.

Recent Trends in Deep Learning-Based Optical Character Recognition (딥러닝 기반 광학 문자 인식 기술 동향)

  • Min, G.;Lee, A.;Kim, K.S.;Kim, J.E.;Kang, H.S.;Lee, G.H.
    • Electronics and Telecommunications Trends
    • /
    • v.37 no.5
    • /
    • pp.22-32
    • /
    • 2022
  • Optical character recognition is a primary technology required in different fields, including digitizing archival documents, industrial automation, automatic driving, video analytics, medicine, and financial institution, among others. It was created in 1928 using pattern matching, but with the advent of artificial intelligence, it has since evolved into a high-performance character recognition technology. Recently, methods for detecting curved text and characters existing in a complicated background are being studied. Additionally, deep learning models are being developed in a way to recognize texts in various orientations and resolutions, perspective distortion, illumination reflection and partially occluded text, complex font characters, and special characters and artistic text among others. This report reviews the recent deep learning-based text detection and recognition methods and their various applications.

Methods for Video Caption Extraction and Extracted Caption Image Enhancement (영화 비디오 자막 추출 및 추출된 자막 이미지 향상 방법)

  • Kim, So-Myung;Kwak, Sang-Shin;Choi, Yeong-Woo;Chung, Kyu-Sik
    • Journal of KIISE:Software and Applications
    • /
    • v.29 no.4
    • /
    • pp.235-247
    • /
    • 2002
  • For an efficient indexing and retrieval of digital video data, research on video caption extraction and recognition is required. This paper proposes methods for extracting artificial captions from video data and enhancing their image quality for an accurate Hangul and English character recognition. In the proposed methods, we first find locations of beginning and ending frames of the same caption contents and combine those multiple frames in each group by logical operation to remove background noises. During this process an evaluation is performed for detecting the integrated results with different caption images. After the multiple video frames are integrated, four different image enhancement techniques are applied to the image: resolution enhancement, contrast enhancement, stroke-based binarization, and morphological smoothing operations. By applying these operations to the video frames we can even improve the image quality of phonemes with complex strokes. Finding the beginning and ending locations of the frames with the same caption contents can be effectively used for the digital video indexing and browsing. We have tested the proposed methods with the video caption images containing both Hangul and English characters from cinema, and obtained the improved results of the character recognition.

Design of System for Character Recognition and Improvement of the tire side using a Laser Sensor (레이저 센서를 이용한 타이어 옆면 인식 및 개선 시스템 설계)

  • Jang, Hyun-young;Jang, Jong-Wook
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2016.05a
    • /
    • pp.267-270
    • /
    • 2016
  • Currently, tire has a variety of information to know the information of the tire are displayed on the tire side. the market are marked with raised letters showing a variety of information relating to the tires. Such information enables people and tire related companies to distinguish between tires upon the information marked on the tires. Generally, people see the information including max press, manufacturing date, etc. Accordingly, studies on automated recognition of raised letters on tire by using image processing technology have been presented consistently. However, they lack a method for recognition of letters and improvement of the recognition. Moreover, the raised letters have been previously recognized through image processing. Further, to obtain the character recognition of a conventional side in video, it is suitably utilized the effects of lighting time of acquisition, so as part of the background and the character has a gray level values between approximately the same, is the part that is not relatively clear are many scattered. In this paper, we see the characters of the tire side using the laser sensor, recognition, was designed for character recognition of the tire side.

  • PDF

Extraction and Recognition of Character from MPEG-2 news Video Images (MPEG-2 뉴스영상에서 문자영역 추출 및 문자 인식)

  • Park, Yeong-Gyu;Kim, Seong-Guk;Yu, Won-Yeong;Kim, Jun-Cheol;Lee, Jun-Hwan
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.5
    • /
    • pp.1410-1417
    • /
    • 1999
  • In this paper, we propose the method of extracting the caption regions from news video and the method of recognizing the captions that can be used mainly for content-based indexing and retrieving the MPEG-2 compressed news for NOD(News On Demand). The proposed method can reduce the searching time on detecting caption frames with minimum MPEG-2 decoding, and effectively eliminate the noise in caption regions by deliberately devised preprocessing. Because the kind of fonts that are used for captions is not various in the news video, an enhanced template matching method is used for recognizing characters. We could obtain good recognition result in the experiment of sports news video by the proposed methods.

  • PDF

Automatic Poster Generation System Using Protagonist Face Analysis

  • Yeonhwi You;Sungjung Yong;Hyogyeong Park;Seoyoung Lee;Il-Young Moon
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.4
    • /
    • pp.287-293
    • /
    • 2023
  • With the rapid development of domestic and international over-the-top markets, a large amount of video content is being created. As the volume of video content increases, consumers tend to increasingly check data concerning the videos before watching them. To address this demand, video summaries in the form of plot descriptions, thumbnails, posters, and other formats are provided to consumers. This study proposes an approach that automatically generates posters to effectively convey video content while reducing the cost of video summarization. In the automatic generation of posters, face recognition and clustering are used to gather and classify character data, and keyframes from the video are extracted to learn the overall atmosphere of the video. This study used the facial data of the characters and keyframes as training data and employed technologies such as DreamBooth, a text-to-image generation model, to automatically generate video posters. This process significantly reduces the time and cost of video-poster production.

Automatic Indexing for the Content-based Retrieval of News Video (뉴스 비디오의 내용기반 검색을 위한 자동 인덱싱)

  • Yang, Myung-Sup;Yoo, Cheol-Jung;Chang, Ok-Bae
    • The Transactions of the Korea Information Processing Society
    • /
    • v.5 no.5
    • /
    • pp.1130-1139
    • /
    • 1998
  • This paper presents an integrated solution for the content-based news video indexing and the retrieval. Currently, it is impossible to automatically index a general video, but we can index a specific structural video such as news videos. Our proposed model extracts automatically the key frames by using the structured knowledge of news and consists of the news item segmentation, caption recognition and search browser modules. We present above three modules in the following: the news event segmentation module recognizes an anchor-person shot based on face recognition, and then its news event are divided by the anchor-person's frame information. The caption recognition module detects the caption-frames with the caption characteristics, extracts their character region by the using split-merge method, and then recognizes characters with OCR software. Finally, the search browser module could make a various of searching mechanism possible.

  • PDF