• Title/Summary/Keyword: 문자 인식 기술

Search Result 297, Processing Time 0.046 seconds

The wavelet Transform as a Preprocessing for Character Recognition (웨이브릿변환을 이용한 문자인식 전처리 기술에 관한연구)

  • Choi, Hwan-Soo;Kong, Seong-pil
    • Proceedings of the KIEE Conference
    • /
    • 1997.11a
    • /
    • pp.405-407
    • /
    • 1997
  • 본 논문은 자동차 번호판 용도문자를 인식하개 위한 전처리 과정으로써 웨이브릿 변환을 적용한 연구에 관해 기술한다. 웨이브릿 변환에 의하여 여과된 고주파 대역의 영상은 수평방향, 수직방향, 대각선 방향의 윤관석 형태로 세 개의 대역에 존재하게 되는데, 대상영상이 고주파 대역의 에너지량이 적게 나타나는 반면에 저주파 대역의 에너지량은 크므로 용도문자의 인식 과정에서 저주파 대역 부분만을 이용하였다. 저주파 대역에서 $20{\times}20$크기의 영상을 추출하고 영상을 정규화 하여 오츠알고리즘을 통한 이치화 과정을 거친 다음 역전파 신경망으로 인식함으로써 기존의 단순축소 방법보다 향상된 결과를 실험을 통하여 확인할 수 있었다.

  • PDF

Software Architecture for Embedded OCR System Development (임베디드 OCR시스템 개발을 위한 소프트웨어 아키텍쳐)

  • Kim, Se-Ho;Pack, Jae-Hwa
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2005.11b
    • /
    • pp.862-864
    • /
    • 2005
  • 최근 임베디드 환경에서는 정보 처리를 위한 문자 인식 기술이 많이 요구되고 있다. 하지만 임베디드 환경에서의 문자인식 시스템(Opticai Character Recognition)은 제약적인 자원으로 인하여 플랫폼에 크게 의존하는 문제점을 안고 있어 재사용성을 기대하기 힘들다. 그렇지만 임베디드 환경에서 플랫폼에 독립적인 즉, 재사용이 가능한 모범적인 소프트웨어 아키텍쳐는 없다. 따라서, 본 논문에서는 임베디드 환경에서의 문자 인식 시스템 개발시 플랫폼에 독립적인 즉, 재사용이 가능한 소프트웨어 아키텍쳐를 제안하였다. 또한 제안한 아키텍쳐를 바탕으로 실제 임베디드 환경(WIPI, Qt)에 문자인식 시스템에 적용시켜보았으며, 더 이상 플랫폼에 의존적이지 않음을 확인 해 볼 수 있다.

  • PDF

$\emph{A Priori}$ and the Local Font Classification (연역적이고 국부적인 영문자의 폰트 분류법)

  • 정민철
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.3 no.4
    • /
    • pp.245-250
    • /
    • 2002
  • This paper presents a priori and the local font classification method. The font classification uses ascenders, descenders, and serifs extracted from a word image. The gradient features of those sub-images are extracted, and used as an input to a neural network classifier to produce font classification results. The font classification determines 2-font styles (upright or slant), 3-font groups (serif, sans serif, or typewriter), and 7-font names (PostScript fonts such as Avant Garde, Helvetica, Bookman, New Century Schoolbook, Palatino, Times, or Courier). The proposed a priori and local font classification method allows an OCR system consisting of various font-specific character segmentation tools and various mono-font character recognizers.

  • PDF

The Algorithm Implementation on The Recognition - Technology Through the use of ID Verification (인식기술을 이용한 신원확인 알고리즘 구현)

  • Bang, Gul-Won;Kim, Byung-Ki;Cho, Wan-Hyun
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2002.04b
    • /
    • pp.1247-1250
    • /
    • 2002
  • 본 논문은 일반적인 신원확인 방법인 주민등록증의 사진과 본인 실물을 육안으로 비교로 판별하던 것을 컴퓨터를 이용하여 자동으로 판별할 수 있게 하는 알고리즘을 제안한다. 생체이식 기술은 이용하여 본인 확인하는 방법은 점차 보편화 있다, 이런 생체인식기술 즉 지문인식과 문자인식, 홀로그램 인식기술을 접목하여 주민등록상의 지문이미지와 지문입력기에서 입력받은 생체지문을 비교판별하고 문자이미지를 데이터화하여 주민등록번호가 민원인의 Index Key 되며 홀로그램으로 주민등록증의 진위여부를 판별하는 방법을 제공한다.

  • PDF

Recognition of Container Identifier using Color Information and Contour Following (컬러 정보와 윤곽선 추적을 이용한 컨테이너 식별자 인식)

  • Kim Pyeoung-Kee
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.11 no.3
    • /
    • pp.40-46
    • /
    • 2006
  • Automatic recognition of container identifier is one of key factor to implement port automation and increase distribution throughput. In this paper, I propose a method of container identifier recognition on various input images using color based edge detection and character verification algorithm, I tested the proposed method on 350 container images and it showed good results.

  • PDF

A Method for Thresholding and Correction of Skew in Camera Document Images (카메라 문서 영상의 이진화 및 기울어짐 보정 방법)

  • Jang Dae-Geun;Chun Byung-Tae
    • Journal of the Korea Society of Computer and Information
    • /
    • v.10 no.3 s.35
    • /
    • pp.143-150
    • /
    • 2005
  • Camera image is very sensitive to illumination that result in difficulties for recognizing character. Also Camera captured document images have not only skew but also vignetting effect and geometric distortion. Vignetting effect make it difficult to separate characters from the document images. Geometric distortion, occurred by the mismatch of angle and center position between the document image and the camera, make the shape of characters to be distorted, so that the character recognition is more difficult than the case of using scanner. In this paper, we propose a method that can increase the performance of character recognition by correcting the geometric distortion of document images using a linear approximation which changes the quadrilateral region to the rectangle one. The proposed method also determine the quadrilateral transform region automatically, using the alignment of character lines and the skewed angles of characters located in the edges of each character line. Proposed method, therefore, can correct the geometric distortion without getting positional information from camera.

  • PDF

Text Region Detection Method Using Table Border Pseudo Label (표의 테두리 유사 라벨을 활용한 문자 영역 검출 방법)

  • Han, Jeong Hoon;Park, Se Jin;Moon, Young Shik
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.10
    • /
    • pp.1271-1279
    • /
    • 2020
  • Text region detection is a technology that detects text area in handwriting or printed documents. The detected text areas are digitized through a recognition step, which is used in various fields depending on the purpose of use. However, the detection result of the small text unit is not suitable for the industrial field. In addition, the border of tables in the document that it causes miss-detected results, which has an adverse effect on the recognition step. To solve the issues, we propose a method for detecting text region using the border information of the table. In order to utilize the border information of the table, the proposed method adjusts the flow of two decoders. Experimentally, we show improved performance using the table border pseudo label based on weak supervised learning.

A Study on the Vehicle License Plate Recognition Using Convolutional Neural Networks(CNNs) (CNN 기법을 이용한 자동차 번호판 인식법 연구)

  • Nkundwanayo Seth;Gyoo-Soo Chae
    • Journal of Advanced Technology Convergence
    • /
    • v.2 no.4
    • /
    • pp.7-11
    • /
    • 2023
  • In this study, we presented a method to recognize vehicle license plates using CNN techniques. A vehicle plate is normally used for the official identification purposes by the authorities. Most regular Optical Character Recognition (OCR) techniques perform well in recognizing printed characters on documents but cannot make out the registration number on the number plates. Besides, the existing approaches to plate number detection require that the vehicle is stationary and not in motion. To address these challenges to number plate detection we make the following contributions. We create a database of captured vehicle number plate's images and recognize the number plate character using Convolutional Neural Networks. The results of this study can be usefully used in parking management systems and enforcement cameras.

Text Cues-based Image Matching Method for Navigation (네비게이션을 위한 문자영상기반의 영상매칭 방법)

  • Park, An-Jin;Jung, Kee-Chul
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2005.11b
    • /
    • pp.631-633
    • /
    • 2005
  • 유비쿼터스 시대가 다가오면서, 많은 사람들은 모르는 장소에서 자신의 위치와 목적지까지의 경로에 대한 정보를 알고 싶어할 것이다. 기존의 네비게이션(navigation)을 위한 비전기술은 고차원과 저차원 특징값을 이용하였다. 텍스춰 정보, 색상 히스토그램과 같은 저차원 특징값은 영상의 특징을 정확하게 표현하기 어려우며, 마커와 같은 고차원 정보는 실험환경을 구축하는데 어려움이 있다. 우리는 기존 저/고차원의 특징값 대신, 영상의 특징을 표현하고 인덱싱(indexing)하기 위한 유용한 정보를 많이 포함하고 있으며, 실제환경에서 널리 분포되어있는 중차원 특징값인 문자영상을 이용한다. 문자영상추출은 MLP(Multi-layer perceptron)와 CAMShift알고리즘을 결합한 방법을 이용하며, 서로 다른 장소지만 같은 문자를 가진 곳에서 인식을 수행하기 위해 문자영상의 크기와 기울기를 기반으로 한 영상 검색공간을 대상으로 영상매칭을 수행한다. 실험에서 문자영상을 포함하는 직사각형 검색공간으로 인해 다양한 크기와 기울기에서 높은 인식률을 보이며, 간단한 계산으로 빠른 수행시간을 가진다.

  • PDF

Automatic gasometer reading system using selective optical character recognition (관심 문자열 인식 기술을 이용한 가스계량기 자동 검침 시스템)

  • Lee, Kyohyuk;Kim, Taeyeon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.1-25
    • /
    • 2020
  • In this paper, we suggest an application system architecture which provides accurate, fast and efficient automatic gasometer reading function. The system captures gasometer image using mobile device camera, transmits the image to a cloud server on top of private LTE network, and analyzes the image to extract character information of device ID and gas usage amount by selective optical character recognition based on deep learning technology. In general, there are many types of character in an image and optical character recognition technology extracts all character information in an image. But some applications need to ignore non-of-interest types of character and only have to focus on some specific types of characters. For an example of the application, automatic gasometer reading system only need to extract device ID and gas usage amount character information from gasometer images to send bill to users. Non-of-interest character strings, such as device type, manufacturer, manufacturing date, specification and etc., are not valuable information to the application. Thus, the application have to analyze point of interest region and specific types of characters to extract valuable information only. We adopted CNN (Convolutional Neural Network) based object detection and CRNN (Convolutional Recurrent Neural Network) technology for selective optical character recognition which only analyze point of interest region for selective character information extraction. We build up 3 neural networks for the application system. The first is a convolutional neural network which detects point of interest region of gas usage amount and device ID information character strings, the second is another convolutional neural network which transforms spatial information of point of interest region to spatial sequential feature vectors, and the third is bi-directional long short term memory network which converts spatial sequential information to character strings using time-series analysis mapping from feature vectors to character strings. In this research, point of interest character strings are device ID and gas usage amount. Device ID consists of 12 arabic character strings and gas usage amount consists of 4 ~ 5 arabic character strings. All system components are implemented in Amazon Web Service Cloud with Intel Zeon E5-2686 v4 CPU and NVidia TESLA V100 GPU. The system architecture adopts master-lave processing structure for efficient and fast parallel processing coping with about 700,000 requests per day. Mobile device captures gasometer image and transmits to master process in AWS cloud. Master process runs on Intel Zeon CPU and pushes reading request from mobile device to an input queue with FIFO (First In First Out) structure. Slave process consists of 3 types of deep neural networks which conduct character recognition process and runs on NVidia GPU module. Slave process is always polling the input queue to get recognition request. If there are some requests from master process in the input queue, slave process converts the image in the input queue to device ID character string, gas usage amount character string and position information of the strings, returns the information to output queue, and switch to idle mode to poll the input queue. Master process gets final information form the output queue and delivers the information to the mobile device. We used total 27,120 gasometer images for training, validation and testing of 3 types of deep neural network. 22,985 images were used for training and validation, 4,135 images were used for testing. We randomly splitted 22,985 images with 8:2 ratio for training and validation respectively for each training epoch. 4,135 test image were categorized into 5 types (Normal, noise, reflex, scale and slant). Normal data is clean image data, noise means image with noise signal, relfex means image with light reflection in gasometer region, scale means images with small object size due to long-distance capturing and slant means images which is not horizontally flat. Final character string recognition accuracies for device ID and gas usage amount of normal data are 0.960 and 0.864 respectively.