• Title/Summary/Keyword: Optical character recognition

Search Result 181, Processing Time 0.027 seconds

Artificial intelligence wearable platform that supports the life cycle of the visually impaired (시각장애인의 라이프 사이클을 지원하는 인공지능 웨어러블 플랫폼)

  • Park, Siwoong;Kim, Jeung Eun;Kang, Hyun Seo;Park, Hyoung Jun
    • Journal of Platform Technology
    • /
    • v.8 no.4
    • /
    • pp.20-28
    • /
    • 2020
  • In this paper, a voice, object, and optical character recognition platform including voice recognition-based smart wearable devices, smart devices, and web AI servers was proposed as an appropriate technology to help the visually impaired to live independently by learning the life cycle of the visually impaired in advance. The wearable device for the visually impaired was designed and manufactured with a reverse neckband structure to increase the convenience of wearing and the efficiency of object recognition. And the high-sensitivity small microphone and speaker attached to the wearable device was configured to support the voice recognition interface function consisting of the app of the smart device linked to the wearable device. From experimental results, the voice, object, and optical character recognition service used open source and Google APIs in the web AI server, and it was confirmed that the accuracy of voice, object and optical character recognition of the service platform achieved an average of 90% or more.

  • PDF

Character Segmentation using Side Profile Pattern (측면윤곽 패턴을 이용한 접합 문자 분할 연구)

  • Jung Minchul
    • Journal of Intelligence and Information Systems
    • /
    • v.10 no.3
    • /
    • pp.1-10
    • /
    • 2004
  • In this paper, a new character segmentation algorithm of machine printed character recognition is proposed. The new approach of the proposed character segmentation algorithm overcomes the weak points of both feature-based approaches and recognition-based approaches in character segmentation. This paper defines side profiles of touching characters. The character segmentation algorithm gives a candidate single character in touching characters by side profiles, without any help of character recognizer. It segments touching characters and decides the candidate single character by side profiles. This paper also defines cutting cost, which makes the proposed character segmentation find an optimal segmenting path. The performance of the proposed character segmentation algorithm in this paper has been obtained using a real envelope reader system, which can recognize addresses in U.S. mail pieces and sort the mail pieces. 3359 mail pieces were tested. The improvement was from $68.92\%\;to\;80.08\%$ by the proposed character segmentation.

  • PDF

Synthesis of Multiplexed MACE Filter for Optical Korean Character Recognition (인쇄체 한글의 광학적 인식을 위한 다중 MACE 필터의 합성)

  • 김정우;김철수;배장근;도양회;김수중
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.19 no.12
    • /
    • pp.2364-2375
    • /
    • 1994
  • For the efficient recognition of printed Korean characters, a multiplexed minimum average correlation energy(MMACE) filter is proposed. Proposed method solved the disadvantages of the tree structure algorithm which recognition system is very huge and recognition method is sophisticated. Using only one consonant MMACE filter and one vowel one, we recognized the full Korean character. Each MMACE filter is multiplexed by 4 K-tuple MACE filters which are synthesized by 24 consonants and vowels. Hence the proposed MMACE filter and the correlation distribution plane are divided by 4 subregion. We obtained the binary codes for the Korean character recognition from each correlation distribution subplane. And the obtained codes are compared with the truth table for consonants and vowels in computer. We can recognize the full Korean characters when substitute the corresponded consonant or vowel font of the consistent code to the correlation peak place in the output correlation plane. The computer simulation and optical experiment results show that the proposed compact Korean character recognition system using the MMACE filters has high discrimination capability.

  • PDF

Development an Android based OCR Application for Hangul Food Menu (한글 음식 메뉴 인식을 위한 OCR 기반 어플리케이션 개발)

  • Lee, Gyu-Cheol;Yoo, Jisang
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.21 no.5
    • /
    • pp.951-959
    • /
    • 2017
  • In this paper, we design and implement an Android-based Hangul food menu recognition application that recognizes characters from images captured by a smart phone. Optical Character Recognition (OCR) technology is divided into preprocessing, recognition and post-processing. In the preprocessing process, the characters are extracted using Maximally Stable Extremal Regions (MSER). In recognition process, Tesseract-OCR, a free OCR engine, is used to recognize characters. In the post-processing process, the wrong result is corrected by using the dictionary DB for the food menu. In order to evaluate the performance of the proposed method, experiments were conducted to compare the recognition performance using the actual menu plate as the DB. The recognition rate measurement experiment with OCR Instantly Free, Text Scanner and Text Fairy, which is a character recognizing application in Google Play Store, was conducted. The experimental results show that the proposed method shows an average recognition rate of 14.1% higher than other techniques.

Convolutional Neural Networks for Character-level Classification

  • Ko, Dae-Gun;Song, Su-Han;Kang, Ki-Min;Han, Seong-Wook
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.6 no.1
    • /
    • pp.53-59
    • /
    • 2017
  • Optical character recognition (OCR) automatically recognizes text in an image. OCR is still a challenging problem in computer vision. A successful solution to OCR has important device applications, such as text-to-speech conversion and automatic document classification. In this work, we analyze character recognition performance using the current state-of-the-art deep-learning structures. One is the AlexNet structure, another is the LeNet structure, and the other one is the SPNet structure. For this, we have built our own dataset that contains digits and upper- and lower-case characters. We experiment in the presence of salt-and-pepper noise or Gaussian noise, and report the performance comparison in terms of recognition error. Experimental results indicate by five-fold cross-validation that the SPNet structure (our approach) outperforms AlexNet and LeNet in recognition error.

Automated Bar Placing Model Generation for Augmented Reality Using Recognition of Reinforced Concrete Details (부재 일람표 도면 인식을 활용한 증강현실 배근모델 자동 생성)

  • Park, U-Yeol;An, Sung-Hoon
    • Journal of the Korea Institute of Building Construction
    • /
    • v.20 no.3
    • /
    • pp.289-296
    • /
    • 2020
  • This study suggests a methodology for automatically extracting placing information from 2D reinforced concrete details drawings and generating a 3D reinforcement placing model to develop a mobile augmented reality for bar placing work. To make it easier for users to acquire placing information, it is suggested that users takes pictures of structural drawings using a camera built into a mobile device and extract placing information using vision recognition and the OCR(Optical Character Registration) tool. In addition, an augmented reality app is implemented using the game engine to allow users to automatically generate 3D reinforcement placing model and review the 3D models by superimposing them with real images. Details are described for application to the proposed methodology using the previously developed programming tools, and the results of implementing reinforcement augmented reality models for typical members at construction sites are reviewed. It is expected that the methodology presented as a result of application can be used for learning bar placing work or construction review.

Lightweight Deep Learning Model of Optical Character Recognition for Laundry Management (세탁물 관리를 위한 문자인식 딥러닝 모델 경량화)

  • Im, Seung-Jin;Lee, Sang-Hyeop;Park, Jang-Sik
    • Journal of the Korean Society of Industry Convergence
    • /
    • v.25 no.6_3
    • /
    • pp.1285-1291
    • /
    • 2022
  • In this paper, we propose a low-cost, low-power embedded environment-based deep learning lightweight model for input images to recognize laundry management codes. Laundry franchise companies mainly use barcode recognition-based systems to record laundry consignee information and laundry information for laundry collection management. Conventional laundry collection management systems using barcodes require barcode printing costs, and due to barcode damage and contamination, it is necessary to improve the cost of reprinting the barcode book in its entirety of 1 billion won annually. It is also difficult to do. Recognition performance is improved by applying the VGG model with 7 layers, which is a reduced-transformation of the VGGNet model for number recognition. As a result of the numerical recognition experiment of service parts drawings, the proposed method obtained a significantly improved result over the conventional method with an F1-Score of 0.95.

A Vehicular License Plate Recognition Framework For Skewed Images

  • Arafat, M.Y.;Khairuddin, A.S.M.;Paramesran, R.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.11
    • /
    • pp.5522-5540
    • /
    • 2018
  • Vehicular license plate (LP) recognition system has risen as a significant field of research recently because various explorations are currently being conducted by the researchers to cope with the challenges of LPs which include different illumination and angular situations. This research focused on restricted conditions such as using image of only one vehicle, stationary background, no angular adjustment of the skewed images. A real time vehicular LP recognition scheme is proposed for the skewed images for detection, segmentation and recognition of LP. In this research, a polar co-ordinate transformation procedure is implemented to adjust the skewed vehicular images. Besides that, window scanning procedure is utilized for the candidate localization that is based on the texture characteristics of the image. Then, connected component analysis (CCA) is implemented to the binary image for character segmentation where the pixels get connected in an eight-point neighbourhood process. Finally, optical character recognition is implemented for the recognition of the characters. For measuring the performance of this experiment, 300 skewed images of different illumination conditions with various tilt angles have been tested. The results show that proposed method able to achieve accuracy of 96.3% in localizing, 95.4% in segmenting and 94.2% in recognizing the LPs with an average localization time of 0.52s.

Improvement OCR Algorithm for Efficient Book Catalog RetrievalTechnology (효과적인 도서목록 검색을 위한 개선된 OCR알고리즘에 관한 연구)

  • HeWen, HeWen;Baek, Young-Hyun;Moon, Sung-Ryong
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.47 no.1
    • /
    • pp.152-159
    • /
    • 2010
  • Existing character recognition algorithm recognize characters in simple conditional. It has the disadvantage that recognition rates often drop drastically when input document image has low quality, rotated text, various font or size text because of external noise or data loss. In this paper, proposes the optical character recognition algorithm which using bicubic interpolation method for the catalog retrieval when the input image has rotated text, blurred, various font and size. In this paper, applied optical character recognition algorithm consist of detection and recognition part. Detection part applied roberts and hausdorff distance algorithm for correct detection the catalog of book. Recognition part applied bicubic interpolation to interpolate data loss due to low quality, various font and size text. By the next time, applied rotation for the bicubic interpolation result image to slant proofreading. Experimental results show that proposal method can effectively improve recognition rate 6% and search-time 1.077s process result.

Scene Text Recognition Performance Improvement through an Add-on of an OCR based Classifier (OCR 엔진 기반 분류기 애드온 결합을 통한 이미지 내부 텍스트 인식 성능 향상)

  • Chae, Ho-Yeol;Seok, Ho-Sik
    • Journal of IKEEE
    • /
    • v.24 no.4
    • /
    • pp.1086-1092
    • /
    • 2020
  • An autonomous agent for real world should be able to recognize text in scenes. With the advancement of deep learning, various DNN models have been utilized for transformation, feature extraction, and predictions. However, the existing state-of-the art STR (Scene Text Recognition) engines do not achieve the performance required for real world applications. In this paper, we introduce a performance-improvement method through an add-on composed of an OCR (Optical Character Recognition) engine and a classifier for STR engines. On instances from IC13 and IC15 datasets which a STR engine failed to recognize, our method recognizes 10.92% of unrecognized characters.