• Title/Summary/Keyword: Intelligent Character Recognition

Search Result 62, Processing Time 0.038 seconds

Character Recognition of Vehicle Number Plate using Modular Neural Network (모듈라 신경망을 이용한 자동차 번호판 문자인식)

  • Park, Chang-Seok;Kim, Byeong-Man;Seo, Byung-Hoon;Lee, Kwang-Ho
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.13 no.4
    • /
    • pp.409-415
    • /
    • 2003
  • Recently, the modular learning are very popular and receive much attention for pattern classification. The modular learning method based on the "divide and conquer" strategy can not only solve the complex problems, but also reach a better result than a single classifier′s on the learning quality and speed. In the neural network area, some researches that take the modular learning approach also have been made to improve classification performance. In this paper, we propose a simple modular neural network for characters recognition of vehicle number plate and evaluate its performance on the clustering methods of feature vectors used in constructing subnetworks. We implement two clustering method, one is grouping similar feature vectors by K-means clustering algorithm, the other grouping unsimilar feature vectors by our proposed algorithm. The experiment result shows that our algorithm achieves much better performance.

Design of a Korean Character Vehicle License Plate Recognition System (퍼지 ARTMAP에 의한 한글 차량 번호판 인식 시스템 설계)

  • Xing, Xiong;Choi, Byung-Jae
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.20 no.2
    • /
    • pp.262-266
    • /
    • 2010
  • Recognizing a license plate of a vehicle has widely been issued. In this thesis, firstly, mean shift algorithm is used to filter and segment a color vehicle image in order to get candidate regions. These candidate regions are then analyzed and classified in order to decide whether a candidate region contains a license plate. We then present an approach to recognize a vehicle's license plate using the Fuzzy ARTMAP neural network, a relatively new architecture of the neural network family. We show that the proposed system is well to recognize the license plate and shows some compute simulations.

Detecting Numeric and Character Areas of Low-quality License Plate Images using YOLOv4 Algorithm (YOLOv4 알고리즘을 이용한 저품질 자동차 번호판 영상의 숫자 및 문자영역 검출)

  • Lee, Jeonghwan
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.18 no.4
    • /
    • pp.1-11
    • /
    • 2022
  • Recently, research on license plate recognition, which is a core technology of an intelligent transportation system(ITS), is being actively conducted. In this paper, we propose a method to extract numbers and characters from low-quality license plate images by applying the YOLOv4 algorithm. YOLOv4 is a one-stage object detection method using convolution neural network including BACKBONE, NECK, and HEAD parts. It is a method of detecting objects in real time rather than the previous two-stage object detection method such as the faster R-CNN. In this paper, we studied a method to directly extract number and character regions from low-quality license plate images without additional edge detection and image segmentation processes. In order to evaluate the performance of the proposed method we experimented with 500 license plate images. In this experiment, 350 images were used for training and the remaining 150 images were used for the testing process. Computer simulations show that the mean average precision of detecting number and character regions on vehicle license plates was about 93.8%.

Automatic gasometer reading system using selective optical character recognition (관심 문자열 인식 기술을 이용한 가스계량기 자동 검침 시스템)

  • Lee, Kyohyuk;Kim, Taeyeon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.1-25
    • /
    • 2020
  • In this paper, we suggest an application system architecture which provides accurate, fast and efficient automatic gasometer reading function. The system captures gasometer image using mobile device camera, transmits the image to a cloud server on top of private LTE network, and analyzes the image to extract character information of device ID and gas usage amount by selective optical character recognition based on deep learning technology. In general, there are many types of character in an image and optical character recognition technology extracts all character information in an image. But some applications need to ignore non-of-interest types of character and only have to focus on some specific types of characters. For an example of the application, automatic gasometer reading system only need to extract device ID and gas usage amount character information from gasometer images to send bill to users. Non-of-interest character strings, such as device type, manufacturer, manufacturing date, specification and etc., are not valuable information to the application. Thus, the application have to analyze point of interest region and specific types of characters to extract valuable information only. We adopted CNN (Convolutional Neural Network) based object detection and CRNN (Convolutional Recurrent Neural Network) technology for selective optical character recognition which only analyze point of interest region for selective character information extraction. We build up 3 neural networks for the application system. The first is a convolutional neural network which detects point of interest region of gas usage amount and device ID information character strings, the second is another convolutional neural network which transforms spatial information of point of interest region to spatial sequential feature vectors, and the third is bi-directional long short term memory network which converts spatial sequential information to character strings using time-series analysis mapping from feature vectors to character strings. In this research, point of interest character strings are device ID and gas usage amount. Device ID consists of 12 arabic character strings and gas usage amount consists of 4 ~ 5 arabic character strings. All system components are implemented in Amazon Web Service Cloud with Intel Zeon E5-2686 v4 CPU and NVidia TESLA V100 GPU. The system architecture adopts master-lave processing structure for efficient and fast parallel processing coping with about 700,000 requests per day. Mobile device captures gasometer image and transmits to master process in AWS cloud. Master process runs on Intel Zeon CPU and pushes reading request from mobile device to an input queue with FIFO (First In First Out) structure. Slave process consists of 3 types of deep neural networks which conduct character recognition process and runs on NVidia GPU module. Slave process is always polling the input queue to get recognition request. If there are some requests from master process in the input queue, slave process converts the image in the input queue to device ID character string, gas usage amount character string and position information of the strings, returns the information to output queue, and switch to idle mode to poll the input queue. Master process gets final information form the output queue and delivers the information to the mobile device. We used total 27,120 gasometer images for training, validation and testing of 3 types of deep neural network. 22,985 images were used for training and validation, 4,135 images were used for testing. We randomly splitted 22,985 images with 8:2 ratio for training and validation respectively for each training epoch. 4,135 test image were categorized into 5 types (Normal, noise, reflex, scale and slant). Normal data is clean image data, noise means image with noise signal, relfex means image with light reflection in gasometer region, scale means images with small object size due to long-distance capturing and slant means images which is not horizontally flat. Final character string recognition accuracies for device ID and gas usage amount of normal data are 0.960 and 0.864 respectively.

Geometrical Reorientation of Distorted Road Sign using Projection Transformation for Road Sign Recognition (도로표지판 인식을 위한 사영 변환을 이용한 왜곡된 표지판의 기하교정)

  • Lim, Hee-Chul;Deb, Kaushik;Jo, Kang-Hyun
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.15 no.11
    • /
    • pp.1088-1095
    • /
    • 2009
  • In this paper, we describe the reorientation method of distorted road sign by using projection transformation for improving recognition rate of road sign. RSR (Road Sign Recognition) is one of the most important topics for implementing driver assistance in intelligent transportation systems using pattern recognition and vision technology. The RS (Road Sign) includes direction of road or place name, and intersection for obtaining the road information. We acquire input images from mounted camera on vehicle. However, the road signs are often appeared with rotation, skew, and distortion by perspective camera. In order to obtain the correct road sign overcoming these problems, projection transformation is used to transform from 4 points of image coordinate to 4 points of world coordinate. The 4 vertices points are obtained using the trajectory as the distance from the mass center to the boundary of the object. Then, the candidate areas of road sign are transformed from distorted image by using homography transformation matrix. Internal information of reoriented road signs is segmented with arrow and the corresponding indicated place name. Arrow area is the largest labeled one. Also, the number of group of place names equals to that of arrow heads. Characters of the road sign are segmented by using vertical and horizontal histograms, and each character is recognized by using SAD (Sum of Absolute Difference). From the experiments, the proposed method has shown the higher recognition results than the image without reorientation.

A Hangul Element Separation for the Hand-written Character Recognition (필기체 인식을 위한 한글 자소분리)

  • Baek, Nam-U
    • 한국ITS학회:학술대회논문집
    • /
    • 2004.11a
    • /
    • pp.208-211
    • /
    • 2004
  • 본 연구는 필기체 한글 문자를 인식하기 위하여 한글 문자구조를 6개 기본구조로 분류한다. 각각의 한글 자음과 모음을 7-세크먼트, '/'(Left-Incline), '$\backslash$'(Right-Incline), '-'(Left-Right), '$\mid$'(UP-Down), 'c'(Circle), 'ㄱ'(Right-down), 'ㄴ'(Down-Right) 분리한다. 분리된 7-세크먼트에 대해 한글이 쓰여지는 위치에 따라 8개의 기본구조로 정의하여 세크먼트를 분리하여 레벨화한다. 따라서 본 연구는 문자를 자소(자음과모음)로 하여 7-세크먼트로 분리하는 필기체 자소분리 구조를 제시한다.

  • PDF

Feature Extraction and Recognition of Myanmar Characters Based on Deep Learning (딥러닝 기반 미얀마 문자의 특징 추출 및 인식)

  • Ohnmar, Khin;Lee, Sung-Keun
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.17 no.5
    • /
    • pp.977-984
    • /
    • 2022
  • Recently, with the economic development of Southeast Asia, the use of information devices is widely spreading, and the demand for application services using intelligent character recognition is increasing. This paper discusses deep learning-based feature extraction and recognition of Myanmar, one of the Southeast Asian countries. Myanmar alphabet (33 letters) and Myanmar numerals (10 numbers) are used for feature extraction. In this paper, the number of nine features are extracted and more than three new features are proposed. Extracted features of each characters and numbers are expressed with successful results. In the recognition part, convolutional neural networks are used to assess its execution on character distinction. Its algorithm is implemented on captured image data-sets and its implementation is evaluated. The precision of models on the input data set is 96 % and uses a real-time input image.

Automatic Display of an Additional Explanation on a Keyword Written by a Lecturer for e-Learning Using a Pen Capture Tool on Whiteboard and Two Cameras

  • Nishikimi, Kazuyuki;Yada, Yuuki;Tsuruoka, Shinji;Yoshikawa, Tomohiro;Shinogi, Tsuyoshi
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2003.09a
    • /
    • pp.102-105
    • /
    • 2003
  • "e-Leaning" system is classified by lecture time into two types, that is, "synchronous type" spent the same lecture time between the lecturer and students, and "asynchronous type" spent the different lecture time. The size of image database is huge, and there are some problem on the management of the lecture image database in "asynchronous type" e-Learning system. The one of them is that the time tag for the database management must be added manually at present, and the cost of the addition of the time tag causes a serious problem. To resolve the problem, we will use the character recognition for the characters written by the lecturer on whiteboard, and will add the recognized character as a keyword to the tag of the image database. If the database would have the keyword, we could retrieve the database by the keyword efficiently, and the student could select the interested lecture scene only in the full lecture database.

  • PDF

Implementation of ROS-Based Intelligent Unmanned Delivery Robot System (ROS 기반 지능형 무인 배송 로봇 시스템의 구현)

  • Seong-Jin Kong;Won-Chang Lee
    • Journal of IKEEE
    • /
    • v.27 no.4
    • /
    • pp.610-616
    • /
    • 2023
  • In this paper, we implement an unmanned delivery robot system with Robot Operating System(ROS)-based mobile manipulator, and introduce the technologies employed for the system implementation. The robot consists of a mobile robot capable of autonomous navigation inside the building using an elevator and a Selective Compliance Assembly Robot Arm(SCARA)-Type manipulator equipped with a vacuum pump. The robot can determines the position and orientation for picking up a package through image segmentation and corner detection using the camera on the manipulator. The proposed system has a user interface implemented to check the delivery status and determine the real-time location of the robot through a web server linked to the application and ROS, and recognizes the shipment and address at the delivery station through You Only Look Once(YOLO) and Optical Character Recognition(OCR). The effectiveness of the system is validated through delivery experiments conducted within a 4-story building.

An Elliptic Approach to Learning Discriminabts

  • KARBOU, Fatiha;KARBOU, Fatima;KARBOU, M.
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 1998.06a
    • /
    • pp.143-147
    • /
    • 1998
  • It sis wisely stated that the most valuable knowledge that a person can acquire is the knowledge of how to learn. The human's learning is characterized by the ability to extract relationships between the different characters of a given situation . The ellipse is a first approach of comparison. We assimilate each character to a half axis of the ellipse and the result is a geometrical figure that varies according to values of the two characters. Thus, we take into account the two characters as an alone entity.

  • PDF