• Title/Summary/Keyword: Arabic Characters Recognition

Search Result 6, Processing Time 0.029 seconds

Enhanced technique for Arabic handwriting recognition using deep belief network and a morphological algorithm for solving ligature segmentation

  • Essa, Nada;El-Daydamony, Eman;Mohamed, Ahmed Atwan
    • ETRI Journal
    • /
    • v.40 no.6
    • /
    • pp.774-787
    • /
    • 2018
  • Arabic handwriting segmentation and recognition is an area of research that has not yet been fully understood. Dealing with Arabic ligature segmentation, where the Arabic characters are connected and unconstrained naturally, is one of the fundamental problems when dealing with the Arabic script. Arabic character-recognition techniques consider ligatures as new classes in addition to the classes of the Arabic characters. This paper introduces an enhanced technique for Arabic handwriting recognition using the deep belief network (DBN) and a new morphological algorithm for ligature segmentation. There are two main stages for the implementation of this technique. The first stage involves an enhanced technique of the Sari segmentation algorithm, where a new ligature segmentation algorithm is developed. The second stage involves the Arabic character recognition using DBNs and support vector machines (SVMs). The two stages are tested on the IFN/ENIT and HACDB databases, and the results obtained proved the effectiveness of the proposed algorithm compared with other existing systems.

A study on the Recognition of Hand-written Characters and Arabic numbers by Neural Networks (신경회로망을 이용한 필기체 한글 자모음 및 숫자인식에 관한 연구)

  • Oh, Dong-Su;Lee, Eun-Un;Yoo, Jae-Guen;Nam, Moon-Hyun
    • Proceedings of the KIEE Conference
    • /
    • 1991.07a
    • /
    • pp.900-904
    • /
    • 1991
  • In this paper, our study for the recognition of Hand-written Korean characters, Arabic numbers and alphabets by neural netwoks. This System extracts feature of character by using the MESH feature point of handwritten character, Arabic numbers and alphabets. To reduce the input image data, features are extracted from each input images. A MLP(multi-layer perceptron) with one hidden layer was trained with a modified BEP(back error propagation) algorithm. This method extracts feature sets of the characters directly from the scanner and can enhance computation speed without using the special preprocesses such as size normalization, smoothing, and thinning.

  • PDF

Arabic Words Extraction and Character Recognition from Picturesque Image Macros with Enhanced VGG-16 based Model Functionality Using Neural Networks

  • Ayed Ahmad Hamdan Al-Radaideh;Mohd Shafry bin Mohd Rahim;Wad Ghaban;Majdi Bsoul;Shahid Kamal;Naveed Abbas
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.7
    • /
    • pp.1807-1822
    • /
    • 2023
  • Innovation and rapid increased functionality in user friendly smartphones has encouraged shutterbugs to have picturesque image macros while in work environment or during travel. Formal signboards are placed with marketing objectives and are enriched with text for attracting people. Extracting and recognition of the text from natural images is an emerging research issue and needs consideration. When compared to conventional optical character recognition (OCR), the complex background, implicit noise, lighting, and orientation of these scenic text photos make this problem more difficult. Arabic language text scene extraction and recognition adds a number of complications and difficulties. The method described in this paper uses a two-phase methodology to extract Arabic text and word boundaries awareness from scenic images with varying text orientations. The first stage uses a convolution autoencoder, and the second uses Arabic Character Segmentation (ACS), which is followed by traditional two-layer neural networks for recognition. This study presents the way that how can an Arabic training and synthetic dataset be created for exemplify the superimposed text in different scene images. For this purpose a dataset of size 10K of cropped images has been created in the detection phase wherein Arabic text was found and 127k Arabic character dataset for the recognition phase. The phase-1 labels were generated from an Arabic corpus of quotes and sentences, which consists of 15kquotes and sentences. This study ensures that Arabic Word Awareness Region Detection (AWARD) approach with high flexibility in identifying complex Arabic text scene images, such as texts that are arbitrarily oriented, curved, or deformed, is used to detect these texts. Our research after experimentations shows that the system has a 91.8% word segmentation accuracy and a 94.2% character recognition accuracy. We believe in the future that the researchers will excel in the field of image processing while treating text images to improve or reduce noise by processing scene images in any language by enhancing the functionality of VGG-16 based model using Neural Networks.

Development of character recognition system for the mixed font style in the steel processing material

  • Lee, Jong-Hak;Park, Sang-Gug;Park, Soo-Young
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.1431-1434
    • /
    • 2005
  • In the steel production line, the molten metal of a furnace is transformed into billet and then moves to the heating furnace of the hot rolling mill. This paper describes about the development of recognition system for the characters, which was marked at the billet material by use template-marking plate and hand written method, in the steel plant. For the recognition of template-marked characters, we propose PSVM algorithm. And for the recognition of hand written character, we propose combination methods of CCD algorithm and PSVM algorithm. The PSVM algorithm need some more time than the conventional KLT or SVM algorithm. The CCD algorithm makes shorter classification time than the PSVM algorithm and good for the classification of closed curve characters from Arabic numerals. For the confirmation of algorithm, we have compared our algorithm with conventional methods such as KLT classifier and one-to-one SVM. The recognition rate of experimented billet characters shows that the proposing PSVM algorithm is 97 % for the template-marked characters and combinational algorithm of CCD & PSVM is 95.5 % for the hand written characters. The experimental results show that our proposing method has higher recognition rate than that of the conventional methods for the template-marked characters and hand written characters. By using our algorithm, we have installed real time character recognition system at the billet processing line of the steel-iron plant.

  • PDF

The FE-MCBP for Recognition of the Tilted New-Type Vehicle License Plate (기울어진 신규차량번호판 인식을 위한 FE-MCBP)

  • Koo, Gun-Seo
    • Journal of the Korea Society of Computer and Information
    • /
    • v.12 no.5
    • /
    • pp.73-81
    • /
    • 2007
  • This paper presents how to recognize the new-type vehicle license plate using multi-link recognizer after extract the features from characters. In order to assist this task, this paper proposed FE-MCBP to recognize each character that got through image preprocess, extract range of vehicle license plate and extract process of each character. FE-MCBP is the recognizer based on the features of the character, The recognizer is employed to identify the new-type vehicle licence plates which have both the hangul and the arabic numeral characters. And its recognition rate is improved 9.7 percent than the back propagation recognizer before. Also it makes use of extract of linear component and region coordinate generation technology to normalize a image of the tilted vehicle license plate. The recognition system of the new-type vehicle license plate make possible recognize a image of the tilted vehicle license plate when using this system. Also, this system can recognize the tilted or imperfect vehicle licence plates.

  • PDF

Automatic gasometer reading system using selective optical character recognition (관심 문자열 인식 기술을 이용한 가스계량기 자동 검침 시스템)

  • Lee, Kyohyuk;Kim, Taeyeon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.1-25
    • /
    • 2020
  • In this paper, we suggest an application system architecture which provides accurate, fast and efficient automatic gasometer reading function. The system captures gasometer image using mobile device camera, transmits the image to a cloud server on top of private LTE network, and analyzes the image to extract character information of device ID and gas usage amount by selective optical character recognition based on deep learning technology. In general, there are many types of character in an image and optical character recognition technology extracts all character information in an image. But some applications need to ignore non-of-interest types of character and only have to focus on some specific types of characters. For an example of the application, automatic gasometer reading system only need to extract device ID and gas usage amount character information from gasometer images to send bill to users. Non-of-interest character strings, such as device type, manufacturer, manufacturing date, specification and etc., are not valuable information to the application. Thus, the application have to analyze point of interest region and specific types of characters to extract valuable information only. We adopted CNN (Convolutional Neural Network) based object detection and CRNN (Convolutional Recurrent Neural Network) technology for selective optical character recognition which only analyze point of interest region for selective character information extraction. We build up 3 neural networks for the application system. The first is a convolutional neural network which detects point of interest region of gas usage amount and device ID information character strings, the second is another convolutional neural network which transforms spatial information of point of interest region to spatial sequential feature vectors, and the third is bi-directional long short term memory network which converts spatial sequential information to character strings using time-series analysis mapping from feature vectors to character strings. In this research, point of interest character strings are device ID and gas usage amount. Device ID consists of 12 arabic character strings and gas usage amount consists of 4 ~ 5 arabic character strings. All system components are implemented in Amazon Web Service Cloud with Intel Zeon E5-2686 v4 CPU and NVidia TESLA V100 GPU. The system architecture adopts master-lave processing structure for efficient and fast parallel processing coping with about 700,000 requests per day. Mobile device captures gasometer image and transmits to master process in AWS cloud. Master process runs on Intel Zeon CPU and pushes reading request from mobile device to an input queue with FIFO (First In First Out) structure. Slave process consists of 3 types of deep neural networks which conduct character recognition process and runs on NVidia GPU module. Slave process is always polling the input queue to get recognition request. If there are some requests from master process in the input queue, slave process converts the image in the input queue to device ID character string, gas usage amount character string and position information of the strings, returns the information to output queue, and switch to idle mode to poll the input queue. Master process gets final information form the output queue and delivers the information to the mobile device. We used total 27,120 gasometer images for training, validation and testing of 3 types of deep neural network. 22,985 images were used for training and validation, 4,135 images were used for testing. We randomly splitted 22,985 images with 8:2 ratio for training and validation respectively for each training epoch. 4,135 test image were categorized into 5 types (Normal, noise, reflex, scale and slant). Normal data is clean image data, noise means image with noise signal, relfex means image with light reflection in gasometer region, scale means images with small object size due to long-distance capturing and slant means images which is not horizontally flat. Final character string recognition accuracies for device ID and gas usage amount of normal data are 0.960 and 0.864 respectively.