• Title/Summary/Keyword: mobile camera module

Search Result 117, Processing Time 0.024 seconds

A Robust Staff Line Height and Staff Line Space Estimation for the Preprocessing of Music Score Recognition (악보인식 전처리를 위한 강건한 오선 두께와 간격 추정 방법)

  • Na, In-Seop;Kim, Soo-Hyung;Nquyen, Trung Quy
    • Journal of Internet Computing and Services
    • /
    • v.16 no.1
    • /
    • pp.29-37
    • /
    • 2015
  • In this paper, we propose a robust pre-processing module for camera-based Optical Music Score Recognition (OMR) on mobile device. The captured images likely suffer for recognition from many distortions such as illumination, blur, low resolution, etc. Especially, the complex background music sheets recognition are difficult. Through any symbol recognition system, the staff line height and staff line space are used many times and have a big impact on recognition module. A robust and accurate staff line height and staff line space are essential. Some staff line height and staff line space are proposed for binary image. But in case of complex background music sheet image, the binarization results from common binarization algorithm are not satisfactory. It can cause incorrect staff line height and staff line space estimation. We propose a robust staff line height and staff line space estimation by using run-length encoding technique on edge image. Proposed method is composed of two steps, first step, we conducted the staff line height and staff line space estimation based on edge image using by Sobel operator on image blocks. Each column of edge image is encoded by run-length encoding algorithm Second step, we detect the staff line using by Stable Path algorithm and removal the staff line using by adaptive Line Track Height algorithm which is to track the staff lines positions. The result has shown that robust and accurate estimation is possible even in complex background cases.

Smart Camera Technology to Support High Speed Video Processing in Vehicular Network (차량 네트워크에서 고속 영상처리 기반 스마트 카메라 기술)

  • Son, Sanghyun;Kim, Taewook;Jeon, Yongsu;Baek, Yunju
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.40 no.1
    • /
    • pp.152-164
    • /
    • 2015
  • A rapid development of semiconductors, sensors and mobile network technologies has enable that the embedded device includes high sensitivity sensors, wireless communication modules and a video processing module for vehicular environment, and many researchers have been actively studying the smart car technology combined on the high performance embedded devices. The vehicle is increased as the development of society, and the risk of accidents is increasing gradually. Thus, the advanced driver assistance system providing the vehicular status and the surrounding environment of the vehicle to the driver using various sensor data is actively studied. In this paper, we design and implement the smart vehicular camera device providing the V2X communication and gathering environment information. And we studied the method to create the metadata from a received video data and sensor data using video analysis algorithm. In addition, we invent S-ROI, D-ROI methods that set a region of interest in a video frame to improve calculation performance. We performed the performance evaluation for two ROI methods. As the result, we confirmed the video processing speed that S-ROI is 3.0 times and D-ROI is 4.8 times better than a full frame analysis.

Development of Realtime Multimedia Streaming Service using Mobile Smart Devices (모바일 스마트 단말을 활용한 실시간 멀티미디어 스트리밍 서비스 개발)

  • Park, Mi-Ryong;Sim, Han-Eug
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.14 no.4
    • /
    • pp.51-56
    • /
    • 2014
  • Thesedays, there are many smart device applications developed, especially on the using various sensors included in the smart device. Smart devices have several sensors which are camera, GPS, mike, and communication module for collecting ubiquitous environment, and many applications are developed by using such sensors. In this paper, we developed the multimedia stream architecture and examined the smart device applications based on open source with front and back-end server clouds for developing the conceptual architecture. Also, we examined the back-end distributed servers, realtime multimedia stream transferring, multi-media store, and media relay for other server and smart devices. We test the examined architecture on the real target environment to collect the SIP initial setup time, media stream delay, and end-to-end play time. The test results show that there have good network operation environment to provide realtime multimedia services, and we need to improve the end-to-end play time by minimizing the initial setup time.

A Study On Low-cost LPR(License Plate Recognition) System Based On Smart Cam System using Android (안드로이드 기반 스마트 캠 방식의 저가형 자동차 번호판 인식 시스템 구현에 관한 연구)

  • Lee, Hee-Yeol;Lee, Seung-Ho
    • Journal of IKEEE
    • /
    • v.18 no.4
    • /
    • pp.471-477
    • /
    • 2014
  • In this paper, we propose a low-cost license plate recognition system based on smart cam system using Android. The proposed system consists of a portable device and server. Potable device Hardware consists of ARM Cortex-A9 (S5PV210) processor control unit, a power supply device, wired and wireless communication, input/output unit. We develope Linux kernel and dedicated device driver for WiFi module and camera. The license plate recognition algorithm is consisted of setting candidate plates areas with canny edge detector, extracting license plate number with Labeling, recognizing with template matching, etc. The number that is recognized by the device is transmitted to the remote server via the user mobile phone, and the server re-transfer the vehicle information in the database to the portable device. To verify the utility of the proposed system, user photographs the license plate of any vehicle in the natural environment. Confirming the recognition result, the recognition rate was 95%. The proposed system was suitable for low cost portable license plate recognition device, it enabled the stability of the system when used long time by using the Android operating system.

Hybrid (refrctive/diffractive) lens design for the ultra-compact camera module (초소형 영상 전송 모듈용 DOE(Diffractive optical element)렌즈의 설계 및 평가)

  • Lee, Hwan-Seon;Rim, Cheon-Seog;Jo, jae-Heung;Chang, Soo;Lim, Hyun-Kyu
    • Korean Journal of Optics and Photonics
    • /
    • v.12 no.3
    • /
    • pp.240-249
    • /
    • 2001
  • A high speed ultra-compact lens with a diffractive optical element (DOE) is designed, which can be applied to mobile communication devices such as IMT2000, PDA, notebook computer, etc. The designed hybrid lens has sufficiently high performance of less than f/2.2, compact size of 3.3 mm (1st surf. to image), and wide field angle of more than 30 deg. compared with the specifications of a single lens. By proper choice of the aspheric and DOE surface which has very large negative dispersion, we can correct chromatic and high order aberrations through the optimization technique. From Seidel third order aberration theory and Sweatt modeling, the initial data and surface configurations, that is, the combination condition of the DOE and the aspherical surface are obtained. However, due to the consideration of diffraction efficiency of a DOE, we can choose only four cases as the optimization input, and present the best solution after evaluating and comparing those four cases. On the other hand, we also report dramatic improvement in optical performance by inserting another refractive lens (so-called, field flattener), that keeps the refractive power of an original DOE lens and makes the petzval sum zero in the original DOE lens system. ystem.

  • PDF

Change Attention-based Vehicle Scratch Detection System (변화 주목 기반 차량 흠집 탐지 시스템)

  • Lee, EunSeong;Lee, DongJun;Park, GunHee;Lee, Woo-Ju;Sim, Donggyu;Oh, Seoung-Jun
    • Journal of Broadcast Engineering
    • /
    • v.27 no.2
    • /
    • pp.228-239
    • /
    • 2022
  • In this paper, we propose an unmanned vehicle scratch detection deep learning model for car sharing services. Conventional scratch detection models consist of two steps: 1) a deep learning module for scratch detection of images before and after rental, 2) a manual matching process for finding newly generated scratches. In order to build a fully automatic scratch detection model, we propose a one-step unmanned scratch detection deep learning model. The proposed model is implemented by applying transfer learning and fine-tuning to the deep learning model that detects changes in satellite images. In the proposed car sharing service, specular reflection greatly affects the scratch detection performance since the brightness of the gloss-treated automobile surface is anisotropic and a non-expert user takes a picture with a general camera. In order to reduce detection errors caused by specular reflected light, we propose a preprocessing process for removing specular reflection components. For data taken by mobile phone cameras, the proposed system can provide high matching performance subjectively and objectively. The scores for change detection metrics such as precision, recall, F1, and kappa are 67.90%, 74.56%, 71.08%, and 70.18%, respectively.

Automatic gasometer reading system using selective optical character recognition (관심 문자열 인식 기술을 이용한 가스계량기 자동 검침 시스템)

  • Lee, Kyohyuk;Kim, Taeyeon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.1-25
    • /
    • 2020
  • In this paper, we suggest an application system architecture which provides accurate, fast and efficient automatic gasometer reading function. The system captures gasometer image using mobile device camera, transmits the image to a cloud server on top of private LTE network, and analyzes the image to extract character information of device ID and gas usage amount by selective optical character recognition based on deep learning technology. In general, there are many types of character in an image and optical character recognition technology extracts all character information in an image. But some applications need to ignore non-of-interest types of character and only have to focus on some specific types of characters. For an example of the application, automatic gasometer reading system only need to extract device ID and gas usage amount character information from gasometer images to send bill to users. Non-of-interest character strings, such as device type, manufacturer, manufacturing date, specification and etc., are not valuable information to the application. Thus, the application have to analyze point of interest region and specific types of characters to extract valuable information only. We adopted CNN (Convolutional Neural Network) based object detection and CRNN (Convolutional Recurrent Neural Network) technology for selective optical character recognition which only analyze point of interest region for selective character information extraction. We build up 3 neural networks for the application system. The first is a convolutional neural network which detects point of interest region of gas usage amount and device ID information character strings, the second is another convolutional neural network which transforms spatial information of point of interest region to spatial sequential feature vectors, and the third is bi-directional long short term memory network which converts spatial sequential information to character strings using time-series analysis mapping from feature vectors to character strings. In this research, point of interest character strings are device ID and gas usage amount. Device ID consists of 12 arabic character strings and gas usage amount consists of 4 ~ 5 arabic character strings. All system components are implemented in Amazon Web Service Cloud with Intel Zeon E5-2686 v4 CPU and NVidia TESLA V100 GPU. The system architecture adopts master-lave processing structure for efficient and fast parallel processing coping with about 700,000 requests per day. Mobile device captures gasometer image and transmits to master process in AWS cloud. Master process runs on Intel Zeon CPU and pushes reading request from mobile device to an input queue with FIFO (First In First Out) structure. Slave process consists of 3 types of deep neural networks which conduct character recognition process and runs on NVidia GPU module. Slave process is always polling the input queue to get recognition request. If there are some requests from master process in the input queue, slave process converts the image in the input queue to device ID character string, gas usage amount character string and position information of the strings, returns the information to output queue, and switch to idle mode to poll the input queue. Master process gets final information form the output queue and delivers the information to the mobile device. We used total 27,120 gasometer images for training, validation and testing of 3 types of deep neural network. 22,985 images were used for training and validation, 4,135 images were used for testing. We randomly splitted 22,985 images with 8:2 ratio for training and validation respectively for each training epoch. 4,135 test image were categorized into 5 types (Normal, noise, reflex, scale and slant). Normal data is clean image data, noise means image with noise signal, relfex means image with light reflection in gasometer region, scale means images with small object size due to long-distance capturing and slant means images which is not horizontally flat. Final character string recognition accuracies for device ID and gas usage amount of normal data are 0.960 and 0.864 respectively.