• Title/Summary/Keyword: Focus Distance of Camera

Search Result 44, Processing Time 0.019 seconds

Implementation of Infant Learning Content using Augmented Reality (증강현실을 이용한 유아용 학습 콘텐츠의 구현)

  • Lee, Jong-Hyeok;Cho, Hyun-Wook
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.15 no.1
    • /
    • pp.257-263
    • /
    • 2011
  • Recently as AR(Augmented Reality) is focus of attention, AR is applied to various fields and is expected its valuable use. In this paper, we implemented the system based on Goblin XNA which supports high resolution model file and higher AR. We confirmed the relation of model output among the number of marker, the location and changes of camera distance. And we produced the infantile studying contents using AR and embodied. In implemented contents, we showed the familiar character to infants on each page marker. As the result of it, we can raise their concentration and at a time studying supporters can use the contents easily as well. Also we put 3 marker on each page of contents to recognize it smoothly in case one part of it is hidden by any obstacle. Finally we maximized the learning effect such as presence and immersion in studying through reinforcing 3D models according to the every situation.

Image Recognition Using Colored-hear Transformation Based On Human Synesthesia (인간의 공감각에 기반을 둔 색청변환을 이용한 영상 인식)

  • Shin, Seong-Yoon;Moon, Hyung-Yoon;Pyo, Seong-Bae
    • Journal of the Korea Society of Computer and Information
    • /
    • v.13 no.2
    • /
    • pp.135-141
    • /
    • 2008
  • In this paper, we propose colored-hear recognition that distinguishing feature of synesthesia for human sensing by shared vision and specific sense of hearing. We perceived what potential influence of human's structured object recognition by visual analysis through the camera, So we've studied how to make blind persons can feel similar vision of real object. First of all, object boundaries are detected in the image data representing a specific scene. Then, four specific features such as object location in the image focus, feeling of average color, distance information of each object, and object area are extracted from picture. Finally, mapping these features to the audition factors. The audition factors are used to recognize vision for blind persons. Proposed colored-hear transformation for recognition can get fast and detail perception, and can be transmit information for sense at the same time. Thus, we were get a food result when applied this concepts to blind person's case of image recognition.

  • PDF

Intuitive Manipulation of Deformable Cloth Object Based on Augmented Reality for Mobile Game (모바일 게임을 위한 증강현실 기반 직관적 변형 직물객체 조작)

  • Kim, Sang-Joon;Hong, Min;Choi, Yoo-Joo
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.7 no.4
    • /
    • pp.159-168
    • /
    • 2018
  • In recent, mobile augmented reality game which has been attracting high attention is considered to be an good approach to increase immersion. In conventional augmented reality-based games that recognize target objects using a mobile camera and show the matching game characters, touch-based interaction is mainly used. In this paper, we propose an intuitive interaction method which manipulates a deformable game object by moving a target image of augmented reality in order to enhacne the immersion of the game. In the proposed method, the deformable object is intuitively manipulated by calculating the distance and direction between the target images and by adjusting the external force applied to the deformable object using them. In this paper, we focus on the cloth deformable object which is widely used for natural object animation in game contents and implement natural cloth simulation interacting with game objects represented by wind and rigid objects. In the experiments, we compare the previous commercial cloth model with the proposed method and show the proposed method can represent cloth animation more realistically.

Automatic gasometer reading system using selective optical character recognition (관심 문자열 인식 기술을 이용한 가스계량기 자동 검침 시스템)

  • Lee, Kyohyuk;Kim, Taeyeon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.1-25
    • /
    • 2020
  • In this paper, we suggest an application system architecture which provides accurate, fast and efficient automatic gasometer reading function. The system captures gasometer image using mobile device camera, transmits the image to a cloud server on top of private LTE network, and analyzes the image to extract character information of device ID and gas usage amount by selective optical character recognition based on deep learning technology. In general, there are many types of character in an image and optical character recognition technology extracts all character information in an image. But some applications need to ignore non-of-interest types of character and only have to focus on some specific types of characters. For an example of the application, automatic gasometer reading system only need to extract device ID and gas usage amount character information from gasometer images to send bill to users. Non-of-interest character strings, such as device type, manufacturer, manufacturing date, specification and etc., are not valuable information to the application. Thus, the application have to analyze point of interest region and specific types of characters to extract valuable information only. We adopted CNN (Convolutional Neural Network) based object detection and CRNN (Convolutional Recurrent Neural Network) technology for selective optical character recognition which only analyze point of interest region for selective character information extraction. We build up 3 neural networks for the application system. The first is a convolutional neural network which detects point of interest region of gas usage amount and device ID information character strings, the second is another convolutional neural network which transforms spatial information of point of interest region to spatial sequential feature vectors, and the third is bi-directional long short term memory network which converts spatial sequential information to character strings using time-series analysis mapping from feature vectors to character strings. In this research, point of interest character strings are device ID and gas usage amount. Device ID consists of 12 arabic character strings and gas usage amount consists of 4 ~ 5 arabic character strings. All system components are implemented in Amazon Web Service Cloud with Intel Zeon E5-2686 v4 CPU and NVidia TESLA V100 GPU. The system architecture adopts master-lave processing structure for efficient and fast parallel processing coping with about 700,000 requests per day. Mobile device captures gasometer image and transmits to master process in AWS cloud. Master process runs on Intel Zeon CPU and pushes reading request from mobile device to an input queue with FIFO (First In First Out) structure. Slave process consists of 3 types of deep neural networks which conduct character recognition process and runs on NVidia GPU module. Slave process is always polling the input queue to get recognition request. If there are some requests from master process in the input queue, slave process converts the image in the input queue to device ID character string, gas usage amount character string and position information of the strings, returns the information to output queue, and switch to idle mode to poll the input queue. Master process gets final information form the output queue and delivers the information to the mobile device. We used total 27,120 gasometer images for training, validation and testing of 3 types of deep neural network. 22,985 images were used for training and validation, 4,135 images were used for testing. We randomly splitted 22,985 images with 8:2 ratio for training and validation respectively for each training epoch. 4,135 test image were categorized into 5 types (Normal, noise, reflex, scale and slant). Normal data is clean image data, noise means image with noise signal, relfex means image with light reflection in gasometer region, scale means images with small object size due to long-distance capturing and slant means images which is not horizontally flat. Final character string recognition accuracies for device ID and gas usage amount of normal data are 0.960 and 0.864 respectively.