• 제목/요약/키워드: Monocular Camera

검색결과 111건 처리시간 0.025초

Error Minimized Laser Beam Point Detection Using Mono-Camera (한 개의 카메라를 이용한 최소오차 레이저 빔 포인터 위치 검출)

  • Lee, Wang-Heon;Lee, Hyun-Chang
    • Journal of the Korea Society of Computer and Information
    • /
    • 제12권6호
    • /
    • pp.69-76
    • /
    • 2007
  • The main stream of presentation is interrupted because of the direct manipulation of their PC frequently so as to control the screen and file open and so on. A variety of products have been developed to solve these inconveniences of the conventional laser beam pointer [LBP] by simply adding a mouse function to the previous LBP. However. the LBPs fully supporting a mouse function are not yet appeared. In this paper. we developed the LBP fully fulfilling a mouse function using mono-camera as well as a robust image processing and analyzed the position detection accuracy. Finally we verified the developed LBP does not only fulfill a mouse function but also solve the defects of the current laser pointer such as inconvenient installation and Position detection errors due to the illumination and viewing direction changes.

  • PDF

Moving Object Extraction and Relative Depth Estimation of Backgrould regions in Video Sequences (동영상에서 물체의 추출과 배경영역의 상대적인 깊이 추정)

  • Park Young-Min;Chang Chu-Seok
    • The KIPS Transactions:PartB
    • /
    • 제12B권3호
    • /
    • pp.247-256
    • /
    • 2005
  • One of the classic research problems in computer vision is that of stereo, i.e., the reconstruction of three dimensional shape from two or more images. This paper deals with the problem of extracting depth information of non-rigid dynamic 3D scenes from general 2D video sequences taken by monocular camera, such as movies, documentaries, and dramas. Depth of the blocks are extracted from the resultant block motions throughout following two steps: (i) calculation of global parameters concerned with camera translations and focal length using the locations of blocks and their motions, (ii) calculation of each block depth relative to average image depth using the global parameters and the location of the block and its motion, Both singular and non-singular cases are experimented with various video sequences. The resultant relative depths and ego-motion object shapes are virtually identical to human vision.

Monocular Vision Based Localization System using Hybrid Features from Ceiling Images for Robot Navigation in an Indoor Environment (실내 환경에서의 로봇 자율주행을 위한 천장영상으로부터의 이종 특징점을 이용한 단일비전 기반 자기 위치 추정 시스템)

  • Kang, Jung-Won;Bang, Seok-Won;Atkeson, Christopher G.;Hong, Young-Jin;Suh, Jin-Ho;Lee, Jung-Woo;Chung, Myung-Jin
    • The Journal of Korea Robotics Society
    • /
    • 제6권3호
    • /
    • pp.197-209
    • /
    • 2011
  • This paper presents a localization system using ceiling images in a large indoor environment. For a system with low cost and complexity, we propose a single camera based system that utilizes ceiling images acquired from a camera installed to point upwards. For reliable operation, we propose a method using hybrid features which include natural landmarks in a natural scene and artificial landmarks observable in an infrared ray domain. Compared with previous works utilizing only infrared based features, our method reduces the required number of artificial features as we exploit both natural and artificial features. In addition, compared with previous works using only natural scene, our method has an advantage in the convergence speed and robustness as an observation of an artificial feature provides a crucial clue for robot pose estimation. In an experiment with challenging situations in a real environment, our method was performed impressively in terms of the robustness and accuracy. To our knowledge, our method is the first ceiling vision based localization method using features from both visible and infrared rays domains. Our system can be easily utilized with a variety of service robot applications in a large indoor environment.

Deep Learning Based On-Device Augmented Reality System using Multiple Images (다중영상을 이용한 딥러닝 기반 온디바이스 증강현실 시스템)

  • Jeong, Taehyeon;Park, In Kyu
    • Journal of Broadcast Engineering
    • /
    • 제27권3호
    • /
    • pp.341-350
    • /
    • 2022
  • In this paper, we propose a deep learning based on-device augmented reality (AR) system in which multiple input images are used to implement the correct occlusion in a real environment. The proposed system is composed of three technical steps; camera pose estimation, depth estimation, and object augmentation. Each step employs various mobile frameworks to optimize the processing on the on-device environment. Firstly, in the camera pose estimation stage, the massive computation involved in feature extraction is parallelized using OpenCL which is the GPU parallelization framework. Next, in depth estimation, monocular and multiple image-based depth image inference is accelerated using the mobile deep learning framework, i.e. TensorFlow Lite. Finally, object augmentation and occlusion handling are performed on the OpenGL ES mobile graphics framework. The proposed augmented reality system is implemented as an application in the Android environment. We evaluate the performance of the proposed system in terms of augmentation accuracy and the processing time in the mobile as well as PC environments.

Multi-Scale, Multi-Object and Real-Time Face Detection and Head Pose Estimation Using Deep Neural Networks (다중크기와 다중객체의 실시간 얼굴 검출과 머리 자세 추정을 위한 심층 신경망)

  • Ahn, Byungtae;Choi, Dong-Geol;Kweon, In So
    • The Journal of Korea Robotics Society
    • /
    • 제12권3호
    • /
    • pp.313-321
    • /
    • 2017
  • One of the most frequently performed tasks in human-robot interaction (HRI), intelligent vehicles, and security systems is face related applications such as face recognition, facial expression recognition, driver state monitoring, and gaze estimation. In these applications, accurate head pose estimation is an important issue. However, conventional methods have been lacking in accuracy, robustness or processing speed in practical use. In this paper, we propose a novel method for estimating head pose with a monocular camera. The proposed algorithm is based on a deep neural network for multi-task learning using a small grayscale image. This network jointly detects multi-view faces and estimates head pose in hard environmental conditions such as illumination change and large pose change. The proposed framework quantitatively and qualitatively outperforms the state-of-the-art method with an average head pose mean error of less than $4.5^{\circ}$ in real-time.

Real Time Traffic Signal Recognition Using HSI and YCbCr Color Models and Adaboost Algorithm (HSI/YCbCr 색상모델과 에이다부스트 알고리즘을 이용한 실시간 교통신호 인식)

  • Park, Sanghoon;Lee, Joonwoong
    • Transactions of the Korean Society of Automotive Engineers
    • /
    • 제24권2호
    • /
    • pp.214-224
    • /
    • 2016
  • This paper proposes an algorithm to effectively detect the traffic lights and recognize the traffic signals using a monocular camera mounted on the front windshield glass of a vehicle in day time. The algorithm consists of three main parts. The first part is to generate the candidates of a traffic light. After conversion of RGB color model into HSI and YCbCr color spaces, the regions considered as a traffic light are detected. For these regions, edge processing is applied to extract the borders of the traffic light. The second part is to divide the candidates into traffic lights and non-traffic lights using Haar-like features and Adaboost algorithm. The third part is to recognize the signals of the traffic light using a template matching. Experimental results show that the proposed algorithm successfully detects the traffic lights and recognizes the traffic signals in real time in a variety of environments.

Development of Omnidirectional Active Marker for Motion Capture System with a Monocular PSD Camera (단안 PSD 카메라를 이용한 모션캡쳐 시스템을 위한 전방향성 능동마커 개발)

  • Seo, Pyeong-Won;Ryu, Young-Kee;Oh, Choon-Suk
    • Proceedings of the KIEE Conference
    • /
    • 대한전기학회 2008년도 학술대회 논문집 정보 및 제어부문
    • /
    • pp.379-381
    • /
    • 2008
  • 본 논문에서는 가정용 비디오 게임에 사용 가능한 고속의 저가형 모션캡쳐, 시스템에 사용되는 전 방향 특성을 갖는 IR 능동 마커의 개발을 목표로 하고 있다. 현재 영화나 게임에서 모션캡쳐를 응용한 시스템 및 컨텐츠들이 많이 선보기고 있으며, 인기를 모으고 있는 추세이다. 이러한 흐름에 맞추어 우리는 이미 저가이면서 속도가 빠른 PSD(Position Sensitive Detector) 센서를 이용만 스테레오 비젼 기반의 PSD 모션캡쳐 시스템(Stereo vision based PSD motion capture system)과 광량 보정 기반의 단일 PSD 모션캡쳐 시스템(Intensity Calibration based single PSD motion capture system) 그리고 일정간격의 두 능동마커 기반의 단안 PSD 모션캡쳐 시스템(Two active markers at fixed distance based single PSD motion capture system)등을 소개한 바 있다. 본 논문에서 제안하는 전방향 특성을 갖는 IR 능동 마커는 일정간격의 두 능동마커 기반의 단안 PSD 모션캡쳐 시스템에 적용하여 보다 정밀한 3차원 좌표 측정을 할 수 있도록 한다. 이를 위해 본 논문에서는 동일 특성을 갖는 마커를 제작하고 평가하여 일정간격의 두 능동마커 기반의 단안 PSD 모션캡쳐 시스템에 적합한 마커의 제작 방법을 제안하였다.

  • PDF

A Gaze Detection Technique Using a Monocular Camera System (단안 카메라 환경에서의 시선 위치 추적)

  • 박강령;김재희
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • 제26권10B호
    • /
    • pp.1390-1398
    • /
    • 2001
  • 시선 위치 추적이란 사용자가 모니터 상의 어느 지점을 쳐다보고 있는 지를 파악해 내는 기술이다. 시선 위치를 파악하기 위해 본 논문에서는 2차원 카메라 영상으로부터 얼굴 영역 및 얼굴 특징점을 추출한다. 초기에 모니터상의 3 지점을 쳐다볼 때 얼굴 특징점들은 움직임의 변화를 나타내며, 이로부터 카메라 보정 및 매개변수 추정 방법을 이용하여 얼굴특징점의 3차원 위치를 추정한다. 이후 사용자가 모니터 상의 또 다른 지점을 쳐다볼 때 얼굴 특징점의 변화된 3차원 위치는 3차원 움직임 추정방법 및 아핀변환을 이용하여 구해낸다. 이로부터 변화된 얼굴 특징점 및 이러한 얼굴 특징점으로 구성된 얼굴평면이 구해지며, 이러한 평면의 법선으로부터 모니터 상의 시선위치를 구할 수 있다. 실험 결과 19인치 모니터를 사용하여 모니터와 사용자까지의 거리를 50∼70cm정도 유지하였을 때 약 2.08인치의 시선위치에러 성능을 얻었다. 이 결과는 Rikert의 논문에서 나타낸 시선위치추적 성능(5.08cm 에러)과 비슷한 결과를 나타낸다. 그러나 Rikert의 방법은 모니터와 사용자 얼굴까지의 거리는 항상 고정시켜야 한다는 단점이 있으며, 얼굴의 자연스러운 움직임(회전 및 이동)이 발생하는 경우 시선위치추적 에러가 증가되는 문제점이 있다. 동시에 그들의 방법은 사용자 얼굴의 뒤 배경에 복잡한 물체가 없는 것으로 제한조건을 두고 있으며 처리 시간이 상당히 오래 걸리는 문제점이 있다. 그러나 본 논문에서 제안하는 시선 위치 추적 방법은 배경이 복잡한 사무실 환경에서도 사용가능하며, 약 3초 이내의 처리 시간(200MHz Pentium PC)이 소요됨을 알 수 있었다.

  • PDF

CONSIDERATION OF THE RELATION BETWEEN DISTANCE AND CHANGE OF PANEL COLOR BASED ON AERIAL PERSPECTIVE

  • Horiuchi, Hitoshi;Kaneko, Satoru;Sato, Mie;Ozaki, Koichi;Kasuga, Masao
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 한국방송공학회 2009년도 IWAIT
    • /
    • pp.695-698
    • /
    • 2009
  • Three-dimensional (3D) shape recognition and distance recognition methods utilizing monocular camera systems have been required for field of virtual-reality, computer graphics, measurement technology and robot technology. There have been many studies regarding 3D shape and distance recognition based on geometric and optical information, and it is now possible to accurately measure the geometric information of an object at short range distances. However, these methods cannot currently be applied to long range objects. In the field of virtual-reality, all visual objects must be presented at widely varying ranges, even though some objects will be hazed over. In order to achieve distance recognition from a landscape image, we focused on the use of aerial perspective to simulate a type of depth perception and investigated the relationship between distance and color perception. The applicability of our proposed method was demonstrated in experimental results.

  • PDF

Feature Based Techniques for a Driver's Distraction Detection using Supervised Learning Algorithms based on Fixed Monocular Video Camera

  • Ali, Syed Farooq;Hassan, Malik Tahir
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제12권8호
    • /
    • pp.3820-3841
    • /
    • 2018
  • Most of the accidents occur due to drowsiness while driving, avoiding road signs and due to driver's distraction. Driver's distraction depends on various factors which include talking with passengers while driving, mood disorder, nervousness, anger, over-excitement, anxiety, loud music, illness, fatigue and different driver's head rotations due to change in yaw, pitch and roll angle. The contribution of this paper is two-fold. Firstly, a data set is generated for conducting different experiments on driver's distraction. Secondly, novel approaches are presented that use features based on facial points; especially the features computed using motion vectors and interpolation to detect a special type of driver's distraction, i.e., driver's head rotation due to change in yaw angle. These facial points are detected by Active Shape Model (ASM) and Boosted Regression with Markov Networks (BoRMaN). Various types of classifiers are trained and tested on different frames to decide about a driver's distraction. These approaches are also scale invariant. The results show that the approach that uses the novel ideas of motion vectors and interpolation outperforms other approaches in detection of driver's head rotation. We are able to achieve a percentage accuracy of 98.45 using Neural Network.