• Title/Summary/Keyword: Image-based localization

Search Result 258, Processing Time 0.027 seconds

Image Feature-Based Real-Time RGB-D 3D SLAM with GPU Acceleration (GPU 가속화를 통한 이미지 특징점 기반 RGB-D 3차원 SLAM)

  • Lee, Donghwa;Kim, Hyongjin;Myung, Hyun
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.19 no.5
    • /
    • pp.457-461
    • /
    • 2013
  • This paper proposes an image feature-based real-time RGB-D (Red-Green-Blue Depth) 3D SLAM (Simultaneous Localization and Mapping) system. RGB-D data from Kinect style sensors contain a 2D image and per-pixel depth information. 6-DOF (Degree-of-Freedom) visual odometry is obtained through the 3D-RANSAC (RANdom SAmple Consensus) algorithm with 2D image features and depth data. For speed up extraction of features, parallel computation is performed with GPU acceleration. After a feature manager detects a loop closure, a graph-based SLAM algorithm optimizes trajectory of the sensor and builds a 3D point cloud based map.

A Study on Localization of Text in Natural Scene Images (자연 영상에서의 정확한 문자 검출에 관한 연구)

  • Choi, Mi-Young;Kim, Gye-Young;Choi, Hyung-Il
    • Journal of the Korea Society of Computer and Information
    • /
    • v.13 no.5
    • /
    • pp.77-84
    • /
    • 2008
  • This paper proposes a new approach to eliminate the reflectance component for the localization of text in natural scene images. Natural scene images normally have an illumination component as well as a reflectance component. It is well known that a reflectance component usually obstructs the task of detecting and recognizing objects like texts in the scene, since it blurs out an overall image. We have developed an approach that efficiently removes reflectance components while Preserving illumination components. We decided whether an input image hits Normal or Polarized for determining the light environment, using the histogram which consisted of a red component. In the normal image, we acquired the text region without additional processing. Otherwise we removed light reflecting from the object using homomorphic filtering in the polarized image. And then this decided the each text region based on the color merging technique and the Saliency Map. Finally, we localized text region on these two candidate regions.

  • PDF

Externalization of sound image in 3D sound system based on headphone

  • Youngsik Yoon;Park, Youngjin
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2002.10a
    • /
    • pp.51.3-51
    • /
    • 2002
  • 3D sound user often finds the results that the sound image appear to originate either inside, or close to, the head when he uses headphone-based binaural system. This phenomenon is called in-head localization(IHL). The main factors were chosen to evaluate externalization performance : individualized HRTFs, near-field HRTF characteristics and reverberation. Direct comparison was conducted among them, especially two factors\ulcorner reverberation and near-field HRTFs.

  • PDF

Perceptual Localization of a phantom sound image for Ultrahigh-Definition TV (UHD TV를 위한 가상 음상의 인지 위치)

  • Lee, Young-Woo;Kim, Sun-Min
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.47 no.5
    • /
    • pp.9-17
    • /
    • 2010
  • This paper presents a localization perception of a phantom sound image for ultrahigh-definition TV with respect to various loudspeaker configurations; two-horizontal, two-vertical and triplet loudspeakers. Vector base amplitude panning algorithm with modification for non-equidistant loudspeaker setup is applied to create the phantom sound image. In order to practically study the localization performance in real situation, the listening tests were conducted at the on-axis and off-axis positions of TV in normal listening room. A method of adjustment which can reduce the ambiguity of a perceived angle is exploited to evaluate the angles of octave-band signals. The subjects changed the panning angle until the real sound source and virtually panned source were coincident. A spatial blurring can be measured by examining the differences of the panning angles perceived with respect to each band. The listening tests show that the triplet panning method has better performance than vertical panning in view of perceptual localization and spatial blurring at both on-axis and off-axis positions.

Development of Omnidirectional Ranging System Based on Structured Light Image (구조광 영상기반 전방향 거리측정 시스템 개발)

  • Shin, Jin;Yi, Soo-Yeong
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.18 no.5
    • /
    • pp.479-486
    • /
    • 2012
  • In this paper, a ranging system is proposed that is able to measure 360 degree omnidirectional distances to environment objects. The ranging system is based on the structured light imaging system with catadioptric omnidirectional mirror. In order to make the ranging system robust against environmental illumination, efficient structured light image processing algorithms are developed; sequential integration of difference images with modulated structured light and radial search based on Bresenham line drawing algorithm. A dedicated FPGA image processor is developed to speed up the overall image processing. Also the distance equation is derived in the omnidirectional imaging system with a hyperbolic mirror. It is expected that the omnidirectional ranging system is useful for mapping and localization of mobile robot. Experiments are carried out to verify the performance of the proposed ranging system.

Image Authentication and Restoration Using Digital Watermarking by Quantization of Integer Wavelet Transform Coefficients

  • Ahsan, Tanveer;Chong, Ui-Pil
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.13 no.4
    • /
    • pp.187-193
    • /
    • 2012
  • An image authentication scheme for gray scale image through embedding a digital watermark by quantization of Integer Wavelet Transform (IWT) coefficients of the image is proposed in this paper. Proposed method is designed to detect modification of an image and to identify tampered location of the image. To embed the watermark mid-frequency band of a second level IWT was used. An approximation of the original image based on LL band was stored in LSB bits of the pixel data as a recovery mark for restoration of the image. Watermarked image has achieved a good PSNR of 40 dB compared to original cover image. Restored image quality was also very good with a PSNR of more than 35 dB compared to unmodified watermarked image even when 25% of the received image is cropped. Thus, the proposed method ensures a proper balance between the fidelity of the watermarked image and the quality of the restored image.

A Method for Body Keypoint Localization based on Object Detection using the RGB-D information (RGB-D 정보를 이용한 객체 탐지 기반의 신체 키포인트 검출 방법)

  • Park, Seohee;Chun, Junchul
    • Journal of Internet Computing and Services
    • /
    • v.18 no.6
    • /
    • pp.85-92
    • /
    • 2017
  • Recently, in the field of video surveillance, a Deep Learning based learning method has been applied to a method of detecting a moving person in a video and analyzing the behavior of a detected person. The human activity recognition, which is one of the fields this intelligent image analysis technology, detects the object and goes through the process of detecting the body keypoint to recognize the behavior of the detected object. In this paper, we propose a method for Body Keypoint Localization based on Object Detection using RGB-D information. First, the moving object is segmented and detected from the background using color information and depth information generated by the two cameras. The input image generated by rescaling the detected object region using RGB-D information is applied to Convolutional Pose Machines for one person's pose estimation. CPM are used to generate Belief Maps for 14 body parts per person and to detect body keypoints based on Belief Maps. This method provides an accurate region for objects to detect keypoints an can be extended from single Body Keypoint Localization to multiple Body Keypoint Localization through the integration of individual Body Keypoint Localization. In the future, it is possible to generate a model for human pose estimation using the detected keypoints and contribute to the field of human activity recognition.

Bayesian Sensor Fusion of Monocular Vision and Laser Structured Light Sensor for Robust Localization of a Mobile Robot (이동 로봇의 강인 위치 추정을 위한 단안 비젼 센서와 레이저 구조광 센서의 베이시안 센서융합)

  • Kim, Min-Young;Ahn, Sang-Tae;Cho, Hyung-Suck
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.16 no.4
    • /
    • pp.381-390
    • /
    • 2010
  • This paper describes a procedure of the map-based localization for mobile robots by using a sensor fusion technique in structured environments. A combination of various sensors with different characteristics and limited sensibility has advantages in view of complementariness and cooperation to obtain better information on the environment. In this paper, for robust self-localization of a mobile robot with a monocular camera and a laser structured light sensor, environment information acquired from two sensors is combined and fused by a Bayesian sensor fusion technique based on the probabilistic reliability function of each sensor predefined through experiments. For the self-localization using the monocular vision, the robot utilizes image features consisting of vertical edge lines from input camera images, and they are used as natural landmark points in self-localization process. However, in case of using the laser structured light sensor, it utilizes geometrical features composed of corners and planes as natural landmark shapes during this process, which are extracted from range data at a constant height from the navigation floor. Although only each feature group of them is sometimes useful to localize mobile robots, all features from the two sensors are simultaneously used and fused in term of information for reliable localization under various environment conditions. To verify the advantage of using multi-sensor fusion, a series of experiments are performed, and experimental results are discussed in detail.

Improvement of front-back sound localization characteristics in headphone-based 3D sound generation (헤드폰 기반의 입체음향 생성에서 앞/뒤 음상정위 특성 개선)

  • 김경훈;김시호;배건성;최송인;박만호
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.8C
    • /
    • pp.1142-1148
    • /
    • 2004
  • A binaural filtering method using HRTF DB is generally used to make the headphone-based 3D sound. But it can make some confusion between front and back directions or between up and down directions due to the non-individual HRTF depending on each listener. To reduce the confusion of sound image localization, we propose a new method to boost the spectral cue by modifying HRTF spectra with spectrum difference between front and back directions. Informal listening tests show that the proposed method improves the front-back sound localization characteristics much better than the conventional methods

A Real-time Vehicle Localization Algorithm for Autonomous Parking System (자율 주차 시스템을 위한 실시간 차량 추출 알고리즘)

  • Hahn, Jong-Woo;Choi, Young-Kyu
    • Journal of the Semiconductor & Display Technology
    • /
    • v.10 no.2
    • /
    • pp.31-38
    • /
    • 2011
  • This paper introduces a video based traffic monitoring system for detecting vehicles and obstacles on the road. To segment moving objects from image sequence, we adopt the background subtraction algorithm based on the local binary patterns (LBP). Recently, LBP based texture analysis techniques are becoming popular tools for various machine vision applications such as face recognition, object classification and so on. In this paper, we adopt an extension of LBP, called the Diagonal LBP (DLBP), to handle the background subtraction problem arise in vision-based autonomous parking systems. It reduces the code length of LBP by half and improves the computation complexity drastically. An edge based shadow removal and blob merging procedure are also applied to the foreground blobs, and a pose estimation technique is utilized for calculating the position and heading angle of the moving object precisely. Experimental results revealed that our system works well for real-time vehicle localization and tracking applications.