• 제목/요약/키워드: visual localization

검색결과 144건 처리시간 0.024초

Visual Positioning System based on Voxel Labeling using Object Simultaneous Localization And Mapping

  • Jung, Tae-Won;Kim, In-Seon;Jung, Kye-Dong
    • International Journal of Advanced Culture Technology
    • /
    • 제9권4호
    • /
    • pp.302-306
    • /
    • 2021
  • Indoor localization is one of the basic elements of Location-Based Service, such as indoor navigation, location-based precision marketing, spatial recognition of robotics, augmented reality, and mixed reality. We propose a Voxel Labeling-based visual positioning system using object simultaneous localization and mapping (SLAM). Our method is a method of determining a location through single image 3D cuboid object detection and object SLAM for indoor navigation, then mapping to create an indoor map, addressing it with voxels, and matching with a defined space. First, high-quality cuboids are created from sampling 2D bounding boxes and vanishing points for single image object detection. And after jointly optimizing the poses of cameras, objects, and points, it is a Visual Positioning System (VPS) through matching with the pose information of the object in the voxel database. Our method provided the spatial information needed to the user with improved location accuracy and direction estimation.

수중 구조물 형상의 영상 정보를 이용한 수중로봇 위치인식 기법 (Localization of AUV Using Visual Shape Information of Underwater Structures)

  • 정종대;최수영;최현택;명현
    • 한국해양공학회지
    • /
    • 제29권5호
    • /
    • pp.392-397
    • /
    • 2015
  • An autonomous underwater vehicle (AUV) can perform flexible operations even in complex underwater environments because of its autonomy. Localization is one of the key components of this autonomous navigation. Because the inertial navigation system of an AUV suffers from drift, observing fixed objects in an inertial reference system can enhance the localization performance. In this paper, we propose a method of AUV localization using visual measurements of underwater structures. A camera measurement model that emulates the camera’s observations of underwater structures is designed in a particle filtering framework. Then, the particle weight is updated based on the extracted visual information of the underwater structures. The proposed method is validated based on the results of experiments performed in a structured basin environment.

Audio-visual Spatial Coherence Judgments in the Peripheral Visual Fields

  • 이채봉;강대기
    • 융합신호처리학회논문지
    • /
    • 제16권2호
    • /
    • pp.35-39
    • /
    • 2015
  • Auditory and visual stimuli presented in the peripheral visual field were perceived as spatially coincident when the auditory stimulus was presented five to seven degrees outwards from the direction of the visual stimulus. Furthermore, judgments of the perceived distance between auditory and visual stimuli presented in the periphery did not increase when an auditory stimulus was presented in the peripheral side of the visual stimulus. As to the origin of this phenomenon, there would seem to be two possibilities. One is that the participants could not perceptually distinguish the distance on the peripheral side because of the limitation of accuracy perception. The other is that the participants could distinguish the distances, but could not evaluate them because of the insufficient experimental setup of auditory stimuli. In order to confirm which of these two alternative explanations is valid, we conducted an experiment similar to that of our previous study using a sufficient number of loudspeakers for the presentation of auditory stimuli. Results revealed that judgments of perceived distance increased on the peripheral side. This indicates that we can perceive discrimination between audio and visual stimuli on the peripheral side.

휴모노이드 로봇을 위한 시청각 정보 기반 음원 정위 시스템 구현 (Implementation of Sound Source Localization Based on Audio-visual Information for Humanoid Robots)

  • 박정욱;나승유;김진영
    • 음성과학
    • /
    • 제11권4호
    • /
    • pp.29-42
    • /
    • 2004
  • This paper presents an implementation of real-time speaker localization using audio-visual information. Four channels of microphone signals are processed to detect vertical as well as horizontal speaker positions. At first short-time average magnitude difference function(AMDF) signals are used to determine whether the microphone signals are human voices or not. And then the orientation and distance information of the sound sources can be obtained through interaural time difference. Finally visual information by a camera helps get finer tuning of the angles to speaker. Experimental results of the real-time localization system show that the performance improves to 99.6% compared to the rate of 88.8% when only the audio information is used.

  • PDF

Human-Robot Interaction in Real Environments by Audio-Visual Integration

  • Kim, Hyun-Don;Choi, Jong-Suk;Kim, Mun-Sang
    • International Journal of Control, Automation, and Systems
    • /
    • 제5권1호
    • /
    • pp.61-69
    • /
    • 2007
  • In this paper, we developed not only a reliable sound localization system including a VAD(Voice Activity Detection) component using three microphones but also a face tracking system using a vision camera. Moreover, we proposed a way to integrate three systems in the human-robot interaction to compensate errors in the localization of a speaker and to reject unnecessary speech or noise signals entering from undesired directions effectively. For the purpose of verifying our system's performances, we installed the proposed audio-visual system in a prototype robot, called IROBAA(Intelligent ROBot for Active Audition), and demonstrated how to integrate the audio-visual system.

실내 물류 환경에서 라이다-카메라 약결합 기반 맵핑 및 위치인식과 네비게이션 방법 (Loosely Coupled LiDAR-visual Mapping and Navigation of AMR in Logistic Environments)

  • 최병희;강경수;노예진;조영근
    • 로봇학회논문지
    • /
    • 제17권4호
    • /
    • pp.397-406
    • /
    • 2022
  • This paper presents an autonomous mobile robot (AMR) system and operation algorithms for logistic and factory facilities without magnet-lines installation. Unlike widely used AMR systems, we propose an EKF-based loosely coupled fusion of LiDAR measurements and visual markers. Our method first constructs occupancy grid and visual marker map in the mapping process and utilizes prebuilt maps for precise localization. Also, we developed a waypoint-based navigation pipeline for robust autonomous operation in unconstrained environments. The proposed system estimates the robot pose using by updating the state with the fusion of visual marker and LiDAR measurements. Finally, we tested the proposed method in indoor environments and existing factory facilities for evaluation. In experimental results, this paper represents the performance of our system compared to the well-known LiDAR-based localization and navigation system.

시각주의 모델을 적용한 실내 복도에서의 위치인식 기법 (An Approach for Localization Around Indoor Corridors Based on Visual Attention Model)

  • 윤국열;최선욱;이종호
    • 제어로봇시스템학회논문지
    • /
    • 제17권2호
    • /
    • pp.93-101
    • /
    • 2011
  • For mobile robot, recognizing its current location is very important to navigate autonomously. Especially, loop closing detection that robot recognize location where it has visited before is a kernel problem to solve localization. A considerable amount of research has been conducted on loop closing detection and localization based on appearance because vision sensor has an advantage in terms of costs and various approaching methods to solve this problem. In case of scenes that consist of repeated structures like in corridors, perceptual aliasing in which, the two different locations are recognized as the same, occurs frequently. In this paper, we propose an improved method to recognize location in the scenes which have similar structures. We extracted salient regions from images using visual attention model and calculated weights using distinctive features in the salient region. It makes possible to emphasize unique features in the scene to classify similar-looking locations. In the results of corridor recognition experiments, proposed method showed improved recognition performance. It shows 78.2% in the accuracy of single floor corridor recognition and 71.5% for multi floor corridors recognition.

무인 시스템의 자율 주행을 위한 영상기반 항법기술 동향 (Survey on Visual Navigation Technology for Unmanned Systems)

  • 김현진;서호성;김표진;이충근
    • 한국항행학회논문지
    • /
    • 제19권2호
    • /
    • pp.133-139
    • /
    • 2015
  • 이 논문에서는 영상정보를 기반으로 한 무인 시스템의 자율 항법기술에 대한 동향을 요약한다. 영상기반 항법기술로는 비주얼 서보잉, 비주얼 오도메트리, 영상 기반 SLAM(simultaneous localization and mapping)이 있다. 비주얼 서보잉은 목표 이미지와 현재 이미지 사이의 피쳐 차이로부터 원하는 속도 입력을 계산하여 무인 로봇을 목표 자세로 유도하는 데 사용된다. 비주얼 오도메트리는 무인 시스템이 영상정보를 바탕으로 자신의 이동 궤적을 추정하는 기술로, 기존의 dead-reckoning 방식보다 정확성을 향상시킬 수 있다. 영상 기반 SLAM은 무인 시스템이 영상 정보를 활용하여 미지의 환경에 대한 지도를 구축함과 동시에 자신의 위치를 결정해 나가는 기술로, 정확히 알지 못하는 환경에서 무인차량이나 무인기를 운용하는데 필수적이다. 이러한 기술들이 적용된 해외의 연구 사례들을 살펴봄으로써 영상기반 항법기술의 동향을 파악할 수 있었다.

국소 집단 최적화 기법을 적용한 비정형 해저면 환경에서의 비주얼 SLAM (Visual SLAM using Local Bundle Optimization in Unstructured Seafloor Environment)

  • 홍성훈;김진환
    • 로봇학회논문지
    • /
    • 제9권4호
    • /
    • pp.197-205
    • /
    • 2014
  • As computer vision algorithms are developed on a continuous basis, the visual information from vision sensors has been widely used in the context of simultaneous localization and mapping (SLAM), called visual SLAM, which utilizes relative motion information between images. This research addresses a visual SLAM framework for online localization and mapping in an unstructured seabed environment that can be applied to a low-cost unmanned underwater vehicle equipped with a single monocular camera as a major measurement sensor. Typically, an image motion model with a predefined dimensionality can be corrupted by errors due to the violation of the model assumptions, which may lead to performance degradation of the visual SLAM estimation. To deal with the erroneous image motion model, this study employs a local bundle optimization (LBO) scheme when a closed loop is detected. The results of comparison between visual SLAM estimation with LBO and the other case are presented to validate the effectiveness of the proposed methodology.

실제 실내 환경에서 이동로봇의 위상학적 위치 추정 (Topological Localization of Mobile Robots in Real Indoor Environment)

  • 박영빈;서일홍;최병욱
    • 로봇학회논문지
    • /
    • 제4권1호
    • /
    • pp.25-33
    • /
    • 2009
  • One of the main problems of topological localization in a real indoor environment is variations in the environment caused by dynamic objects and changes in illumination. Another problem arises from the sense of topological localization itself. Thus, a robot must be able to recognize observations at slightly different positions and angles within a certain topological location as identical in terms of topological localization. In this paper, a possible solution to these problems is addressed in the domain of global topological localization for mobile robots, in which environments are represented by their visual appearance. Our approach is formulated on the basis of a probabilistic model called the Bayes filter. Here, marginalization of dynamics in the environment, marginalization of viewpoint changes in a topological location, and fusion of multiple visual features are employed to measure observations reliably, and action-based view transition model and action-associated topological map are used to predict the next state. We performed experiments to demonstrate the validity of our proposed approach among several standard approaches in the field of topological localization. The results clearly demonstrated the value of our approach.

  • PDF