• Title/Summary/Keyword: Camera pose

Search Result 277, Processing Time 0.024 seconds

BIM model-based structural damage localization using visual-inertial odometry

  • Junyeon Chung;Kiyoung Kim;Hoon Sohn
    • Smart Structures and Systems
    • /
    • v.31 no.6
    • /
    • pp.561-571
    • /
    • 2023
  • Ensuring the safety of a structure necessitates that repairs are carried out based on accurate inspections and records of damage information. Traditional methods of recording damage rely on individual paper-based documents, making it challenging for inspectors to accurately record damage locations and track chronological changes. Recent research has suggested the adoption of building information modeling (BIM) to record detailed damage information; however, localizing damages on a BIM model can be time-consuming. To overcome this limitation, this study proposes a method to automatically localize damages on a BIM model in real-time, utilizing consecutive images and measurements from an inertial measurement unit in close proximity to damages. The proposed method employs a visual-inertial odometry algorithm to estimate the camera pose, detect damages, and compute the damage location in the coordinate of a prebuilt BIM model. The feasibility and effectiveness of the proposed method were validated through an experiment conducted on a campus building. Results revealed that the proposed method successfully localized damages on the BIM model in real-time, with a root mean square error of 6.6 cm.

Mono-Vision Based Satellite Relative Navigation Using Active Contour Method (능동 윤곽 기법을 적용한 단일 영상 기반 인공위성 상대항법)

  • Kim, Sang-Hyeon;Choi, Han-Lim;Shim, Hyunchul
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.43 no.10
    • /
    • pp.902-909
    • /
    • 2015
  • In this paper, monovision based relative navigation for a satellite proximity operation is studied. The chaser satellite only uses one camera sensor to observe the target satellite and conducts image tracking to obtain the target pose information. However, by using only mono-vision, it is hard to get the depth information which is related to the relative distance to the target. In order to resolve the well-known difficulty in computing the depth information with the use of a single camera, the active contour method is adopted for the image tracking process. The active contour method provides the size of target image, which can be utilized to indirectly calculate the relative distance between the chaser and the target. 3D virtual reality is used in order to model the space environment where two satellites make relative motion and produce the virtual camera images. The unscented Kalman filter is used for the chaser satellite to estimate the relative position of the target in the process of glideslope approaching. Closed-loop simulations are conducted to analyze the performance of the relative navigation with the active contour method.

Estimation of Manhattan Coordinate System using Convolutional Neural Network (합성곱 신경망 기반 맨하탄 좌표계 추정)

  • Lee, Jinwoo;Lee, Hyunjoon;Kim, Junho
    • Journal of the Korea Computer Graphics Society
    • /
    • v.23 no.3
    • /
    • pp.31-38
    • /
    • 2017
  • In this paper, we propose a system which estimates Manhattan coordinate systems for urban scene images using a convolutional neural network (CNN). Estimating the Manhattan coordinate system from an image under the Manhattan world assumption is the basis for solving computer graphics and vision problems such as image adjustment and 3D scene reconstruction. We construct a CNN that estimates Manhattan coordinate systems based on GoogLeNet [1]. To train the CNN, we collect about 155,000 images under the Manhattan world assumption by using the Google Street View APIs and calculate Manhattan coordinate systems using existing calibration methods to generate dataset. In contrast to PoseNet [2] that trains per-scene CNNs, our method learns from images under the Manhattan world assumption and thus estimates Manhattan coordinate systems for new images that have not been learned. Experimental results show that our method estimates Manhattan coordinate systems with the median error of $3.157^{\circ}$ for the Google Street View images of non-trained scenes, as test set. In addition, compared to an existing calibration method [3], the proposed method shows lower intermediate errors for the test set.

A Study on the Estimation of Multi-Object Social Distancing Using Stereo Vision and AlphaPose (Stereo Vision과 AlphaPose를 이용한 다중 객체 거리 추정 방법에 관한 연구)

  • Lee, Ju-Min;Bae, Hyeon-Jae;Jang, Gyu-Jin;Kim, Jin-Pyeong
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.10 no.7
    • /
    • pp.279-286
    • /
    • 2021
  • Recently, We are carrying out a policy of physical distancing of at least 1m from each other to prevent the spreading of COVID-19 disease in public places. In this paper, we propose a method for measuring distances between people in real time and an automation system that recognizes objects that are within 1 meter of each other from stereo images acquired by drones or CCTVs according to the estimated distance. A problem with existing methods used to estimate distances between multiple objects is that they do not obtain three-dimensional information of objects using only one CCTV. his is because three-dimensional information is necessary to measure distances between people when they are right next to each other or overlap in two dimensional image. Furthermore, they use only the Bounding Box information to obtain the exact coordinates of human existence. Therefore, in this paper, to obtain the exact two-dimensional coordinate value in which a person exists, we extract a person's key point to detect the location, convert it to a three-dimensional coordinate value using Stereo Vision and Camera Calibration, and estimate the Euclidean distance between people. As a result of performing an experiment for estimating the accuracy of 3D coordinates and the distance between objects (persons), the average error within 0.098m was shown in the estimation of the distance between multiple people within 1m.

A Home-Based Remote Rehabilitation System with Motion Recognition for Joint Range of Motion Improvement (관절 가동범위 향상을 위한 원격 모션 인식 재활 시스템)

  • Kim, Kyungah;Chung, Wan-Young
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.20 no.3
    • /
    • pp.151-158
    • /
    • 2019
  • Patients with disabilities from various reasons such as disasters, injuries or chronic illness or elderly with limited body motion range due to aging are recommended to participate in rehabilitation programs at hospitals. But typically, it's not as simple for them to commute without help as they have limited access outside of the home. Also, regarding the perspectives of hospitals, having to maintain the workforce and have them take care of the rehabilitation sessions leads them to more expenses in cost aspects. For those reasons, in this paper, a home-based remote rehabilitation system using motion recognition is developed without needing help from others. This system can be executed by a personal computer and a stereo camera at home, the real-time user motion status is monitored using motion recognition feature. The system tracks the joint range of motion(Joint ROM) of particular body parts of users to check the body function improvement. For demonstration, total of 4 subjects with various ages and health conditions participated in this project. Their motion data were collected during all 3 exercise sessions, and each session was repeated 9 times per person and was compared in the results.

Spectrum Analysis and Detection of Ships Based on Aerial Hyperspectral Remote Sensing Experiments (항공 초분광 원격탐사 실험 기반 선박 스펙트럼 분석 및 탐지)

  • Jae-Jin Park;Kyung-Ae Park;Tae-Sung Kim;Moonjin Lee
    • Journal of the Korean earth science society
    • /
    • v.45 no.3
    • /
    • pp.214-223
    • /
    • 2024
  • The recent increase in maritime traffic and coastal leisure activities has led to a rise in various marine accidents. These incidents not only result in damage to human life and property but also pose a significant risk of marine pollution involving oil and hazardous and noxious substances (HNS) spills. Therefore, effective ship monitoring is crucial for preparing and for responding to marine accidents. This study conducted an aerial experiment utilizing hyperspectral remote sensing to develop a maritime ship monitoring system. Hyperspectral aerial measurements were carried out around Gungpyeong Port in the western coastal region of the Korean Peninsula, and spectral libraries were constructed for various ship decks. The spectral correlation similarity (SCS) technique was employed for ship detection, analyzing the spatial similarity distribution between hyperspectral images and ship spectra. As a result, 15 ships were detected in the hyperspectral images. The color of each ship's deck was classified based on the highest spectral similarity. The detected ships were verified by matching them with high-resolution digital mapping camera (DMC) images. This foundational study on the application of aerial hyperspectral sensors for maritime ship detection demonstrates their potential role in future remote sensing-based ship monitoring systems.

Geometric Correction of Lips Using Lip Information (입술정보를 이용한 입술모양의 기하학적 보정)

  • 황동국;박희정;전병민
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.6C
    • /
    • pp.834-841
    • /
    • 2004
  • There can be lips transformed geometrically in the lip images according to the location or the pose of camera and speaker. This transformation of the lip images changes geometric information of original lip phases. Therefore, for enhancing global lip information by using partial information of lips to correct lip phases transformed geometrically, in this paper we propose a method that can geometrically correct lips. The method is composed of two steps - the feature-deciding step and the correcting step. In the former, it is for us to extract key points and features of source image according to the its lip model and to create that of target image according to the its lip model. In the latter, we decide mapping relation after partition a source and target image based on information extracted in the previous step into each 4 regions. and then, after mapping, we unite corrected sub-images to a result image. As experiment image, we use fames that contain pronunciation on short vowels of the Korean language and use lip symmetry for evaluating the proposed algorithm. In experiment result, the correcting rate of the lower lip than the upper lip and that of lips moving largely than little was highly enhanced.

Vision-based Sensor Fusion of a Remotely Operated Vehicle for Underwater Structure Diagnostication (수중 구조물 진단용 원격 조종 로봇의 자세 제어를 위한 비전 기반 센서 융합)

  • Lee, Jae-Min;Kim, Gon-Woo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.21 no.4
    • /
    • pp.349-355
    • /
    • 2015
  • Underwater robots generally show better performances for tasks than humans under certain underwater constraints such as. high pressure, limited light, etc. To properly diagnose in an underwater environment using remotely operated underwater vehicles, it is important to keep autonomously its own position and orientation in order to avoid additional control efforts. In this paper, we propose an efficient method to assist in the operation for the various disturbances of a remotely operated vehicle for the diagnosis of underwater structures. The conventional AHRS-based bearing estimation system did not work well due to incorrect measurements caused by the hard-iron effect when the robot is approaching a ferromagnetic structure. To overcome this drawback, we propose a sensor fusion algorithm with the camera and AHRS for estimating the pose of the ROV. However, the image information in the underwater environment is often unreliable and blurred by turbidity or suspended solids. Thus, we suggest an efficient method for fusing the vision sensor and the AHRS with a criterion which is the amount of blur in the image. To evaluate the amount of blur, we adopt two methods: one is the quantification of high frequency components using the power spectrum density analysis of 2D discrete Fourier transformed image, and the other is identifying the blur parameter based on cepstrum analysis. We evaluate the performance of the robustness of the visual odometry and blur estimation methods according to the change of light and distance. We verify that the blur estimation method based on cepstrum analysis shows a better performance through the experiments.

Three Dimensional Tracking of Road Signs based on Stereo Vision Technique (스테레오 비전 기술을 이용한 도로 표지판의 3차원 추적)

  • Choi, Chang-Won;Choi, Sung-In;Park, Soon-Yong
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.20 no.12
    • /
    • pp.1259-1266
    • /
    • 2014
  • Road signs provide important safety information about road and traffic conditions to drivers. Road signs include not only common traffic signs but also warning information regarding unexpected obstacles and road constructions. Therefore, accurate detection and identification of road signs is one of the most important research topics related to safe driving. In this paper, we propose a 3-D vision technique to automatically detect and track road signs in a video sequence which is acquired from a stereo vision camera mounted on a vehicle. First, color information is used to initially detect the sign candidates. Second, the SVM (Support Vector Machine) is employed to determine true signs from the candidates. Once a road sign is detected in a video frame, it is continuously tracked from the next frame until it is disappeared. The 2-D position of a detected sign in the next frame is predicted by the 3-D motion of the vehicle. Here, the 3-D vehicle motion is acquired by using the 3-D pose information of the detected sign. Finally, the predicted 2-D position is corrected by template-matching of the scaled template of the detected sign within a window area around the predicted position. Experimental results show that the proposed method can detect and track many types of road signs successfully. Tracking comparisons with two different methods are shown.

Adaptive Keyframe-Based Tracking for Augmented Books (증강 책을 위한 적응형 키프레임 기반 트래킹)

  • Yoo, Jae-Sang;Cho, Kyu-Sung;Yang, Hyun-S.
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.4
    • /
    • pp.502-506
    • /
    • 2010
  • An augmented book is an application that augments such multimedia elements as virtual 3D objects generated by computer graphics, movie clips, or sound clips to a real book using AR technologies. It is intended to bring additional education and entertainment effects to users. For augmented books, this paper proposes an adaptive keyframe-based page tracking method to estimate the camera's 6 DOF pose in real-time after recognizing a page and performing wide-baseline keypoint matching. For a page tracking, proposed method in this paper chooses a proper keyframe and performs a tracking in two step of coarse-to-fine stage. As a result, the proposed method in this paper guarantees a robust tracking to view-point and illumination variations and real-time.