• 제목/요약/키워드: pose estimation

검색결과 388건 처리시간 0.027초

HigherHRNet 기반의 발추정 기법을 통한 횡단보도 보행자 인식 (Pedestrian Recognition of Crosswalks Using Foot Estimation Techniques Based on HigherHRNet)

  • 정경민;한주훈;이현
    • 대한임베디드공학회논문지
    • /
    • 제16권5호
    • /
    • pp.171-177
    • /
    • 2021
  • It is difficult to accurately extract features of pedestrian because the pedestrian is photographed at a crosswalk using a camera positioned higher than the pedestrian. In addition, it is more difficult to extract features when a part of the pedestrian's body is covered by an umbrella or parasol or when the pedestrian is holding an object. Representative methods to solve this problem include Object Detection, Instance Segmentation, and Pose Estimation. Among them, this study intends to use the Pose Estimation method. In particular, we intend to increase the recognition rate of pedestrians in crosswalks by maintaining the image resolution through HigherHRNet and applying the foot estimation technique. Finally, we show the superiority of the proposed method by applying and analyzing several data sets covered by body parts to the existing method and the proposed method.

Markerless camera pose estimation framework utilizing construction material with standardized specification

  • Harim Kim;Heejae Ahn;Sebeen Yoon;Taehoon Kim;Thomas H.-K. Kang;Young K. Ju;Minju Kim;Hunhee Cho
    • Computers and Concrete
    • /
    • 제33권5호
    • /
    • pp.535-544
    • /
    • 2024
  • In the rapidly advancing landscape of computer vision (CV) technology, there is a burgeoning interest in its integration with the construction industry. Camera calibration is the process of deriving intrinsic and extrinsic parameters that affect when the coordinates of the 3D real world are projected onto the 2D plane, where the intrinsic parameters are internal factors of the camera, and extrinsic parameters are external factors such as the position and rotation of the camera. Camera pose estimation or extrinsic calibration, which estimates extrinsic parameters, is essential information for CV application at construction since it can be used for indoor navigation of construction robots and field monitoring by restoring depth information. Traditionally, camera pose estimation methods for cameras relied on target objects such as markers or patterns. However, these methods, which are marker- or pattern-based, are often time-consuming due to the requirement of installing a target object for estimation. As a solution to this challenge, this study introduces a novel framework that facilitates camera pose estimation using standardized materials found commonly in construction sites, such as concrete forms. The proposed framework obtains 3D real-world coordinates by referring to construction materials with certain specifications, extracts the 2D coordinates of the corresponding image plane through keypoint detection, and derives the camera's coordinate through the perspective-n-point (PnP) method which derives the extrinsic parameters by matching 3D and 2D coordinate pairs. This framework presents a substantial advancement as it streamlines the extrinsic calibration process, thereby potentially enhancing the efficiency of CV technology application and data collection at construction sites. This approach holds promise for expediting and optimizing various construction-related tasks by automating and simplifying the calibration procedure.

사각형 특징 기반 Visual SLAM을 위한 자세 추정 방법 (A Camera Pose Estimation Method for Rectangle Feature based Visual SLAM)

  • 이재민;김곤우
    • 로봇학회논문지
    • /
    • 제11권1호
    • /
    • pp.33-40
    • /
    • 2016
  • In this paper, we propose a method for estimating the pose of the camera using a rectangle feature utilized for the visual SLAM. A warped rectangle feature as a quadrilateral in the image by the perspective transformation is reconstructed by the Coupled Line Camera algorithm. In order to fully reconstruct a rectangle in the real world coordinate, the distance between the features and the camera is needed. The distance in the real world coordinate can be measured by using a stereo camera. Using properties of the line camera, the physical size of the rectangle feature can be induced from the distance. The correspondence between the quadrilateral in the image and the rectangle in the real world coordinate can restore the relative pose between the camera and the feature through obtaining the homography. In order to evaluate the performance, we analyzed the result of proposed method with its reference pose in Gazebo robot simulator.

스테레오 영상을 이용한 3차원 포즈 추정 (3D Head Pose Estimation Using The Stereo Image)

  • 양욱일;송환종;이용욱;손광훈
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2003년도 하계종합학술대회 논문집 Ⅳ
    • /
    • pp.1887-1890
    • /
    • 2003
  • This paper presents a three-dimensional (3D) head pose estimation algorithm using the stereo image. Given a pair of stereo image, we automatically extract several important facial feature points using the disparity map, the gabor filter and the canny edge detector. To detect the facial feature region , we propose a region dividing method using the disparity map. On the indoor head & shoulder stereo image, a face region has a larger disparity than a background. So we separate a face region from a background by a divergence of disparity. To estimate 3D head pose, we propose a 2D-3D Error Compensated-SVD (EC-SVD) algorithm. We estimate the 3D coordinates of the facial features using the correspondence of a stereo image. We can estimate the head pose of an input image using Error Compensated-SVD (EC-SVD) method. Experimental results show that the proposed method is capable of estimating pose accurately.

  • PDF

RBFNNs 패턴분류기와 객체 추적 알고리즘을 이용한 얼굴인식 및 추적 시스템 설계 (Design of Face Recognition and Tracking System by Using RBFNNs Pattern Classifier with Object Tracking Algorithm)

  • 오승훈;오성권;김진율
    • 전기학회논문지
    • /
    • 제64권5호
    • /
    • pp.766-778
    • /
    • 2015
  • In this paper, we design a hybrid system for recognition and tracking realized with the aid of polynomial based RBFNNs pattern classifier and particle filter. The RBFNN classifier is built by learning the training data for diverse pose images. The optimized parameters of RBFNN classifier are obtained by Particle Swarm Optimization(PSO). Testing data for pose image is used as a face image obtained under real situation, where the face image is detected by AdaBoost algorithm. In order to improve the recognition performance for a detected image, pose estimation as preprocessing step is carried out before the face recognition step. PCA is used for pose estimation, the pose of detected image is assigned for the built pose by considering the featured difference between the previously built pose image and the newly detected image. The recognition of detected image is performed through polynomial based RBFNN pattern classifier, and if the detected image is equal to target for tracking, the target will be traced by particle filter in real time. Moreover, when tracking is failed by PF, Adaboost algorithm detects facial area again, and the procedures of both the pose estimation and the image recognition are repeated as mentioned above. Finally, experimental results are compared and analyzed by using Honda/UCSD data known as benchmark DB.

빈피킹을 위한 스테레오 비전 기반의 제품 라벨의 3차원 자세 추정 (Stereo Vision-Based 3D Pose Estimation of Product Labels for Bin Picking)

  • 우다야 위제나야카;최성인;박순용
    • 제어로봇시스템학회논문지
    • /
    • 제22권1호
    • /
    • pp.8-16
    • /
    • 2016
  • In the field of computer vision and robotics, bin picking is an important application area in which object pose estimation is necessary. Different approaches, such as 2D feature tracking and 3D surface reconstruction, have been introduced to estimate the object pose accurately. We propose a new approach where we can use both 2D image features and 3D surface information to identify the target object and estimate its pose accurately. First, we introduce a label detection technique using Maximally Stable Extremal Regions (MSERs) where the label detection results are used to identify the target objects separately. Then, the 2D image features on the detected label areas are utilized to generate 3D surface information. Finally, we calculate the 3D position and the orientation of the target objects using the information of the 3D surface.

A Framework for Real Time Vehicle Pose Estimation based on synthetic method of obtaining 2D-to-3D Point Correspondence

  • Yun, Sergey;Jeon, Moongu
    • 한국정보처리학회:학술대회논문집
    • /
    • 한국정보처리학회 2014년도 춘계학술발표대회
    • /
    • pp.904-907
    • /
    • 2014
  • In this work we present a robust and fast approach to estimate 3D vehicle pose that can provide results under a specific traffic surveillance conditions. Such limitations are expressed by single fixed CCTV camera that is located relatively high above the ground, its pitch axes is parallel to the reference plane and the camera focus assumed to be known. The benefit of our framework that it does not require prior training, camera calibration and does not heavily rely on 3D model shape as most common technics do. Also it deals with a bad shape condition of the objects as we focused on low resolution surveillance scenes. Pose estimation task is presented as PnP problem to solve it we use well known "POSIT" algorithm [1]. In order to use this algorithm at least 4 non coplanar point's correspondence is required. To find such we propose a set of techniques based on model and scene geometry. Our framework can be applied in real time video sequence. Results for estimated vehicle pose are shown in real image scene.

스테레오 영상을 이용한 얼굴 포즈 추정 (Face Pose Estimation using Stereo Image)

  • 소인미;강선경;김영운;이지근;정성태
    • 한국컴퓨터정보학회논문지
    • /
    • 제11권3호
    • /
    • pp.151-159
    • /
    • 2006
  • 본 논문에서는 두 대의 카메라 영상으로부터 얼굴의 포즈를 추정하는 방법을 제안한다. 제안된 방법은 먼저 두 얼굴 영상으로부터 대응되는 눈썹, 눈, 입의 특징점을 추출한 다음, 스테레오 비전의 삼각법에 의해 특징점에 대한 3차원 위치를 계산한다. 그 다음에는 특징점으로 부터 삼각형을 생성하고 그 삼각형에 수직 방향을 계산함으로써 얼굴의 포즈를 계산한다. 계산된 얼굴의 포즈를 3D 얼굴 모델에 적용해 본 결과 본 논문에서 제안된 방법이 정확한 얼굴 포즈를 추정할 수 있음을 알 수 있었다.

  • PDF

로봇 팔을 활용한 정리작업을 위한 물체 자세추정 및 이미지 매칭 (Pose Estimation and Image Matching for Tidy-up Task using a Robot Arm)

  • 박정란;조현준;송재복
    • 로봇학회논문지
    • /
    • 제16권4호
    • /
    • pp.299-305
    • /
    • 2021
  • In this study, the task of robotic tidy-up is to clean the current environment up exactly like a target image. To perform a tidy-up task using a robot, it is necessary to estimate the pose of various objects and to classify the objects. Pose estimation requires the CAD model of an object, but these models of most objects in daily life are not available. Therefore, this study proposes an algorithm that uses point cloud and PCA to estimate the pose of objects without the help of CAD models in cluttered environments. In addition, objects are usually detected using a deep learning-based object detection. However, this method has a limitation in that only the learned objects can be recognized, and it may take a long time to learn. This study proposes an image matching based on few-shot learning and Siamese network. It was shown from experiments that the proposed method can be effectively applied to the robotic tidy-up system, which showed a success rate of 85% in the tidy-up task.