• 제목/요약/키워드: pose estimation

검색결과 388건 처리시간 0.031초

얼굴 포즈 추정을 이용한 다중 RGB-D 카메라 기반의 2D - 3D 얼굴 인증을 위한 시스템 (2D - 3D Human Face Verification System based on Multiple RGB-D Camera using Head Pose Estimation)

  • 김정민;이성철;김학일
    • 정보보호학회논문지
    • /
    • 제24권4호
    • /
    • pp.607-616
    • /
    • 2014
  • 현재 영상감시 시스템에서 얼굴 인식을 통한 사람의 신원 확인은 정면 얼굴이 아닌 관계로 매우 어려운 기술에 속한다. 일반적인 사람들의 얼굴 영상과 입력된 얼굴 영상을 비교하여 유사도를 파악하고 신원을 확인 하는 기술은 각도의 차이에 따라 정확도의 오차가 심해진다. 이런 문제를 해결하기 위해 본 논문에서는 POSIT을 사용하여 얼굴 포즈 측정을 하고, 추정된 각도를 이용하여 3D 얼굴 영상을 제작 후 매칭 하여 일반적인 정면 영상끼리의 매칭이 아닌 rotated face를 이용한 매칭을 해보기로 한다. 얼굴을 매칭 하는 데는 상용화된 얼굴인식 알고리즘을 사용하였다. 얼굴 포즈 추정은 $10^{\circ}$이내의 오차를 보였고, 얼굴인증 성능은 약 95% 정도임을 확인하였다.

관성 센서를 이용한 휴머노이드 로봇용 3축 자세 추정 알고리듬 개발 (Development of 3-Dimensional Pose Estimation Algorithm using Inertial Sensors for Humanoid Robot)

  • 이아람;김정한
    • 제어로봇시스템학회논문지
    • /
    • 제14권2호
    • /
    • pp.133-140
    • /
    • 2008
  • In this paper, a small and effective attitude estimation system for a humanoid robot was developed. Four small inertial sensors were packed and used for inertial measurements(3D accelerometer and three 1D gyroscopes.) An effective 3D pose estimation algorithm for low cost DSP using an extended Kalman filter was developed and evaluated. The 3D pose estimation algorithm has a very simple structure composed by 3 modules of a linear acceleration estimator, an external acceleration detector and an pseudo-accelerometer output estimator. The algorithm also has an effective switching structure based on probability and simple feedback loop for the extended Kalman filter. A special test equipment using linear motor for the testing of the 3D pose sensor was developed and the experimental results showed its very fast convergence to real values and effective responses. Popular DSP of TMS320F2812 was used to calculate robot's 3D attitude and translated acceleration, and the whole system were packed in a small size for humanoids robots. The output of the 3D sensors(pitch, roll, 3D linear acceleration, and 3D angular rate) can be transmitted to a humanoid robot at 200Hz frequency.

A Kidnapping Detection Using Human Pose Estimation in Intelligent Video Surveillance Systems

  • Park, Ju Hyun;Song, KwangHo;Kim, Yoo-Sung
    • 한국컴퓨터정보학회논문지
    • /
    • 제23권8호
    • /
    • pp.9-16
    • /
    • 2018
  • In this paper, a kidnapping detection scheme in which human pose estimation is used to classify accurately between kidnapping cases and normal ones is proposed. To estimate human poses from input video, human's 10 joint information is extracted by OpenPose library. In addition to the features which are used in the previous study to represent the size change rates and the regularities of human activities, the human pose estimation features which are computed from the location of detected human's joints are used as the features to distinguish kidnapping situations from the normal accompanying ones. A frame-based kidnapping detection scheme is generated according to the selection of J48 decision tree model from the comparison of several representative classification models. When a video has more frames of kidnapping situation than the threshold ratio after two people meet in the video, the proposed scheme detects and notifies the occurrence of kidnapping event. To check the feasibility of the proposed scheme, the detection accuracy of our newly proposed scheme is compared with that of the previous scheme. According to the experiment results, the proposed scheme could detect kidnapping situations more 4.73% correctly than the previous scheme.

포즈 추정 기반 얼굴 인식 시스템 설계 : 포즈 추정 알고리즘 비교 연구 (Design of Face Recognition System Based on Pose Estimation : Comparative Studies of Pose Estimation Algorithms)

  • 김진율;김종범;오성권
    • 전기학회논문지
    • /
    • 제66권4호
    • /
    • pp.672-681
    • /
    • 2017
  • This paper is concerned with the design methodology of face recognition system based on pose estimation. In 2-dimensional face recognition, the variations of facial pose cause the deterioration of recognition performance because object recognition is carried out by using brightness of each pixel on image. To alleviate such problem, the proposed face recognition system deals with Learning Vector Quantizatioin(LVQ) or K-Nearest Neighbor(K-NN) to estimate facial pose on image and then the images obtained from LVQ or K-NN are used as the inputs of networks such as Convolution Neural Networks(CNNs) and Radial Basis Function Neural Networks(RBFNNs). The effectiveness and efficiency of the post estimation using LVQ and K-NN as well as face recognition rate using CNNs and RBFNNs are discussed through experiments carried out by using ICPR and CMU PIE databases.

RGB 이미지 기반 인간 동작 추정을 통한 투구 동작 분석 (Analysis of Pitching Motions by Human Pose Estimation Based on RGB Images)

  • 우영주;주지용;김영관;정희용
    • 스마트미디어저널
    • /
    • 제13권4호
    • /
    • pp.16-22
    • /
    • 2024
  • 투구는 야구의 시작이라 할 만큼 야구에서 주요한 부분을 차지한다. 투구 동작의 정확한 분석은 경기력 향상과 부상 예방 측면에서 매우 중요하다. 올바른 투구 동작을 분석할 때, 현재 주로 사용되는 모션캡처는 환경적으로 치명적인 단점들이 몇 가지 존재한다. 본 논문에서 우리는 이러한 단점들이 존재하는 모션캡처를 대체하기 위하여 RGB 기반의 Human Pose Estimation(HPE) 모델을 활용한 투구 동작의 분석을 제안하며 이에 대한 신뢰도를 검증하기 위해 모션캡처 데이터와 HPE 데이터의 관절 좌표를 Dynamic Time Warping(DTW) 알고리즘의 비교를 통해 두 데이터의 유사도를 검증하였다.

Pose Tracking of Moving Sensor using Monocular Camera and IMU Sensor

  • Jung, Sukwoo;Park, Seho;Lee, KyungTaek
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제15권8호
    • /
    • pp.3011-3024
    • /
    • 2021
  • Pose estimation of the sensor is important issue in many applications such as robotics, navigation, tracking, and Augmented Reality. This paper proposes visual-inertial integration system appropriate for dynamically moving condition of the sensor. The orientation estimated from Inertial Measurement Unit (IMU) sensor is used to calculate the essential matrix based on the intrinsic parameters of the camera. Using the epipolar geometry, the outliers of the feature point matching are eliminated in the image sequences. The pose of the sensor can be obtained from the feature point matching. The use of IMU sensor can help initially eliminate erroneous point matches in the image of dynamic scene. After the outliers are removed from the feature points, these selected feature points matching relations are used to calculate the precise fundamental matrix. Finally, with the feature point matching relation, the pose of the sensor is estimated. The proposed procedure was implemented and tested, comparing with the existing methods. Experimental results have shown the effectiveness of the technique proposed in this paper.

A Multi-Stage Convolution Machine with Scaling and Dilation for Human Pose Estimation

  • Nie, Yali;Lee, Jaehwan;Yoon, Sook;Park, Dong Sun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제13권6호
    • /
    • pp.3182-3198
    • /
    • 2019
  • Vision-based Human Pose Estimation has been considered as one of challenging research subjects due to problems including confounding background clutter, diversity of human appearances and illumination changes in scenes. To tackle these problems, we propose to use a new multi-stage convolution machine for estimating human pose. To provide better heatmap prediction of body joints, the proposed machine repeatedly produces multiple predictions according to stages with receptive field large enough for learning the long-range spatial relationship. And stages are composed of various modules according to their strategic purposes. Pyramid stacking module and dilation module are used to handle problem of human pose at multiple scales. Their multi-scale information from different receptive fields are fused with concatenation, which can catch more contextual information from different features. And spatial and channel information of a given input are converted to gating factors by squeezing the feature maps to a single numeric value based on its importance in order to give each of the network channels different weights. Compared with other ConvNet-based architectures, we demonstrated that our proposed architecture achieved higher accuracy on experiments using standard benchmarks of LSP and MPII pose datasets.

Kinect센서를 이용한 물체 인식 및 자세 추정을 위한 정확도 개선 방법 (A Method for Improving Accuracy of Object Recognition and Pose Estimation by Using Kinect sensor)

  • 김안나;이건규;강기태;김용범;최혁렬
    • 로봇학회논문지
    • /
    • 제10권1호
    • /
    • pp.16-23
    • /
    • 2015
  • This paper presents a method of improving the pose recognition accuracy of objects by using Kinect sensor. First, by using the SURF algorithm, which is one of the most widely used local features point algorithms, we modify inner parameters of the algorithm for efficient object recognition. The proposed method is adjusting the distance between the box filter, modifying Hessian matrix, and eliminating improper key points. In the second, the object orientation is estimated based on the homography. Finally the novel approach of Auto-scaling method is proposed to improve accuracy of object pose estimation. The proposed algorithm is experimentally tested with objects in the plane and its effectiveness is validated.

Pose Estimation of an Object from X-ray Images Based on Principal Axis Analysis

  • Roh, Young-Jun;Cho, Hyung-Suck
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2002년도 ICCAS
    • /
    • pp.97.4-97
    • /
    • 2002
  • 1. Introduction Pose estimation of a three dimensional object has been studied in robot vision area, and it is needed in a number of industrial applications such as process monitoring and control, assembly and PCB inspection. In this research, we propose a new pose estimation method based on principal axes analysis. Here, it is assumed that the locations of x-ray source and the image plane are predetermined and the object geometry is known. To this end, we define a dispersion matrix of an object, which is a discrete form of inertia matrix of the object. It can be determined here from a set of x-ray images, at least three images are required. Then, the pose information is obtained fro...

  • PDF

3차원 자세 추정 기법의 성능 향상을 위한 임의 시점 합성 기반의 고난도 예제 생성 (Hard Example Generation by Novel View Synthesis for 3-D Pose Estimation)

  • 김민지;김성찬
    • 대한임베디드공학회논문지
    • /
    • 제19권1호
    • /
    • pp.9-17
    • /
    • 2024
  • It is widely recognized that for 3D human pose estimation (HPE), dataset acquisition is expensive and the effectiveness of augmentation techniques of conventional visual recognition tasks is limited. We address these difficulties by presenting a simple but effective method that augments input images in terms of viewpoints when training a 3D human pose estimation (HPE) model. Our intuition is that meaningful variants of the input images for HPE could be obtained by viewing a human instance in the images from an arbitrary viewpoint different from that in the original images. The core idea is to synthesize new images that have self-occlusion and thus are difficult to predict at different viewpoints even with the same pose of the original example. We incorporate this idea into the training procedure of the 3D HPE model as an augmentation stage of the input samples. We show that a strategy for augmenting the synthesized example should be carefully designed in terms of the frequency of performing the augmentation and the selection of viewpoints for synthesizing the samples. To this end, we propose a new metric to measure the prediction difficulty of input images for 3D HPE in terms of the distance between corresponding keypoints on both sides of a human body. Extensive exploration of the space of augmentation probability choices and example selection according to the proposed distance metric leads to a performance gain of up to 6.2% on Human3.6M, the well-known pose estimation dataset.