• 제목/요약/키워드: Hand Pose Estimation

검색결과 23건 처리시간 0.025초

Real-time Human Pose Estimation using RGB-D images and Deep Learning

  • 림빈보니카;성낙준;마준;최유주;홍민
    • 인터넷정보학회논문지
    • /
    • 제21권3호
    • /
    • pp.113-121
    • /
    • 2020
  • Human Pose Estimation (HPE) which localizes the human body joints becomes a high potential for high-level applications in the field of computer vision. The main challenges of HPE in real-time are occlusion, illumination change and diversity of pose appearance. The single RGB image is fed into HPE framework in order to reduce the computation cost by using depth-independent device such as a common camera, webcam, or phone cam. However, HPE based on the single RGB is not able to solve the above challenges due to inherent characteristics of color or texture. On the other hand, depth information which is fed into HPE framework and detects the human body parts in 3D coordinates can be usefully used to solve the above challenges. However, the depth information-based HPE requires the depth-dependent device which has space constraint and is cost consuming. Especially, the result of depth information-based HPE is less reliable due to the requirement of pose initialization and less stabilization of frame tracking. Therefore, this paper proposes a new method of HPE which is robust in estimating self-occlusion. There are many human parts which can be occluded by other body parts. However, this paper focuses only on head self-occlusion. The new method is a combination of the RGB image-based HPE framework and the depth information-based HPE framework. We evaluated the performance of the proposed method by COCO Object Keypoint Similarity library. By taking an advantage of RGB image-based HPE method and depth information-based HPE method, our HPE method based on RGB-D achieved the mAP of 0.903 and mAR of 0.938. It proved that our method outperforms the RGB-based HPE and the depth-based HPE.

HSFE Network and Fusion Model based Dynamic Hand Gesture Recognition

  • Tai, Do Nhu;Na, In Seop;Kim, Soo Hyung
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제14권9호
    • /
    • pp.3924-3940
    • /
    • 2020
  • Dynamic hand gesture recognition(d-HGR) plays an important role in human-computer interaction(HCI) system. With the growth of hand-pose estimation as well as 3D depth sensors, depth, and the hand-skeleton dataset is proposed to bring much research in depth and 3D hand skeleton approaches. However, it is still a challenging problem due to the low resolution, higher complexity, and self-occlusion. In this paper, we propose a hand-shape feature extraction(HSFE) network to produce robust hand-shapes. We build a hand-shape model, and hand-skeleton based on LSTM to exploit the temporal information from hand-shape and motion changes. Fusion between two models brings the best accuracy in dynamic hand gesture (DHG) dataset.

Trinocular Vision System을 이용한 물체 자세정보 인식 향상방안 (A Study on the Improvement of Pose Information of Objects by Using Trinocular Vision System)

  • 김종형;장경재;권혁동
    • 한국생산제조학회지
    • /
    • 제26권2호
    • /
    • pp.223-229
    • /
    • 2017
  • Recently, robotic bin-picking tasks have drawn considerable attention, because flexibility is required in robotic assembly tasks. Generally, stereo camera systems have been used widely for robotic bin-picking, but these have two limitations: First, computational burden for solving correspondence problem on stereo images increases calculation time. Second, errors in image processing and camera calibration reduce accuracy. Moreover, the errors in robot kinematic parameters directly affect robot gripping. In this paper, we propose a method of correcting the bin-picking error by using trinocular vision system which consists of two stereo cameras andone hand-eye camera. First, the two stereo cameras, with wide viewing angle, measure object's pose roughly. Then, the 3rd hand-eye camera approaches the object, and corrects the previous measurement of the stereo camera system. Experimental results show usefulness of the proposed method.

Automatic Registration of Two Parts using Robot with Multiple 3D Sensor Systems

  • Ha, Jong-Eun
    • Journal of Electrical Engineering and Technology
    • /
    • 제10권4호
    • /
    • pp.1830-1835
    • /
    • 2015
  • In this paper, we propose an algorithm for the automatic registration of two rigid parts using multiple 3D sensor systems on a robot. Four sets of structured laser stripe system consisted of a camera and a visible laser stripe is used for the acquisition of 3D information. Detailed procedures including extrinsic calibration among four 3D sensor systems and hand/eye calibration of 3D sensing system on robot arm are presented. We find a best pose using search-based pose estimation algorithm where cost function is proposed by reflecting geometric constraints between sensor systems and target objects. A pose with minimum gap and height difference is found by greedy search. Experimental result using demo system shows the robustness and feasibility of the proposed algorithm.

손동작 인식 기반 Virtual Fitting 개발 (Virtual Fitting Development Based on Hand Gesture Recognition)

  • 김승연;유민지;조하정;정승원
    • 한국정보처리학회:학술대회논문집
    • /
    • 한국정보처리학회 2019년도 춘계학술발표대회
    • /
    • pp.596-598
    • /
    • 2019
  • 손동작 인식을 기반으로 한 Virtual fitting 시스템은 Kinect Sensor 를 사용하여 자연스러운 Fitting 을 구현할 수 있다. Kinect Sensor 를 이용한 Pose estimation, Gesture recognition, Virtual fitting 을 구현함으로써 가상으로 의복을 착용하는 소프트웨어를 소개한다.

3차원 손 특징을 이용한 손 동작 인식에 관한 연구 (A study on hand gesture recognition using 3D hand feature)

  • 배철수
    • 한국정보통신학회논문지
    • /
    • 제10권4호
    • /
    • pp.674-679
    • /
    • 2006
  • 본 논문에서는 3차원 손 특징 데이터를 이용한 동작 인식 시스템을 제안하고자 한다. 제안된 시스템은 3차원 센서에 의해 조밀한 범위의 영상을 생성하여 손 동작에 대한 3차원 특징을 추출하여 손 동작을 분류한다. 또한 다양한 조명과 배경하에서의 손을 견실하게 분할하고 색상 정보와 상관이 없어 수화와 같은 복잡한 손 동작에 대해서도 견실한 인식능력을 나타낼 수가 있다. 제안된 방법의 전체적인 순서는 3차원 영상 획득, 팔 분할, 손과 팔목 분할, 손 자세 추정, 3차원 특징 추출, 그리고 동작 분류로 구성되어 있고, 수화 자세에 대한 인식 실험으로 제안된 시스템의 효율성을 입증하였다.

3차원 골격곡선을 이용한 가상혈관 탐색 방안 (Virtual Navigation of Blood Vessels using 3D Curve-Skeletons)

  • 박상진;박형준
    • 한국CDE학회논문집
    • /
    • 제22권1호
    • /
    • pp.89-99
    • /
    • 2017
  • In order to make a virtual endoscopy system effective for exploring the interior of the 3D model of a human organ, it is necessary to generate an accurate navigation path located inside the 3D model and to obtain consistent camera position and pose estimation along the path. In this paper, we propose an approach to virtual navigation of blood vessels, which makes proper use of orthogonal contours and skeleton curves. The approach generates the orthogonal contours and the skeleton curves from the 3D mesh model and its voxel model, all of which represent the blood vessels. For a navigation zone specified by two nodes on the skeleton curves, it computes the shortest path between the two nodes, estimates the positions and poses of a virtual camera at the nodes in the navigation zone, and interpolates the positions and poses to make the camera move smoothly along the path. In addition to keyboard and mouse input, intuitive hand gestures determined by the Leap Motion SDK are used as user interface for virtual navigation of the blood vessels. The proposed approach provides easy and accurate means for the user to examine the interior of 3D blood vessels without any collisions between the camera and their surface. With a simple user study, we present illustrative examples of applying the approach to 3D mesh models of various blood vessels in order to show its quality and usefulness.

RGB-Depth 카메라와 Deep Convolution Neural Networks 기반의 실시간 사람 양손 3D 포즈 추정 (Real-time 3D Pose Estimation of Both Human Hands via RGB-Depth Camera and Deep Convolutional Neural Networks)

  • 박나현;지용빈;기건;김태연;박혜민;김태성
    • 한국정보처리학회:학술대회논문집
    • /
    • 한국정보처리학회 2018년도 추계학술발표대회
    • /
    • pp.686-689
    • /
    • 2018
  • 3D 손 포즈 추정(Hand Pose Estimation, HPE)은 스마트 인간 컴퓨터 인터페이스를 위해서 중요한 기술이다. 이 연구에서는 딥러닝 방법을 기반으로 하여 단일 RGB-Depth 카메라로 촬영한 양손의 3D 손 자세를 실시간으로 인식하는 손 포즈 추정 시스템을 제시한다. 손 포즈 추정 시스템은 4단계로 구성된다. 첫째, Skin Detection 및 Depth cutting 알고리즘을 사용하여 양손을 RGB와 깊이 영상에서 감지하고 추출한다. 둘째, Convolutional Neural Network(CNN) Classifier는 오른손과 왼손을 구별하는데 사용된다. CNN Classifier 는 3개의 convolution layer와 2개의 Fully-Connected Layer로 구성되어 있으며, 추출된 깊이 영상을 입력으로 사용한다. 셋째, 학습된 CNN regressor는 추출된 왼쪽 및 오른쪽 손의 깊이 영상에서 손 관절을 추정하기 위해 다수의 Convolutional Layers, Pooling Layers, Fully Connected Layers로 구성된다. CNN classifier와 regressor는 22,000개 깊이 영상 데이터셋으로 학습된다. 마지막으로, 각 손의 3D 손 자세는 추정된 손 관절 정보로부터 재구성된다. 테스트 결과, CNN classifier는 오른쪽 손과 왼쪽 손을 96.9%의 정확도로 구별할 수 있으며, CNN regressor는 형균 8.48mm의 오차 범위로 3D 손 관절 정보를 추정할 수 있다. 본 연구에서 제안하는 손 포즈 추정 시스템은 가상 현실(virtual reality, VR), 증강 현실(Augmented Reality, AR) 및 융합 현실 (Mixed Reality, MR) 응용 프로그램을 포함한 다양한 응용 분야에서 사용할 수 있다.

Subjective Evaluation on Perceptual Tracking Errors from Modeling Errors in Model-Based Tracking

  • Rhee, Eun Joo;Park, Jungsik;Seo, Byung-Kuk;Park, Jong-Il
    • IEIE Transactions on Smart Processing and Computing
    • /
    • 제4권6호
    • /
    • pp.407-412
    • /
    • 2015
  • In model-based tracking, an accurate 3D model of a target object or scene is mostly assumed to be known or given in advance, but the accuracy of the model should be guaranteed for accurate pose estimation. In many application domains, on the other hand, end users are not highly distracted by tracking errors from certain levels of modeling errors. In this paper, we examine perceptual tracking errors, which are predominantly caused by modeling errors, on subjective evaluation and compare them to computational tracking errors. We also discuss the tolerance of modeling errors by analyzing their permissible ranges.

실시간 손동작 인식을 위한 동작 평면 추정 (Motion Plane Estimation for Real-Time Hand Motion Recognition)

  • 정승대;장경호;정순기
    • 정보처리학회논문지B
    • /
    • 제16B권5호
    • /
    • pp.347-358
    • /
    • 2009
  • 손동작을 인식하는 연구가 오랫동안 이뤄져 왔지만 대개의 시스템들이 값비싼 깊이 카메라를 사용하거나 여러 개의 카메라를 사용해 분석하는 등 그 비용이 크며 작동이 가능한 작업 공간이 지극히 제한적이었다. 본 논문에서는 가전제품을 원격 제어하기 위한 목적으로 두 개의 회전 모터를 사용해 작업 공간을 확대하고 저렴한 일반 카메라를 사용해서 효율적으로 손동작을 인식하기 위한 시스템을 제안한다. 이 시스템은 입력된 카메라의 자세 정보와 이미지상의 2차원적 손가락 위치 정보를 이용하여 3차원 궤적을 추정하고 이를 동작 평면으로 투영시켜 의미 있는 선형 동작 패턴으로 복원한다. 또한 본 논문에서는 개발된 시스템을 테스트하여 주어진 목적에 맞는 정확도를 가지는 작업 영역을 정의한다.