• 제목/요약/키워드: HRI(Human Robot Interaction)

검색결과 77건 처리시간 0.022초

컨볼루셔널 신경망과 케스케이드 안면 특징점 검출기를 이용한 얼굴의 특징점 분류 (Facial Point Classifier using Convolution Neural Network and Cascade Facial Point Detector)

  • 유제훈;고광은;심귀보
    • 제어로봇시스템학회논문지
    • /
    • 제22권3호
    • /
    • pp.241-246
    • /
    • 2016
  • Nowadays many people have an interest in facial expression and the behavior of people. These are human-robot interaction (HRI) researchers utilize digital image processing, pattern recognition and machine learning for their studies. Facial feature point detector algorithms are very important for face recognition, gaze tracking, expression, and emotion recognition. In this paper, a cascade facial feature point detector is used for finding facial feature points such as the eyes, nose and mouth. However, the detector has difficulty extracting the feature points from several images, because images have different conditions such as size, color, brightness, etc. Therefore, in this paper, we propose an algorithm using a modified cascade facial feature point detector using a convolutional neural network. The structure of the convolution neural network is based on LeNet-5 of Yann LeCun. For input data of the convolutional neural network, outputs from a cascade facial feature point detector that have color and gray images were used. The images were resized to $32{\times}32$. In addition, the gray images were made into the YUV format. The gray and color images are the basis for the convolution neural network. Then, we classified about 1,200 testing images that show subjects. This research found that the proposed method is more accurate than a cascade facial feature point detector, because the algorithm provides modified results from the cascade facial feature point detector.

휴머노이드 로봇을 위한 사람 검출, 추적 및 실루엣 추출 시스템 (Human Tracking and Body Silhouette Extraction System for Humanoid Robot)

  • 곽수영;변혜란
    • 한국통신학회논문지
    • /
    • 제34권6C호
    • /
    • pp.593-603
    • /
    • 2009
  • 본 논문은 스테레오 카메라가 이동하는 환경에서 카메라 움직임을 보정하여 새로운 다수의 사람을 검출하는 방법과 검출된 사람을 추적하고, 실루엣을 추출하는 통합된 시스템을 제안한다. 제안하는 시스템은 사람 검출, 추적, 실루엣 추출 3가지 모듈로 구성되어 있으며 3가지 모듈은 카메라가 이동하는 환경을 고려한 것이다. 사람 검출 모듈에서는 카메라 움직임(egomotion) 보정을 이용한 움직이는 영역 추출 결과와 스테레오 정보를 결합하여 움직이는 객체를 검출하였으며, 추적모듈은 변위 정보가 가중된 히스토그램 알고리즘으로 검출된 객체를 추적한다. 실루엣을 추출하는 모듈은 트라이맵(trimap)을 이용하여 사람의 실루엣 부분을 대략적으로 추정하는 단계와 그래프컷(graph cut)을 적용하여 정교하게 실루엣 추출하는 단계로 이루어져 있다. 본 논문에서 제안하는 방법을 실내 환경에서 팬-틸트(pan-tilt) 스테레오 카메라로 획득한 실험데이터를 대상으로 실험한 결과 다수의 사람의 검출 및 추적, 정교한 실루엣 추출이 가능한 것을 확인하였다. 본 논문의 실루엣 추출결과는 제스처 인식이나 걸음걸이 인식 등의 다양한 분야에도 적용가능하다.

Mirror Neuron System 계산 모델을 이용한 모방학습 기반 인간-로봇 인터페이스에 관한 연구 (A Study on Human-Robot Interface based on Imitative Learning using Computational Model of Mirror Neuron System)

  • 고광은;심귀보
    • 한국지능시스템학회논문지
    • /
    • 제23권6호
    • /
    • pp.565-570
    • /
    • 2013
  • 영장류 대뇌 피질 영역 중 거울 뉴런들이 분포한 것으로 추정되는 몇몇 영역은 목적성 행위에 대한 시각 정보를 기반으로 모방학습을 수행함으로써 관측 행동의 의도 인식 기능을 담당한다고 알려졌다. 본 논문은 이러한 거울 뉴런 영역을 모델링 하여 인간-로봇 상호작용 시스템에 적용함으로써, 자동화 된 의도인식 시스템을 개발하고자 한다. 거울 뉴런 시스템 계산 모델은 동적 신경망을 기반으로 구축하였으며, 모델의 입력은 객체와 행위자 동작에 대한 연속된 특징 벡터 집합이고 모델의 모방학습 및 추론과정을 통해 관측자가 수행할 수 있는 움직임 정보를 출력한다. 이를 위해 제한된 실험 공간 내에서 특정 객체와 그에 대한 행위자의 목적성 행동, 즉 의도에 대한 시나리오를 전제로 키넥트 센서를 통해 모델 입력 데이터를 수집하고 가상 로봇 시뮬레이션 환경에서 대응하는 움직임 정보를 계산하여 동작을 수행하는 프레임워크를 개발하였다.

가상 운전 인터페이스를 활용한 자동차 협력 원격조종 (Shared Vehicle Teleoperation using a Virtual Driving Interface)

  • 김재석;이광현;유지환
    • 제어로봇시스템학회논문지
    • /
    • 제21권3호
    • /
    • pp.243-249
    • /
    • 2015
  • In direct vehicle teleoperation, a human operator drives a vehicle at a distance through a pair of master and slave device. However, if there is time delay, it is difficult to remotely drive the vehicle due to slow response. In order to address this problem, we introduced a novel methodology of shared vehicle teleoperation using a virtual driving interface. The methodology was developed with four components: 1) virtual driving environment, 2) interface for virtual driving environment, 3) path generator based on virtual driving trajectory, 4) path following controller. Experimental results showed the effectiveness of the proposed approach in simple and cluttered driving environment as well. In the experiments, we compared two sampling methods, fixed sampling time and user defined instant, and finally merged method showed best remote driving performance in term of completion time and number of collision.

다중크기와 다중객체의 실시간 얼굴 검출과 머리 자세 추정을 위한 심층 신경망 (Multi-Scale, Multi-Object and Real-Time Face Detection and Head Pose Estimation Using Deep Neural Networks)

  • 안병태;최동걸;권인소
    • 로봇학회논문지
    • /
    • 제12권3호
    • /
    • pp.313-321
    • /
    • 2017
  • One of the most frequently performed tasks in human-robot interaction (HRI), intelligent vehicles, and security systems is face related applications such as face recognition, facial expression recognition, driver state monitoring, and gaze estimation. In these applications, accurate head pose estimation is an important issue. However, conventional methods have been lacking in accuracy, robustness or processing speed in practical use. In this paper, we propose a novel method for estimating head pose with a monocular camera. The proposed algorithm is based on a deep neural network for multi-task learning using a small grayscale image. This network jointly detects multi-view faces and estimates head pose in hard environmental conditions such as illumination change and large pose change. The proposed framework quantitatively and qualitatively outperforms the state-of-the-art method with an average head pose mean error of less than $4.5^{\circ}$ in real-time.

안정적인 실시간 얼굴 특징점 추적과 감정인식 응용 (Robust Real-time Tracking of Facial Features with Application to Emotion Recognition)

  • 안병태;김응희;손진훈;권인소
    • 로봇학회논문지
    • /
    • 제8권4호
    • /
    • pp.266-272
    • /
    • 2013
  • Facial feature extraction and tracking are essential steps in human-robot-interaction (HRI) field such as face recognition, gaze estimation, and emotion recognition. Active shape model (ASM) is one of the successful generative models that extract the facial features. However, applying only ASM is not adequate for modeling a face in actual applications, because positions of facial features are unstably extracted due to limitation of the number of iterations in the ASM fitting algorithm. The unaccurate positions of facial features decrease the performance of the emotion recognition. In this paper, we propose real-time facial feature extraction and tracking framework using ASM and LK optical flow for emotion recognition. LK optical flow is desirable to estimate time-varying geometric parameters in sequential face images. In addition, we introduce a straightforward method to avoid tracking failure caused by partial occlusions that can be a serious problem for tracking based algorithm. Emotion recognition experiments with k-NN and SVM classifier shows over 95% classification accuracy for three emotions: "joy", "anger", and "disgust".

바디 제스처 인식을 위한 기초적 신체 모델 인코딩과 선택적 / 비동시적 입력을 갖는 병렬 상태 기계 (Primitive Body Model Encoding and Selective / Asynchronous Input-Parallel State Machine for Body Gesture Recognition)

  • 김주창;박정우;김우현;이원형;정명진
    • 로봇학회논문지
    • /
    • 제8권1호
    • /
    • pp.1-7
    • /
    • 2013
  • Body gesture Recognition has been one of the interested research field for Human-Robot Interaction(HRI). Most of the conventional body gesture recognition algorithms used Hidden Markov Model(HMM) for modeling gestures which have spatio-temporal variabilities. However, HMM-based algorithms have difficulties excluding meaningless gestures. Besides, it is necessary for conventional body gesture recognition algorithms to perform gesture segmentation first, then sends the extracted gesture to the HMM for gesture recognition. This separated system causes time delay between two continuing gestures to be recognized, and it makes the system inappropriate for continuous gesture recognition. To overcome these two limitations, this paper suggests primitive body model encoding, which performs spatio/temporal quantization of motions from human body model and encodes them into predefined primitive codes for each link of a body model, and Selective/Asynchronous Input-Parallel State machine(SAI-PSM) for multiple-simultaneous gesture recognition. The experimental results showed that the proposed gesture recognition system using primitive body model encoding and SAI-PSM can exclude meaningless gestures well from the continuous body model data, while performing multiple-simultaneous gesture recognition without losing recognition rates compared to the previous HMM-based work.