• Title/Summary/Keyword: Robot Gesture Generation

Search Result 7, Processing Time 0.048 seconds

Survey: Gesture Recognition Techniques for Intelligent Robot (지능형 로봇 구동을 위한 제스처 인식 기술 동향)

  • Oh Jae-Yong;Lee Chil-Woo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.10 no.9
    • /
    • pp.771-778
    • /
    • 2004
  • Recently, various applications of robot system become more popular in accordance with rapid development of computer hardware/software, artificial intelligence, and automatic control technology. Formerly robots mainly have been used in industrial field, however, nowadays it is said that the robot will do an important role in the home service application. To make the robot more useful, we require further researches on implementation of natural communication method between the human and the robot system, and autonomous behavior generation. The gesture recognition technique is one of the most convenient methods for natural human-robot interaction, so it is to be solved for implementation of intelligent robot system. In this paper, we describe the state-of-the-art of advanced gesture recognition technologies for intelligent robots according to three methods; sensor based method, feature based method, appearance based method, and 3D model based method. And we also discuss some problems and real applications in the research field.

An Emotional Gesture-based Dialogue Management System using Behavior Network (행동 네트워크를 이용한 감정형 제스처 기반 대화 관리 시스템)

  • Yoon, Jong-Won;Lim, Sung-Soo;Cho, Sung-Bae
    • Journal of KIISE:Software and Applications
    • /
    • v.37 no.10
    • /
    • pp.779-787
    • /
    • 2010
  • Since robots have been used widely recently, research about human-robot communication is in process actively. Typically, natural language processing or gesture generation have been applied to human-robot interaction. However, existing methods for communication among robot and human have their limits in performing only static communication, thus the method for more natural and realistic interaction is required. In this paper, an emotional gesture based dialogue management system is proposed for sophisticated human-robot communication. The proposed system performs communication by using the Bayesian networks and pattern matching, and generates emotional gestures of robots in real-time while the user communicates with the robot. Through emotional gestures robot can communicate the user more efficiently also realistically. We used behavior networks as the gesture generation method to deal with dialogue situations which change dynamically. Finally, we designed a usability test to confirm the usefulness of the proposed system by comparing with the existing dialogue system.

Automatic Generation of Script-Based Robot Gesture and its Application to Steward Robot (스크립트 기반의 로봇 제스처 자동생성 방법 및 집사로봇에의 적용)

  • Kim, Heon-Hui;Lee, Hyong-Euk;Kim, Yong-Hwi;Park, Kwang-Hyun;Bien, Zeung-Nam
    • 한국HCI학회:학술대회논문집
    • /
    • 2007.02a
    • /
    • pp.688-693
    • /
    • 2007
  • 본 논문은 인간과 로봇간의 효과적인 상호작용을 위한 로봇제스쳐의 자동생성 기법을 다룬다. 이는 텍스트 정보 만의 입력으로 의미 있는 단어에 대응되는 특정 제스쳐패턴이 자동적으로 생성되도록 하는 기법으로서 이를 위한 사전조사로 제스쳐가 출현하는 발화시점에서의 단어수집이 우선적으로 요구되었다. 본 논문은 이러한 분석을 위해 두 개 이상의 연속된 제스쳐 패턴을 효과적으로 표현할 수 있는 제스쳐 모델을 제안한다. 또한 제안된 모델이 적용되어 구축된 제스쳐DB와 스크립트 기법을 이용한 로봇제스쳐 자동생성 방법을 제안한다. 제스쳐 생성시스템은 규칙기반의 제스쳐 선택부와 스크립트 기반의 동작 계획부로 구성되고, 집사로봇의 안내기능에 대한 모의실험을 통해 그 효용성을 확인한다.

  • PDF

The Emotional Boundary Decision in a Linear Affect-Expression Space for Effective Robot Behavior Generation (효과적인 로봇 행동 생성을 위한 선형의 정서-표정 공간 내 감정 경계의 결정 -비선형의 제스처 동기화를 위한 정서, 표정 공간의 영역 결정)

  • Jo, Su-Hun;Lee, Hui-Sung;Park, Jeong-Woo;Kim, Min-Gyu;Chung, Myung-Jin
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02a
    • /
    • pp.540-546
    • /
    • 2008
  • In the near future, robots should be able to understand human's emotional states and exhibit appropriate behaviors accordingly. In Human-Human Interaction, the 93% consist of the speaker's nonverbal communicative behavior. Bodily movements provide information of the quantity of emotion. Latest personal robots can interact with human using multi-modality such as facial expression, gesture, LED, sound, sensors and so on. However, a posture needs a position and an orientation only and in facial expression or gesture, movements are involved. Verbal, vocal, musical, color expressions need time information. Because synchronization among multi-modalities is a key problem, emotion expression needs a systematic approach. On the other hand, at low intensity of surprise, the face could be expressed but the gesture could not be expressed because a gesture is not linear. It is need to decide the emotional boundaries for effective robot behavior generation and synchronization with another expressible method. If it is so, how can we define emotional boundaries? And how can multi-modality be synchronized each other?

  • PDF

Implementation of a Spring Backboned Soft Arm Emulating Human Gestures (인간 동작 표현용 스프링 백본 구조 소프트 암의 구현)

  • Yoon, Hyun-Soo;Choi, Jae-Yeon;Oh, Se-Min;Lee, Byeong-Ju;Yoon, Ho-Sup;Cho, Young-Jo
    • The Journal of Korea Robotics Society
    • /
    • v.7 no.2
    • /
    • pp.65-75
    • /
    • 2012
  • This study deals with the design of a spring backboned soft arm, which will be employed for generation of human gesture as an effective means of Human Robot interaction. The special features of the proposed mechanism are the light weight and the flexibility of the whole mechanism by using a spring backbone. Thus, even in the case of collision with human, this device is able to absorb the impact structurally. The kinematics and the design for the soft arm are introduced. The performance of this mechanism was shown through experiment emulating several human gestures expressing human emotion and some service contents. Finally, this soft arm was implemented as the wing mechanism of a penguin robot.

Rotation Invariant 3D Star Skeleton Feature Extraction (회전무관 3D Star Skeleton 특징 추출)

  • Chun, Sung-Kuk;Hong, Kwang-Jin;Jung, Kee-Chul
    • Journal of KIISE:Software and Applications
    • /
    • v.36 no.10
    • /
    • pp.836-850
    • /
    • 2009
  • Human posture recognition has attracted tremendous attention in ubiquitous environment, performing arts and robot control so that, recently, many researchers in pattern recognition and computer vision are working to make efficient posture recognition system. However the most of existing studies is very sensitive to human variations such as the rotation or the translation of body. This is why the feature, which is extracted from the feature extraction part as the first step of general posture recognition system, is influenced by these variations. To alleviate these human variations and improve the posture recognition result, this paper presents 3D Star Skeleton and Principle Component Analysis (PCA) based feature extraction methods in the multi-view environment. The proposed system use the 8 projection maps, a kind of depth map, as an input data. And the projection maps are extracted from the visual hull generation process. Though these data, the system constructs 3D Star Skeleton and extracts the rotation invariant feature using PCA. In experimental result, we extract the feature from the 3D Star Skeleton and recognize the human posture using the feature. Finally we prove that the proposed method is robust to human variations.