• Title/Summary/Keyword: Human Robot Interaction system

Search Result 142, Processing Time 0.036 seconds

Development of Bio-sensor-Based Feature Extraction and Emotion Recognition Model (바이오센서 기반 특징 추출 기법 및 감정 인식 모델 개발)

  • Cho, Ye Ri;Pae, Dong Sung;Lee, Yun Kyu;Ahn, Woo Jin;Lim, Myo Taeg;Kang, Tae Koo
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.67 no.11
    • /
    • pp.1496-1505
    • /
    • 2018
  • The technology of emotion recognition is necessary for human computer interaction communication. There are many cases where one cannot communicate without considering one's emotion. As such, emotional recognition technology is an essential element in the field of communication. n this regard, it is highly utilized in various fields. Various bio-sensor sensors are used for human emotional recognition and can be used to measure emotions. This paper proposes a system for recognizing human emotions using two physiological sensors. For emotional classification, two-dimensional Russell's emotional model was used, and a method of classification based on personality was proposed by extracting sensor-specific characteristics. In addition, the emotional model was divided into four emotions using the Support Vector Machine classification algorithm. Finally, the proposed emotional recognition system was evaluated through a practical experiment.

Open-source robot platform providing offline personalized advertisements (오프라인 맞춤형 광고 제공을 위한 오픈소스 로봇 플랫폼)

  • Kim, Young-Gi;Ryu, Geon-Hee;Hwang, Eui-Song;Lee, Byeong-Ho;Yoo, Jeong-Ki
    • Journal of Convergence for Information Technology
    • /
    • v.10 no.4
    • /
    • pp.1-10
    • /
    • 2020
  • The performance of the personalized product recommendation system for offline shopping malls is poor compared with the one using online environment information since it is difficult to obtain visitors' characteristic information. In this paper, a mobile robot platform is suggested capable of recommending personalized advertisement using customers' sex and age information provided by Face API of MS Azure Cloud service. The performance of the developed robot is verified through locomotion experiments, and the performance of API used for our robot is tested using sampled images from open Asian FAce Dataset (AFAD). The developed robot could be effective in marketing by providing personalized advertisements at offline shopping malls.

Recognition of Hand gesture to Human-Computer Interaction (손 동작을 통한 인간과 컴퓨터간의 상호 작용)

  • Lee, Lae-Kyoung;Kim, Sung-Shin
    • Proceedings of the KIEE Conference
    • /
    • 2000.07d
    • /
    • pp.2930-2932
    • /
    • 2000
  • In this paper. a robust gesture recognition system is designed and implemented to explore the communication methods between human and computer. Hand gestures in the proposed approach are used to communicate with a computer for actions of a high degree of freedom. The user does not need to wear any cumbersome devices like cyber-gloves. No assumption is made on whether the user is wearing any ornaments and whether the user is using the left or right hand gestures. Image segmentation based upon the skin-color and a shape analysis based upon the invariant moments are combined. The features are extracted and used for input vectors to a radial basis function networks(RBFN). Our "Puppy" robot is employed as a testbed. Preliminary results on a set of gestures show recognition rates of about 87% on the a real-time implementation.

  • PDF

Real-Time Face Tracking Algorithm Robust to illumination Variations (조명 변화에 강인한 실시간 얼굴 추적 알고리즘)

  • Lee, Yong-Beom;You, Bum-Jae;Lee, Seong-Whan;Kim, Kwang-Bae
    • Proceedings of the KIEE Conference
    • /
    • 2000.07d
    • /
    • pp.3037-3040
    • /
    • 2000
  • Real-Time object tracking has emerged as an important component in several application areas including machine vision. surveillance. Human-Computer Interaction. image-based control. and so on. And there has been developed various algorithms for a long time. But in many cases. they have showed limited results under uncontrolled situation such as illumination changes or cluttered background. In this paper. we present a novel. computationally efficient algorithm for tracking human face robustly under illumination changes and cluttered backgrounds. Previous algorithms usually defines color model as a 2D membership function in a color space without consideration for illumination changes. Our new algorithm developed here. however. constructs a 3D color model by analysing plenty of images acquired under various illumination conditions. The algorithm described is applied to a mobile head-eye robot and experimented under various uncontrolled environments. It can track an human face more than 100 frames per second excluding image acquisition time.

  • PDF

Metabolic Rate Estimation for ECG-based Human Adaptive Appliance in Smart Homes (인간 적응형 가전기기를 위한 거주자 심박동 기반 신체활동량 추정)

  • Kim, Hyun-Hee;Lee, Kyoung-Chang;Lee, Suk
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.20 no.5
    • /
    • pp.486-494
    • /
    • 2014
  • Intelligent homes consist of ubiquitous sensors, home networks, and a context-aware computing system. These homes are expected to offer many services such as intelligent air-conditioning, lighting control, health monitoring, and home security. In order to realize these services, many researchers have worked on various research topics including smart sensors with low power consumption, home network protocols, resident and location detection, context-awareness, and scenario and service control. This paper presents the real-time metabolic rate estimation method that is based on measured heart rate for human adaptive appliance (air-conditioner, lighting etc.). This estimation results can provide valuable information to control smart appliances so that they can adjust themselves according to the status of residents. The heart rate based method has been experimentally compared with the location-based method on a test bed.

Emotion Recognition Based on Human Gesture (인간의 제스쳐에 의한 감정 인식)

  • Song, Min-Kook;Park, Jin-Bae;Joo, Young-Hoon
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.17 no.1
    • /
    • pp.46-51
    • /
    • 2007
  • This paper is to present gesture analysis for human-robot interaction. Understanding human emotions through gesture is one of the necessary skills fo the computers to interact intelligently with their human counterparts. Gesture analysis is consisted of several processes such as detecting of hand, extracting feature, and recognizing emotions. For efficient operation we used recognizing a gesture with HMM(Hidden Markov Model). We constructed a large gesture database, with which we verified our method. As a result, our method is successfully included and operated in a mobile system.

Tiny and Blurred Face Alignment for Long Distance Face Recognition

  • Ban, Kyu-Dae;Lee, Jae-Yeon;Kim, Do-Hyung;Kim, Jae-Hong;Chung, Yun-Koo
    • ETRI Journal
    • /
    • v.33 no.2
    • /
    • pp.251-258
    • /
    • 2011
  • Applying face alignment after face detection exerts a heavy influence on face recognition. Many researchers have recently investigated face alignment using databases collected from images taken at close distances and with low magnification. However, in the cases of home-service robots, captured images generally are of low resolution and low quality. Therefore, previous face alignment research, such as eye detection, is not appropriate for robot environments. The main purpose of this paper is to provide a new and effective approach in the alignment of small and blurred faces. We propose a face alignment method using the confidence value of Real-AdaBoost with a modified census transform feature. We also evaluate the face recognition system to compare the proposed face alignment module with those of other systems. Experimental results show that the proposed method has a high recognition rate, higher than face alignment methods using a manually-marked eye position.

Facial expression recognition-based contents preference inference system (얼굴 표정 인식 기반 컨텐츠 선호도 추론 시스템)

  • Lee, Yeon-Gon;Cho, Durkhyun;Jang, Jun Ik;Suh, Il Hong
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2013.01a
    • /
    • pp.201-204
    • /
    • 2013
  • 디지털 컨텐츠의 종류와 양이 폭발적으로 증가하면서 컨텐츠 선호도 투표는 강한 파급력을 지니게 되었다. 하지만 컨텐츠 소비자가 직접 투표를 해야 하는 현재의 방법은 사람들의 투표 참여율이 저조하며, 조작 위험성이 높다는 문제점이 있다. 이에 본 논문에서는 컨텐츠 소비자의 얼굴 표정에 드러나는 감정을 인식함으로써 자동으로 컨텐츠 선호도를 추론하는 시스템을 제안한다. 본 논문에서 제안하는 시스템은 기존의 수동 컨텐츠 선호도 투표 시스템의 문제점인 컨텐츠 소비자의 부담감과 번거로움, 조작 위험성 등을 해소함으로써 보다 편리하고 효율적이며 신뢰도 높은 서비스를 제공하는 것을 목표로 한다. 따라서 본 논문에서는 컨텐츠 선호도 추론 시스템을 구축하기 위한 방법을 구체적으로 제안하고, 실험을 통하여 제안하는 시스템의 실용성과 효율성을 보인다.

  • PDF

Recent Research Trends of Facial Expression Recognition (얼굴표정 인식 기법의 최신 연구 동향)

  • Lee, Min Kyu;Song, Byung Cheol
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2019.11a
    • /
    • pp.128-130
    • /
    • 2019
  • 최근 딥러닝의 급격한 발전과 함께 얼굴표정 인식(facial expression recognition) 기술이 상당한 진보를 이루었다. 얼굴표정 인식은 컴퓨터 비전 분야에서 지속적으로 관심을 받고 있으며, 인포테인먼트 시스템(Infotainment system), 인간-로봇 상호작용(human-robot interaction) 등 다양한 분야에서 활용되고 있다. 그럼에도 불구하고 얼굴표정 인식 분야는 학습 데이터의 부족, 얼굴 각도의 변화 또는 occlusion 등과 같은 많은 문제들이 존재한다. 본 논문은 얼굴표정 인식 분야에서의 위와 같은 고유한 문제들을 다룬 기술들을 포함하여 고전적인 기법부터 최신 기법에 대한 연구 동향을 제시한다.

  • PDF

Hand Shape Classification using Contour Distribution (윤곽 분포를 이용한 이미지 기반의 손모양 인식 기술)

  • Lee, Changmin;Kim, DaeEun
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.20 no.6
    • /
    • pp.593-598
    • /
    • 2014
  • Hand gesture recognition based on vision is a challenging task in human-robot interaction. The sign language of finger spelling alphabets has been tested as a kind of hand gesture. In this paper, we test hand gesture recognition by detecting the contour shape and orientation of hand with visual image. The method has three stages, the first stage of finding hand component separated from the background image, the second stage of extracting the contour feature over the hand component and the last stage of comparing the feature with the reference features in the database. Here, finger spelling alphabets are used to verify the performance of our system and our method shows good performance to discriminate finger alphabets.