• 제목/요약/키워드: Human Machine Interface

검색결과 336건 처리시간 0.035초

지능형 교통체계 기반 인간-기계 인터페이스 시스템 평가를 위한 정신적부하 측정방법의 비교 연구 (Comparative Study of Subjective Mental Workload Assessment Techniques for the Evaluation of ITS-oriented Human-Machine Interface Systems)

  • Cha, Doo-Won;Park, Peom
    • 대한교통학회지
    • /
    • 제19권3호
    • /
    • pp.45-58
    • /
    • 2001
  • 교통시스템 및 차량의 인간-기계 상호작용 및 인간공학적 평가에 널리 사용되는 운전자의 시각행위분석, 생체신호측정(심박수, 호흡수, 뇌파, 피부전기반사 등), 부가작업의 수행도분석, 반응시간 분석, 차량조작변수 분석(속도, 핸들조작, 가속페달조작) 등의 방법들과 함께 운전자 판단에 의한 주관적평가(Subjective Mental Workload Assessment)는 기존 시스템의 평가뿐만 아니라, ITS(Intelligent Transport Systems) 기반 인간-기계 시스템(항법장치, Head-UP Display, AHS Flat Panel Display, Changeable Message Sign, Cellular Phone, 등)의 인간-기계 상호작용 평가, 안전도 평가, 경쟁시스템 및 디자인 대안의 의사결정, 설계가이드라인 제작 등을 위한 하나의 표준화된 평가도구로 자리 잡고 있다. 이에 본 연구에서는 가장 많은 적용빈도와 효과 및 감도를 제시하는 주관적 평가방법인 NASA-TLX (National Aeronautics and Space Administration-Task Load Index), SWAT(Subjective Workload Assessment Technique), MCH(Modified Cooper-Harper) scale, RNASA-TLX(Revision of NASA-TLX)를 대상으로 항법 장치를 사용한 운전자 부하 비교평가 실험결과를 이용하여, 제시된 방법들의 감도 및 피 실험자의 주관적 판단을 통한 상호 비교분석을 실시하였으며, 결과로 위에 제시된 방법들의 사용가이드라인과 ITS 기반 인간-기계 시스템의 인간공학적 평가를 위한 주관적 평가방법의 비교평가결과를 제시하였다.

  • PDF

EOG(Electro-oculogram) 신호를 이용한 Interface 장치 개발 (Development of Interface device with EOG Signal)

  • 김수종;유호선;김영철
    • 대한전기학회:학술대회논문집
    • /
    • 대한전기학회 2006년도 제37회 하계학술대회 논문집 D
    • /
    • pp.1821-1823
    • /
    • 2006
  • This paper presents a development of interface device for electro-oculogram(EOG) signal and it's application to the wireless mouse of wearable PC. The interface device is composed of five bio-electrodes for detecting oculomotor motion, several band-pass filters, instrumentation amplifier and a microprocessor. we have first analyzed impedance characteristics between skin and a bio-electrode. since the impedance highly depends on human face, it's magnitude differs from person. this interface device was applied to develop a wireless mouse for wearable PC, as a Bio Machine Interface(BMI). Where in the prompt on PC monitor is controlled by only EOG signals. this system was implemented in a Head Mount Display(HMD) unit. experimental results show the accuracy of above 90%.

  • PDF

Multimodal Interface Based on Novel HMI UI/UX for In-Vehicle Infotainment System

  • Kim, Jinwoo;Ryu, Jae Hong;Han, Tae Man
    • ETRI Journal
    • /
    • 제37권4호
    • /
    • pp.793-803
    • /
    • 2015
  • We propose a novel HMI UI/UX for an in-vehicle infotainment system. Our proposed HMI UI comprises multimodal interfaces that allow a driver to safely and intuitively manipulate an infotainment system while driving. Our analysis of a touchscreen interface-based HMI UI/UX reveals that a driver's use of such an interface while driving can cause the driver to be seriously distracted. Our proposed HMI UI/UX is a novel manipulation mechanism for a vehicle infotainment service. It consists of several interfaces that incorporate a variety of modalities, such as speech recognition, a manipulating device, and hand gesture recognition. In addition, we provide an HMI UI framework designed to be manipulated using a simple method based on four directions and one selection motion. Extensive quantitative and qualitative in-vehicle experiments demonstrate that the proposed HMI UI/UX is an efficient mechanism through which to manipulate an infotainment system while driving.

HMD를 이용한 사용자 자세 기반 항공 촬영용 쿼드로터 시스템 제어 인터페이스 개발 (A Posture Based Control Interface for Quadrotor Aerial Video System Using Head-Mounted Display)

  • 김재승;정종민;김한솔;황남웅;최윤호;박진배
    • 전기학회논문지
    • /
    • 제64권7호
    • /
    • pp.1056-1063
    • /
    • 2015
  • In this paper, we develop an interface for aerial photograph platform which consists of a quadrotor and a gimbal using the human body and the head posture. As quadrotors have been widely adopted in many industries such as aerial photography, remote surveillance, and maintenance of infrastructures, the demand of aerial video and photograph has been increasing remarkably. Stick type remote controllers are widely used to control a quadrotor, but this method is not an intuitive way of controlling the aerial vehicle and the camera simultaneously. Therefore, a new interface which controls the serial photograph platform is presented. The presented interface uses the human head movement measured by head-mounted display as a reference for controlling the camera angle, and the human body posture measured from Kinect for controlling the attitude of the quadrotor. As the image captured by the camera is displayed on the head-mounted display simultaneously, the user can feel flying experience and intuitively control the quadrotor and the camera. Finally, the performance of the developed system shown to verify the effectiveness and superiority of the presented interface.

Analysis of Ambiguous Adverbial Expression Used for Instruction of Positioning Control

  • Hiratsuka, Shigeyoshi;Inooka, Hikaru;Kajikawa, Shinya
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2001년도 ICCAS
    • /
    • pp.32.1-32
    • /
    • 2001
  • Up to now, many studies on human-machine interface with voice systems have been reported. However, in these systems, precise instructions are necessary for a robot to execute given tasks successfully. In order to make a robot friendly, more ambiguous instructions with some kinds of degree adverbs (, e.g., "move it a little", "lift it more" and so on, ) will be preferable. Therefore, we analyze the relationship between ambiguous instructions and the characteristics of instructed human motion in positioning task to design human friendly interface systems. Several experimental results show that adverbial expressions are mainly divided into three clusters corresponding to the displacement, and that instructors and perators have several differences in distance sense each other.

  • PDF

모의 지능로봇에서 음성신호에 의한 감정인식 (Speech Emotion Recognition by Speech Signals on a Simulated Intelligent Robot)

  • 장광동;권오욱
    • 대한음성학회:학술대회논문집
    • /
    • 대한음성학회 2005년도 추계 학술대회 발표논문집
    • /
    • pp.163-166
    • /
    • 2005
  • We propose a speech emotion recognition method for natural human-robot interface. In the proposed method, emotion is classified into 6 classes: Angry, bored, happy, neutral, sad and surprised. Features for an input utterance are extracted from statistics of phonetic and prosodic information. Phonetic information includes log energy, shimmer, formant frequencies, and Teager energy; Prosodic information includes pitch, jitter, duration, and rate of speech. Finally a patten classifier based on Gaussian support vector machines decides the emotion class of the utterance. We record speech commands and dialogs uttered at 2m away from microphones in 5different directions. Experimental results show that the proposed method yields 59% classification accuracy while human classifiers give about 50%accuracy, which confirms that the proposed method achieves performance comparable to a human.

  • PDF

EPICS 를 이용한 BPM시스템 개발 (1) (Development of BPM System using EPICS (1))

  • 이은희;윤종철;이진원;최진혁;황정연;남상훈
    • 대한전기학회:학술대회논문집
    • /
    • 대한전기학회 2002년도 하계학술대회 논문집 D
    • /
    • pp.2325-2327
    • /
    • 2002
  • 포항 가속기연구소(PAL)에서는 포항방사광가속기(PLS)가 가동을 시작한 1994년 이후 현재까지 사용되어 온 기존의 제어 시스템을 새로운 환경인 EPICS(Experimental Physics and Industrial Control System) 시스템으로 개발하고 있다. EPICS 시스템의 구성은 IOC(Input/Ouput Controller) 와 OPI(Operator Interface)의 2-Layer로 구성되며 이는 MIU(Machine Interfaces Unit), SCC(Subsystem Computer Control System) 그리고 HMI(Human Machine Interface)로 이어지는 기존의 3-Layer 단계 중 SCC단계를 줄여 2-Layer로 구성된다. 이들 두 계층간의 통신은 Client(OPI)/Server(IOC) 구조의 Channel Access를 통해서 이루어진다. 개발중인 EPICS 시스템은 Open Architecture 구조로 IOC와 OPI 각 부분에서 개발시에 사용된 운영체제나 Hardware 를 사용하지 않고 다른 운영체제나 Hardware를 사용하더라도 하나의 공통부분 즉, Channel Access만 있으면 이를 통해 서로 다른 Subsystem IOC의 데이터를 Access할 수 있다. 전체 EPICS 제어시스템 중 저장링 운전의 핵심이 되는 BPM(Beam Position Monitoring) 및 MPS(Magnet Power Supply) 시스템은 IOC부분에 MVME5100(Target Machine) 보드와 vxWorks(Operating System)를 이용하고 OPI부분에는 SUN Workstation(Host Machine)와 Solaris(Operating System)을 사용하여 개발하고 있다. 본 논문에서는 IOC 및 OPI의 설치 절차와 설치 방법에 대해 기술하였다.

  • PDF

가상현실 대화용 가상걸음 장치의 지능제어 (Intelligent Control of a Virtual Walking Machine for Virtual Reality Interface)

  • 윤정원;박장우;류제하
    • 제어로봇시스템학회논문지
    • /
    • 제12권9호
    • /
    • pp.926-934
    • /
    • 2006
  • This paper proposes intelligent control of a virtual walking machine that can generate infinite floor for various surfaces and can provide proprioceptive feedback of walking to a user. This machine allows users to participate in a life-like walking experience in virtual environments with various terrains. The controller of the machine is implemented hierarchically, at low-level for robust actuator control, at mid-level fur platform control to compensate the external forces by foot contact, and at high-level control for generating walking trajectory. The high level controller is suggested to generate continuous walking on an infinite floor for various terrains. For the high level control, each independent platform follows a man foot during the swing phase, while the other platform moves back during single stance phase. During double limb support, two platforms manipulate neutral positions to compensate the offset errors generated by velocity changes. This control can, therefore, satisfy natural walking conditions in any direction. Transition phase between the swing and the stance phases is detected by using simple switch sensor system, while human foot motions are sensed by careful calibration with a magnetic motion tracker attached to the shoe. Experimental results of walking simulations at level ground, slope, and stairs, show that with the proposed machine, a general person can walk naturally on various terrains with safety and without any considerable disturbances. This interface can be applied to various areas such as VR navigations, rehabilitation, and gait analysis.