• Title/Summary/Keyword: multimodal interface

Search Result 54, Processing Time 0.062 seconds

Multimodal Context-aware Service (멀티모달 상황인지 서비스)

  • Jeong, Yeong-Joon;Park, Seong-Soo;Ahn, Se-Yeol
    • 한국IT서비스학회:학술대회논문집
    • /
    • 2006.11a
    • /
    • pp.400-406
    • /
    • 2006
  • 다양한 서비스와의 서비스간 융합이 활발히 이루어지는 유비쿼터스 서비스 환경에서 사용자에게 보다 편리하게 맞춤 서비스를 제공하기 위해서는 다양한 입출력 수단을 통해 상황에 따라 적절한 서비스를 제공할 수 있는 기술이 필요하다. 멀티모달 상황인지 기술은 언제, 어디서나 사용자에게 최적의 서비스를 제공할 수 있도록 다양한 상황정보를 인지하여 적절한 입출력 수단을 사용할 수 있는 UI를 제공함으로써 사용자별 맞춤형 서비스가 가능하게 하는 기술이다. 본 고에서는 멀티모달 및 상황인지 기술에 대한 개요와 KT에서 개발하고 있는 멀티모달 상황인지 플랫폼 및 이를 기반으로 한 홈네트워크 서비스를 소개하고자 한다.

  • PDF

Developing the Design Guideline of Auditory User Interface for Digital Appliances (가전제품의 청각 사용자 인터페이스(AUI) 디자인을 위한 가이드라인 개발 사례)

  • Lee, Ju-Hwan;Jeon, Myoung-Hoon;Han, Kwang-Hee
    • Science of Emotion and Sensibility
    • /
    • v.10 no.3
    • /
    • pp.307-320
    • /
    • 2007
  • In this study, we attempted to provide a distinctive cognitive, emotional 'Auditory User Interface (AUI) Design Guideline' according to home appliance groups and their functions. It is an effort to apply a new design method to practical affairs to overcome the limit of GUI centered appliance design and reflect user multimodal properties by presenting a guideline possible to generate auditory signals intuitively associable with the operational functions. The reason why this study is required is because of frequent instances given rise to annoyance as not systematic application of AUI, but arbitrary mapping. This study tried to provide a useful guideline of AUI in home appliances by extracting the relations with cognitive, emotional properties of a certain device or function induced by several properties of auditory signal and showing the empirical data on the basic mechanism of such relations.

  • PDF

Nano Bio Imaging for NT and BT

  • Moon, DaeWon
    • Proceedings of the Korean Vacuum Society Conference
    • /
    • 2015.08a
    • /
    • pp.51.2-51.2
    • /
    • 2015
  • Understanding interfacial phenomena has been one of the main research issues not only in semiconductors but only in life sciences. I have been trying to meet the atomic scale surface and interface analysis challenges from semiconductor industries and furthermore to extend the application scope to biomedical areas. Optical imaing has been most widely and successfully used for biomedical imaging but complementary ion beam imaging techniques based on mass spectrometry and ion scattering can provide more detailed molecular specific and nanoscale information In this presentation, I will review the 27 years history of medium energy ion scattering (MEIS) development at KRISS and DGIST for nanoanalysis. A electrostatic MEIS system constructed at KRISS after the FOM, Netherland design had been successfully applied for the gate oxide analysis and quantitative surface analysis. Recenlty, we developed time-of-flight (TOF) MEIS system, for the first time in the world. With TOF-MEIS, we reported quantitative compositional profiling with single atomic layer resolution for 0.5~3 nm CdSe/ZnS conjugated QDs and ultra shallow junctions and FINFET's of As implanted Si. With this new TOF-MEIS nano analysis technique, details of nano-structured materials could be measured quantitatively. Progresses in TOF-MEIS analysis in various nano & bio technology will be discussed. For last 10 years, I have been trying to develop multimodal nanobio imaging techniques for cardiovascular and brain tissues. Firstly, in atherosclerotic plaque imaging, using, coherent anti-stokes raman scattering (CARS) and time-of-flight secondary ion mass spectrometry (TOF-SIMS) multimodal analysis showed that increased cholesterol palmitate may contribute to the formation of a necrotic core by increasing cell death. Secondly, surface plasmon resonance imaging ellipsometry (SPRIE) was developed for cell biointerface imaging of cell adhesion, migration, and infiltration dynamics for HUVEC, CASMC, and T cells. Thirdly, we developed an ambient mass spectrometric imaging system for live cells and tissues. Preliminary results on mouse brain hippocampus and hypotahlamus will be presented. In conclusions, multimodal optical and mass spectrometric imaging privides overall structural and morphological information with complementary molecular specific information, which can be a useful methodology for biomedical studies. Future challenges in optical and mass spectrometric imaging for new biomedical applications will be discussed.

  • PDF

Layout Based Multimodal Contents Aughoring Tool for Digilog Book (디지로그 북을 위한 레이아웃 기반 다감각 콘텐츠 저작 도구)

  • Park, Jong-Hee;Woo, Woon-Tack
    • 한국HCI학회:학술대회논문집
    • /
    • 2009.02a
    • /
    • pp.512-515
    • /
    • 2009
  • In this paper, we propose layout based multimodal contents authoring tool for Digilog Book. In authoring step, users create a virtual area using mouse or pen-type device and select property of the area repetitively. After finishing authoring step, system recognizes printed page number and generate page layout including areas and property information. Page layout is represented as a scene graph and stored as XML format. Digilog Book viewer loads stored page layout and analyze properties then augment virtual contents or execute functions based on area. Users can author visual and auditory contents easily by using hybrid interface. In AR environment, system provides area templates in order to help creating area. In addition, proposed authoring tool separates page recognition module from page tracking module. So, it is possible to author many pages using only single marker. As a result of experiment, we showed proposed authoring tool has reasonable performance time in AR environment. We expect that proposed authoring tool would be applicable to many fields such as education and publication.

  • PDF

Emotion Generation Model for Tutoring Agents (교육용 에이전트를 위한 감성 생성 모델)

  • Choo, Moon Won;Choi, Young Mie
    • Proceedings of the Korea Multimedia Society Conference
    • /
    • 2002.05d
    • /
    • pp.812-822
    • /
    • 2002
  • The interface metaphor has been evolved gradually from desktop to agent-oriented paradigm. Multimedia contents could be simply recognized as the multimodal communicational interface. In this respect, the emotional agents are actively focused as the research topics to test the possibility for realizing anthropomorphized and sympathetic interfaces. In this paper, the emotion generation model for tutoring agents is suggested.

  • PDF

Implementation of Pen-Gesture Recognition System for Multimodal User Interface (멀티모달 사용자 인터페이스를 위한 펜 제스처인식기의 구현)

  • 오준택;이우범;김욱현
    • Proceedings of the IEEK Conference
    • /
    • 2000.11c
    • /
    • pp.121-124
    • /
    • 2000
  • In this paper, we propose a pen gesture recognition system for user interface in multimedia terminal which requires fast processing time and high recognition rate. It is realtime and interaction system between graphic and text module. Text editing in recognition system is performed by pen gesture in graphic module or direct editing in text module, and has all 14 editing functions. The pen gesture recognition is performed by searching classification features that extracted from input strokes at pen gesture model. The pen gesture model has been constructed by classification features, ie, cross number, direction change, direction code number, position relation, distance ratio information about defined 15 types. The proposed recognition system has obtained 98% correct recognition rate and 30msec average processing time in a recognition experiment.

  • PDF

Multimodal audiovisual speech recognition architecture using a three-feature multi-fusion method for noise-robust systems

  • Sanghun Jeon;Jieun Lee;Dohyeon Yeo;Yong-Ju Lee;SeungJun Kim
    • ETRI Journal
    • /
    • v.46 no.1
    • /
    • pp.22-34
    • /
    • 2024
  • Exposure to varied noisy environments impairs the recognition performance of artificial intelligence-based speech recognition technologies. Degraded-performance services can be utilized as limited systems that assure good performance in certain environments, but impair the general quality of speech recognition services. This study introduces an audiovisual speech recognition (AVSR) model robust to various noise settings, mimicking human dialogue recognition elements. The model converts word embeddings and log-Mel spectrograms into feature vectors for audio recognition. A dense spatial-temporal convolutional neural network model extracts features from log-Mel spectrograms, transformed for visual-based recognition. This approach exhibits improved aural and visual recognition capabilities. We assess the signal-to-noise ratio in nine synthesized noise environments, with the proposed model exhibiting lower average error rates. The error rate for the AVSR model using a three-feature multi-fusion method is 1.711%, compared to the general 3.939% rate. This model is applicable in noise-affected environments owing to its enhanced stability and recognition rate.

Developing the Design Guideline of Auditory User Interface for Domestic Appliances (가전제품의 청각 사용자 인터페이스(AUI) 설계를 위한 가이드라인 개발 연구)

  • Lee, Ju-Hwan;Jeon, Myoung-Hoon;Ahn, Jeong-Hee;Han, Kwang-Hee
    • 한국HCI학회:학술대회논문집
    • /
    • 2006.02b
    • /
    • pp.1-8
    • /
    • 2006
  • 본 연구는 가전제품의 제품군과 그 기능들에 따라 차별화 가능한 인지적, 감성적 '청각 사용자 인터페이스 디자인 가이드라인(Auditory User Interface Design Guideline)'을 마련하고, 가전제품의 작동기능 정보와 직관적으로 연합 가능한 청각신호(auditory signal)를 제작할 수 있는 지침을 제시하여 GUI 중심의 제품 설계에서 한 차원 확장되고 사용자의 다중감각적 특성이 적용된 디자인 방법을 실무에 적용하고자 하였다. 특히 AUI 에 대한 체계를 확립함으로써 브랜드 정체성(Brand Identity) 및 기업 이미지를 제고할 수 있다는 목적을 함께 고려하였다. 이러한 연구가 필요했던 이유는 가전제품에 대한 소비자의 심적 모형(mental model)과 감성 측면에서의 접근에 대한 요구 때문인데, 이는 AUI 의 체계적 적용이 아닌 임의적 연결(mapping)으로 인한 버저(buzzer) 청각신호의 짜증(annoying) 발생이 빈번한 사례들에서 출발한다. 또한 GUI 의 변화와 수준에 미치지 못하는 AUI 의 업그레이드 필요성과 가전제품에서의 감성 마케팅 경향을 반영하는 의미를 지니고 있다. 이와 함께 멀티미디어 환경의 급속한 확산으로 다중감각적 정보제시(multimodal display)가 요구되는 상황에 걸맞은 시도이다. 본 연구는 특정 가전제품이나 특정 기능이 지니고 있는 인지적, 감성적 차원의 속성을 청각신호(auditory signal)의 다양한 속성들로 유발하는 관계를 추출하고, 이를 형성하는 기본 메커니즘에 대한 경험적 자료를 제시하여, 가전제품의 AUI 디자인에 유용한 가이드라인을 제공하고자 하였다. 그러나 본 논문에서는 연구의 구체적이고 세부적인 결과보다는 전체적인 계획과 진행과정의 절차를 소개하여 관련분야 연구 진행의 참조적 틀을 마련하고자 한다.

  • PDF

A Way to Smart Interface based on the IEEE 1451 Standards for Five-senses Information Device in Ubiquitous Environments (유비쿼터스 환경에서 오감 정보 장치를 위한 IEEE 1451 표준 기반의 스마트 인터페이스 방안)

  • Kim, Dong-Jin;Kim, Jeong-Do;Ham, Yu-Kyung;Lee, Jung-Hwan
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.9 no.2
    • /
    • pp.339-346
    • /
    • 2008
  • Ubiquitous computer can be interfaced with many several peripheral devices for information acquisition. Users should be able to easily use these devices without considering when these devices were interfaced, how to use these devices, and interoperability issues (such as plug and play, the installation of device drivers, and so on). Further, computers and their users need an interface technology that provides five-senses information (the recognition and expressions of the user) such that multimodal interaction can be enabled. In this paper, we proposed an IEEE 1451 standard that uses a smart interface standard for interfacing devices with ubiquitous computer. IEEE 1451 describes the property information of a transducer in the transducer electronic data sheet (TEDS). Further, by using the TEDS format, the interoperability between devices can be enabled by means of the plug-and-play function; it also makes the inconvenient installations of device drivers unnecessary.

A Multimodal Emotion Recognition Using the Facial Image and Speech Signal

  • Go, Hyoun-Joo;Kim, Yong-Tae;Chun, Myung-Geun
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.5 no.1
    • /
    • pp.1-6
    • /
    • 2005
  • In this paper, we propose an emotion recognition method using the facial images and speech signals. Six basic emotions including happiness, sadness, anger, surprise, fear and dislike are investigated. Facia] expression recognition is performed by using the multi-resolution analysis based on the discrete wavelet. Here, we obtain the feature vectors through the ICA(Independent Component Analysis). On the other hand, the emotion recognition from the speech signal method has a structure of performing the recognition algorithm independently for each wavelet subband and the final recognition is obtained from the multi-decision making scheme. After merging the facial and speech emotion recognition results, we obtained better performance than previous ones.