• Title/Summary/Keyword: Voice Menu

Search Result 14, Processing Time 0.022 seconds

A Study on the Usability Evaluation of Earcon Applied to Voice Menu (이어콘을 적용한 음성 메뉴의 사용성 평가에 관한 연구)

  • Lim, Chee-Hwan;Lee, Jae-In;Lee, Sung-Soo
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.28 no.4
    • /
    • pp.55-62
    • /
    • 2005
  • This paper describes the experiment that investigated the possibility of design and evaluation of the usability of earcon applied to voice menu. The earcon has functions in providing navigational cue in the hierarchical menu, and also can be applied to voice menu for improving the recall rate and the response time. In this experiment, participants identified their location with the help of earcon applied to the voice menu with the various earcon parameters. In detail some participants listened to the good quality sound of voice menu using the structured earcon, and they recalled the location they heard. Other participants listened to the good quality sound of voice menu without earcon, and they recalled the location they heard in the same manner. And the response times were checked through their answers. On analyzing the results, we found the earcon applied to voice menu showed the increase of recalling rate from what they heard during experiment. That is the performance of task was better with earcon applied to voice menu than with voice menu without earcon. In the earcon applied to voice menu test, it showed the accuracy of 92.50%, but in voice menu without earcon test, people could only recall 66.25% among given questions. The response time was reduced from 4.98 sec to 3.85 sec. In addition, this experiment showed the 87.5% of participants preferred the earcon applied to voice menu.

Mobile Voice Note File Management Service For Improving Accessibility of the Blind (전맹인의 접근성 향상을 위한 모바일 음성 메모 파일 관리 서비스)

  • Lim, Soon-Bum;Lee, Mi Ji;Choi, Yoo Jin;Yook, Juhye;Park, Joo Hyun;Lee, Jongwoo
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.11
    • /
    • pp.1215-1222
    • /
    • 2019
  • Recently, people with disabilities also search for and collect information from the web through smart devices, and save collected information on smart devices or take notes. For non-disabled people, various memo applications are provided on the market, so it is more convenient to choose according to their preference. However, existing memo services are limited for use by blind people due to the importance of visual information. The problem with blind people when using smart devices is that the screen is not recognized, so it is not possible to check in which location the menu of the application exists. In addition, it is difficult to input and manipulate text, and systematic file management and control are not possible. Therefore, in this paper, we propose the development of voice memo service that blind people can use only voice and hearing information and can operate menu with Bluetooth remote controller. We will develop a system that includes a comprehensive voice file management function for storing, searching, playing, and deleting files, rather than simply storing voice files.

An Experimental Study on Hindrance Factors of Usability of Menu Structure in ARS (ARS 메뉴체계 사용성 저해요소에 대한 실험연구)

  • Kim, Ho-Won;Kim, Hee-Cheol
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.15 no.2
    • /
    • pp.462-470
    • /
    • 2011
  • ARS (Automatic Response Systems) based on VUI (Voice User Interface) and TTI (Touch Tone Interface) are one of the most widely used communication systems. Despite common usages, however, inconvenience of ARS is continually pointed out. This may stem from lack of human-centered studies aside from technological development. In this paper, we provide guidelines for designing ARS by analyzing hindrance factors of usability of ARS menu structure. We had selected two call-centers using ARS, and carried out an experimental study where subjects performed the task of "returning books." After that, they completed questionnaires and interviews. We identified four problems: the complex menu structure, lack of representativeness on the menu name, users' awareness of location, and a difficulty to move among menus. And we partially discussed the ways of avoiding the problems.

Development of a Voice User Interface for Web Browser using VoiceXML (VoiceXML을 이용한 VUI 지원 웹브라우저 개발)

  • Yea SangHoo;Jang MinSeok
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.11 no.2
    • /
    • pp.101-111
    • /
    • 2005
  • The present web informations are mainly described in terms of HTML, which users obtain through input devices such as mouse, keyboard, etc. Thus the existing GUI environment have not supported human's most natural information acquisition means, that is, voice. To solve the problem, several vendors are developing voice user interface. However these products are deficient in man -machine interactivity and their accommodation of existing web environment. This paper presents a VUI(Voice User Interface) supporting web browser by utilizing more and more maturing speech recognition technology and VoiceXML, a markup language derived from XML. It provides users with both interfaces, VUI as well as GUI. In addition, XML Island technology is applied to the bowser in a way that VoiceXML fragments are nested in HTML documents to accommodate the existing web environment. Also for better interactivity, dialogue scenarios for menu, bulletin, and search engine are suggested.

Multi-Modal Controller Usability for Smart TV Control

  • Yu, Jeongil;Kim, Seongmin;Choe, Jaeho;Jung, Eui S.
    • Journal of the Ergonomics Society of Korea
    • /
    • v.32 no.6
    • /
    • pp.517-528
    • /
    • 2013
  • Objective: The objective of this study was to suggest a multi-modal controller type for Smart TV Control. Background: Recently, many issues regarding the Smart TV are arising due to the rising complexity of features in a Smart TV. One of the specific issues involves what type of controller must be utilized in order to perform regulated tasks. This study examines the ongoing trend of the controller. Method: The selected participants had experiences with the Smart TV and were 20 to 30 years of age. A pre-survey determined the first independent variable of five tasks(Live TV, Record, Share, Web, App Store). The second independent variable was the type of controllers(Conventional, Mouse, Voice-Based Remote Controllers). The dependent variables were preference, task completion time, and error rate. The experiment consist a series of three experiments. The first experiment utilized a uni-modal Controller for tasks; the second experiment utilized a dual-modal Controller, while the third experiment utilized a triple-modal Controller. Results: The first experiment revealed that the uni-modal Controller (Conventional, Voice Controller) showed the best results for the Live TV task. The second experiment revealed that the dual-modal Controller(Conventional-Voice, Conventional-Mouse combinations) showed the best results for the Share, Web, App Store tasks. The third experiment revealed that the triple-modal Controller among all the level had not effective compared with dual-modal Controller. Conclusion: In order to control simple tasks in a smart TV, our results showed that a uni-modal Controller was more effective than a dual-modal controller. However, the control of complex tasks was better suited to the dual-modal Controller. User preference for a controller differs according the Smart TV functions. For instance, there was a high user preference for the uni-Controller for simple functions while high user preference appeared for Dual-Controllers when the task was complex. Additionally, in accordance with task characteristics, there was a high user preference for the Voice Controller for channel and volume adjustment. Furthermore, there was a high user preference for the Conventional Controller for menu selection. In situations where the user had to input text, the Voice Controller had the highest preference among users while the Mouse Type, Voice Controller had the highest user preference for performing a search or selecting items on the menu. Application: The results of this study may be utilized in the design of a controller which can effectively carry out the various tasks of the Smart TV.

Design and Implementation of Voice-based Interactive Service KIOSK (음성기반 대화형 서비스 키오스크 설계 및 구현)

  • Kim, Sang-woo;Choi, Dae-june;Song, Yun-Mi;Moon, Il-Young
    • Journal of Practical Engineering Education
    • /
    • v.14 no.1
    • /
    • pp.99-108
    • /
    • 2022
  • As the demand for kiosks increases, more users complain of discomfort. Accordingly, a kiosk that enables easy menu selection and order by producing a voice-based interactive service is produced and provided in the form of a web. It implements voice functions based on the Annyang API and SpeechSynthesis API, and understands the user's intention through Dialogflow. And discuss how to implement this process based on Rest API. In addition, the recommendation system is applied based on collaborative filtering to improve the low consumer accessibility of existing kiosks, and to prevent infection caused by droplets during the use of voice recognition services, it provides the ability to check the wearing of masks before using the service.

A Study on the Correlation Between Sasang Constitution and Sound Characteristics Used Harmonics and Formant Bandwidth (Harmonics(배음)와 Formant Bandwidth(포먼트 폭)를 이용한 음성특성(音聲特性)과 사상체질간(四象體質間)의 상관성(相關性) 연구(硏究))

  • Park, Sung-Jin;Kim, Dal-Rae
    • Journal of Sasang Constitutional Medicine
    • /
    • v.16 no.1
    • /
    • pp.61-73
    • /
    • 2004
  • This study was prepared to investigate the correlation between Sasang constitutional groups and voice characteristics using voice analysis system(in this study, CSL). I focused on the voice characteristics in terms of harmonics, Formant frequency and Formant Bandwidth. The subjects were 71 males. I classified them into three groups, that is Soeumin group, Soyangin group and Taeumin group. The classification method of Constitution used two ways, QSCCII(Questionnarie for the Sasang Constitution Classification II) and Interview with a specialist in Sasang Constitution. So 71 people were categorized into 31 Soeumin(people), 18 Soyangin(people) and 22 Taeumin(people). Pitch is approximately similar to the fundamental frequency(F0) in voices. Shimmer in dB gives an evaluation of the period-to-period variability of the peak-to-peak amplitude within the analyzed voice sample. FFT(Fast Fourier Transform) method in CSL can display sampled voices into harmonics. H1 is the first peak and h2 is the second peak in the harmonics. The amplitude difference of h1 and h2(h1-h2) can be explained as the speaker's phonation type, And Formant frequency and bandwidth can be explained as the speaker's vocal tract. So I checked the harmonics and Formant frequency and Bandwidth as the voice parameters. First I have captured /e/ voices from all subjects using microphone. And then I analyzed /e/ voices with CSL. Power Spectrum and Formant History is the menu in the CSL which can display harmonics and Formant frequency and bandwidth. The results about the correlation between Sasang Constitutional Groups and voice parameters are as follows; 1. There is no significant amplitude difference of harmonics(h1-h2) among three groups. 2. There is the significant difference between Soeumin Group and Soyangin Group in Formant Frequency 1 and Formant Bandwidth 1(p<0.05). Any other parameters have no significance. I assume that Soyangin Group has clearer and brighter voice than Soeumin Group according to the Formant Bandwidth difference. And I think its result has coincidence with the context of "Dongyi Suse Bowon" and "Sasangimhejinam".

  • PDF

Improving Eye-gaze Mouse System Using Mouth Open Detection and Pop Up Menu (입 벌림 인식과 팝업 메뉴를 이용한 시선추적 마우스 시스템 성능 개선)

  • Byeon, Ju Yeong;Jung, Keechul
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.12
    • /
    • pp.1454-1463
    • /
    • 2020
  • An important factor in eye-tracking PC interface for general paralyzed patients is the implementation of the mouse interface, for manipulating the GUI. With a successfully implemented mouse interface, users can generate mouse events exactly at the point of their choosing. However, it is difficult to define this interaction in the eye-tracking interface. This problem has been defined as the Midas touch problem and has been a major focus of eye-tracking research. There have been many attempts to solve this problem using blink, voice input, etc. However, it was not suitable for general paralyzed patients because some of them cannot wink or speak. In this paper, we propose a mouth-pop-up, eye-tracking mouse interface that solves the Midas touch problem as well as becoming a suitable interface for general paralyzed patients using a common RGB camera. The interface presented in this paper implements a mouse interface that detects the opening and closing of the mouth to activate a pop-up menu that the user can select the mouse event. After implementation, a performance experiment was conducted. As a result, we found that the number of malfunctions and the time to perform tasks were reduced compared to the existing method.

Personalized Smart Mirror using Voice Recognition (음성인식을 이용한 개인맞춤형 스마트 미러)

  • Dae-Cheol, Kang;Jong-Seok, Lim;Gil-Ho, Lee;Beom-Hee, Lee;Hyoung-Keun, Park
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.17 no.6
    • /
    • pp.1121-1128
    • /
    • 2022
  • Information about the present invention is made available for business use. You are helping to use the LCD, you can't use the LCD screen. During software configuration, Raspbian was used to provide the system environment. We made our way through the menu and made our financial through play. It provides various information such as weather, weather, apps, streamer music, and web browser search function, and it can be charged. Currently, the 'Google Assistant' will be provided through the GUI within a predetermined time.

Development of Speech-Language Therapy Program kMIT for Aphasic Patients Following Brain Injury and Its Clinical Effects (뇌 손상 후 실어증 환자의 언어치료 프로그램 kMIT의 개발 및 임상적 효과)

  • Kim, Hyun-Gi;Kim, Yun-Hee;Ko, Myoung-Hwan;Park, Jong-Ho;Kim, Sun-Sook
    • Speech Sciences
    • /
    • v.9 no.4
    • /
    • pp.237-252
    • /
    • 2002
  • MIT has been applied for nonfluent aphasic patients on the basis of lateralization of brain hemisphere. However, its applications for different languages have some inquiry for aphasic patients because of prosodic and rhythmic differences. The purpose of this study is to develop the Korean Melodic Intonation Therapy program using personal computer and its clinical effects for nonfluent aphasic patients. The algorithm was composed to voice analog signal, PCM, AMDF, Short-time autocorrelation function and center clipping. The main menu contains pitch, waveform, sound intensity and speech files on window. Aphasic patients' intonation patterns overlay on selected kMIT patterns. Three aphasic patients with or without kMIT training participated in this study. Four affirmative sentences and two interrogative sentences were uttered on CSL by stimulus of ST. VOT, VD, Hold and TD were measured on Spectrogram. In addition, articulation disorders and intonation patterns were evaluated objectively on spectrogram. The results indicated that nonfluent aphasic patients with kMIT training group showed some clinical effects of speech intelligibility based on VOT, TD values, articulation evaluation and prosodic pattern changes.

  • PDF