• Title/Summary/Keyword: Voice command

Search Result 92, Processing Time 0.026 seconds

The Development of Heuristics for Voice Shopping Service through Voice Interface with Display (디스플레이 탑재형 음성 인터페이스를 통한 음성쇼핑 서비스 휴리스틱 개발)

  • Gwon, Hyeon Jeong;Lee, Jee Yeon
    • Journal of the Korean Society for information Management
    • /
    • v.39 no.2
    • /
    • pp.1-33
    • /
    • 2022
  • Voice shopping is gaining attention following the trend of non-contact E-commerce by enabling people to shop via voice command. Therefore, in this study, voice shopping service heuristics using a display-mounted voice interface were developed in preparation for the future where voice shopping becomes a part of daily life in the world. First, as a theoretical approach, a literature survey of 50 papers on the design principles of 'visual interface,' 'voice interface,' and 'shopping service' was conducted to produce a total of 29 draft design principles. Second, as an empirical approach, a focus group interview was conducted on consumer decision-making processes in shopping experiences and information-seeking behavior within the context of shopping to draft the heuristics. This was to supplement the user experience, a weak part of the literature research. Finally, a Delphi survey asked 20 experts in UX, service planning, artificial intelligence development, and shopping to evaluate the heuristics draft developed through the above two stages. After three rounds of Delphi surveys, the final heuristics were proposed.

The University Gusdance System using the Alexa (Alexa를 이용한 대학안내 시스템)

  • Kim, Tae Jin;Kim, Dong Hyun
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.21 no.11
    • /
    • pp.2061-2066
    • /
    • 2017
  • The voice recognition technology is to recognize the voice of an user and execute the command. Recently, the voice recognition is evolving to the artificial intelligence voice recognition by adding the scheme of the natural language processing. The AI voice recognition is exploited to control the IoT devices or provide the information, such as the news or the wether. The University Information which is one of fields serviced by the information provider is mainly presented on the web. However, since too much information are presented on the web, it is difficult for an user to find efficiently the specific information which the user want to know. In this paper, we design and implement the university guidance system to recognize the user voice searching the information and provide the result using the voice. To do this, we classify the university data and design the lambda function to provide the data.

Smart Mirror of Personal Environment using Voice Recognition (음성인식을 이용한 개인환경의 스마트 미러)

  • Yeo, Un-Chan;Park, Sin-Hoo;Moon, Jin-Wan;An, Seong-Won;Han, Yeong-Oh
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.14 no.1
    • /
    • pp.199-204
    • /
    • 2019
  • This paper introduces smart mirror that provides the contents needed for an individual's daily life. When a command that is designated as voice recognition is entered, Smart Mirror is produced that outputs desired contents from a display. The contents of the current smart mirror include time, weather, subway information, schedule and photography. Smart mirror sold for commercial private households is difficult to distribute due to high prices, but the smart mirror production presented in this paper can lower the manufacturing cost and can be more easily used by voice recognition.

A Proposal of Eye-Voice Method based on the Comparative Analysis of Malfunctions on Pointer Click in Gaze Interface for the Upper Limb Disabled (상지장애인을 위한 시선 인터페이스에서 포인터 실행 방법의 오작동 비교 분석을 통한 Eye-Voice 방식의 제안)

  • Park, Joo Hyun;Park, Mi Hyun;Lim, Soon-Bum
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.4
    • /
    • pp.566-573
    • /
    • 2020
  • Computers are the most common tool when using the Internet and utilizing a mouse to select and execute objects. Eye tracking technology is welcomed as an alternative technology to help control computers for users who cannot use their hands due to their disabilities. However, the pointer execution method of the existing eye tracking technique causes many malfunctions. Therefore, in this paper, we developed a gaze tracking interface that combines voice commands to solve the malfunction problem when the upper limb disabled uses the existing gaze tracking technology to execute computer menus and objects. Usability verification was conducted through comparative experiments regarding the improvements of the malfunction. The upper limb disabled who are hand-impaired use eye tracking technology to move the pointer and utilize the voice commands, such as, "okay" while browsing the computer screen for instant clicks. As a result of the comparative experiments on the reduction of the malfunction of pointer execution with the existing gaze interfaces, we verified that our system, Eye-Voice, reduced the malfunction rate of pointer execution and is effective for the upper limb disabled to use.

Development of a User-Customizable Voice Command Application for Android Phones (안드로이드폰용 사용자 정의 음성명령 앱 개발)

  • Gwak, Juri;Lee, Joohyeon;Lim, Hyojin;Lee, Ki Yong
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2014.11a
    • /
    • pp.1171-1174
    • /
    • 2014
  • 최근 Siri, S Voice 등과 같이 스마트폰에 음성으로 명령을 내리는 기능이 보편화되고 있다. 하지만 기존의 음성명령 시스템은 사전에 정의된 명령어가 아니면 사용이 불가능하며 어떤 명령어들이 사용가능한지 확인하기도 어렵다. 본 논문에서는 사용자에게 좀 더 편리한 인터페이스를 제공하기 위한 목적으로, 사용자가 직접 음성명령을 등록할 수 있는 안드로이드폰용 음성명령 앱을 개발한다. 본 논문에서 개발한 음성명령 앱은 사용자가 등록한 임의의 음성을 인식하여 안드로이드폰에 설치된 애플리케이션을 실행시키거나, 특정 번호로 전화를 걸거나, 특정 번호로 문자메시지를 보낼 수 있다. 본 논문에서 개발한 앱을 통해 사용자는 보다 편리하게 스마트폰에 음성명령을 내릴 수 있다.

Voice Command through Facial Recognition Smart Mirror System (얼굴인식을 통한 음성 명령 스마트 거울 시스템)

  • Lee, Se-Hoon;Kim, Su-Min;Park, Hyun-Gyu
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2019.01a
    • /
    • pp.253-254
    • /
    • 2019
  • 본 논문에서는 가정 등에서 사용자의 행동 반경에 가장 많이 있는 거울에 홈 제어 및 근처 전열 기구들을 보다 쉽게 제어 할 수 있도록 Google Speech API와 Open CV 라이브러리를 사용해 음성인식을 통한 홈 제어 방안을 제시하였다. 이를 통해서 바쁜 아침에 화장 등을 하는 경우 두 손을 자유롭게 사용하면서 디바이스를 음성으로 제어 할 수 있는 편리성을 제공하였다.

  • PDF

Cursor Moving by Voice Command using Fuzzy Inference (퍼지 추론을 이용한 음성 명령에 의한 커서 조작)

  • 추명경;손영선
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2000.11a
    • /
    • pp.209-212
    • /
    • 2000
  • 본 논문에서 마우스 대신에 음성으로 명령을 입력하여 퍼지 추론을 통해 윈도우 화면상의 커서를 이동시키는 인터페이스를 구현하였다. 입력된 음성이 대체로 짧은 언어이기에 이를 인식하기 위하여 고립단어 인식에 강한 DTW방식을 사용하였다. DTW방식의 단점중인 하나가 음성길이가 비슷한 명령을 입력하였을 때 표준패턴 중 오차 값이 가장 작은 패턴으로 인식하는 것이다. 예를들면 아주 많이 이동해 라는 음성이 입력되었을 때 동일한 음성길이를 가진 아주 많이 오른쪽으로 인식하는 경우가 있다. 이런 오류를 해결하고자 각 패턴의 DTW 오차 값 범위와 표준 패턴의 음성길이를 기준으로 임계값을 퍼지 추론하여 명령으로서 수락 여부를 결정하였다. 판단이 애매한 부분은 사용자에게 질의를 하여 응답에 따라 수락 여부를 결정하였다.

  • PDF

Speaker Tracking System for Autonomous Mobile Robot (자율형 이동로봇을 위한 전방위 화자 추종 시스템)

  • Lee, Chang-Hoon;Kim, Yong-Hoh
    • Proceedings of the KIEE Conference
    • /
    • 2002.11c
    • /
    • pp.142-145
    • /
    • 2002
  • This paper describes a omni-directionally speaker tracking system for mobile robot interface in real environment. Its purpose is to detect a robust 360-degree sound source and to recognize voice command at a long distance(60-300cm). We consider spatial features, the relation of position and interaural time differences, and realize speaker tracking system using fuzzy inference process based on inference rules generated by its spatial features.

  • PDF

Automatic Vowel Sequence Reproduction for a Talking Robot Based on PARCOR Coefficient Template Matching

  • Vo, Nhu Thanh;Sawada, Hideyuki
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.5 no.3
    • /
    • pp.215-221
    • /
    • 2016
  • This paper describes an automatic vowel sequence reproduction system for a talking robot built to reproduce the human voice based on the working behavior of the human articulatory system. A sound analysis system is developed to record a sentence spoken by a human (mainly vowel sequences in the Japanese language) and to then analyze that sentence to give the correct command packet so the talking robot can repeat it. An algorithm based on a short-time energy method is developed to separate and count sound phonemes. A matching template using partial correlation coefficients (PARCOR) is applied to detect a voice in the talking robot's database similar to the spoken voice. Combining the sound separation and counting the result with the detection of vowels in human speech, the talking robot can reproduce a vowel sequence similar to the one spoken by the human. Two tests to verify the working behavior of the robot are performed. The results of the tests indicate that the robot can repeat a sequence of vowels spoken by a human with an average success rate of more than 60%.

Interface Modeling for Digital Device Control According to Disability Type in Web

  • Park, Joo Hyun;Lee, Jongwoo;Lim, Soon-Bum
    • Journal of Multimedia Information System
    • /
    • v.7 no.4
    • /
    • pp.249-256
    • /
    • 2020
  • Learning methods using various assistive and smart devices have been developed to enable independent learning of the disabled. Pointer control is the most important consideration for the disabled when controlling a device and the contents of an existing graphical user interface (GUI) environment; however, difficulties can be encountered when using a pointer, depending on the disability type; Although there are individual differences depending on the blind, low vision, and upper limb disability, problems arise in the accuracy of object selection and execution in common. A multimodal interface pilot solution is presented that enables people with various disability types to control web interactions more easily. First, we classify web interaction types using digital devices and derive essential web interactions among them. Second, to solve problems that occur when performing web interactions considering the disability type, the necessary technology according to the characteristics of each disability type is presented. Finally, a pilot solution for the multimodal interface for each disability type is proposed. We identified three disability types and developed solutions for each type. We developed a remote-control operation voice interface for blind people and a voice output interface applying the selective focusing technique for low-vision people. Finally, we developed a gaze-tracking and voice-command interface for GUI operations for people with upper-limb disability.