• Title/Summary/Keyword: 시각 음성인식

Search Result 129, Processing Time 0.031 seconds

음성정보기술 국제 표준화 동향

  • 홍기형
    • Korea Information Processing Society Review
    • /
    • v.11 no.2
    • /
    • pp.33-41
    • /
    • 2004
  • 언제, 어디서, 어떤 장치를 사용하더라도 정보의 검색이나 접근이 가능해지는 유비쿼터스 환경이 가시화되고 있어, 음성을 사용한 사용자 인터페이스의 중요성이 증대하고 있다. 음성인식, 합성, 화자인증 등 음성 등의 음성처리엔진의 기술 수준이 상용화 단계에 접어들고, 이동 중과 같이 모니터 등 시각적 인터페이스의 사용이 용이하지 않은 상황에서도 정보 접근기 요구가 증대함에 따라 음성은 정보시스템 의 중요한 인터페이스로 자리매김 하고 있다. 음성정보기술은 인간의 가장 자연스러운 상호작용 수단인 음성을 이용하여. 시스템에 명령을 내리고, 시스템의 명령 수행 결과를 음성으로 전달하는 음성을 이용한 정보 시스템 인터페이스를 구현하기 위 한 기술을 의미한다. 또한, 음성정보시스템은 음성을 이용한 사용자인터페이스가 가능한 정보시스템을 뜻하며, (그림 1)과 같이 음성처리엔진, 사용자 접속망, 사용자 프로파일 및 음성응용시스템으로 구성된다.(중략)

  • PDF

An Optical Character Recognition Method using a Smartphone Gyro Sensor for Visually Impaired Persons (스마트폰 자이로센서를 이용한 시각장애인용 광학문자인식 방법)

  • Kwon, Soon-Kak;Kim, Heung-Jun
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.21 no.4
    • /
    • pp.13-20
    • /
    • 2016
  • It is possible to implement an optical character recognition system using a high-resolution camera mounted on smart phones in the modern society. Further, characters extracted from the implemented application is possible to provide the voice service for the visually impaired person by using TTS. But, it is difficult for the visually impaired person to properly shoot the objects that character information are included, because it is very hard to accurately understand the current state of the object. In this paper, we propose a method of inducing an appropriate shooting for the visually impaired persons by using a smartphone gyro sensor. As a result of simulation using the implemented program, we were able to see that it is possible to recognize the more character from the same object using the proposed method.

New developmental direction of telecommunications for Disabilities Welfare (장애인복지를 위한 정보통신의 발전방향)

  • 박민수
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.4 no.1
    • /
    • pp.35-43
    • /
    • 2000
  • This paper was studied on developmental direction of telecommunications for disabilities welfare. Method of this study is delphi method. Persons with disabilities is classed as motor disability, visual handicap, hearing impairment, and language and speech disorders. Persons with motor disability is needs as follow, speed recognition technology, video recognition technology, breath capacity recognition technology. Persons with visual handicap is needs as follow, display recognition technology, speed recognition technology, text recognition technology, intelligence conversion handling technology, video recognition - speed synthetic technology. Persons with hearing impairment and language - speech disorders is needs as follow, speed signal handling technology, speed recognition technology, intelligence conversion handling technology, video recognition technology, speed synthetic technology the results of this study is as follow: first, disabilities telecommunications organization must be constructed. Second, persons with disabilities in need of universal service. Third, Persons with disabilities in need of information education, Fourth, studying for telecommunications in need of support. Fifth, small telecommunications company in need of support. Sixth, software industry in need of new development. Seventh, Persons with disabilities in need of standard guideline for telecommunications.

  • PDF

The Real-time Shopping System using Multipurpose Visual Language with Voice Recognize (음성인식시스템과 다목적 시각 언어를 연동한 실시간 쇼핑 시스템)

  • Kim, Young-Jong
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.16 no.6
    • /
    • pp.4164-4169
    • /
    • 2015
  • In this paper planed Real-time Shopping System using Multipurpose Visual Language System(MVLS) with voice recognition remote controller. This system has a merit that using existing on-line & off-line shopping system with addition MVLS data. This can realization little modification existing shopping system. Also, customer's a point of view that has a merit to using easy device for shopping. That is no more using difficult device like that keyboard or mouse, and approach to easy device that voice recognition remote controller or smart phone. Especially, aspect of the old and the infirm and disabled persons that information minority group, can easy buy the product using this system. And, the sellers can more easily collection customer's data and using that future sales strategy.

An User-Friendly Kiosk System Based on Deep Learning (딥러닝 기반 사용자 친화형 키오스크 시스템)

  • Su Yeon Kang;Yu Jin Lee;Hyun Ah Jung;Seung A Cho;Hyung Gyu Lee
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.29 no.1
    • /
    • pp.1-13
    • /
    • 2024
  • This study aims to provide a customized dynamic kiosk screen that considers user characteristics to cope with changes caused by increased use of kiosks. In order to optimize the screen composition according to the characteristics of the digital vulnerable group such as the visually impaired, the elderly, children, and wheelchair users, etc., users are classified into nine categories based on real-time analysis of user characteristics (wheelchair use, visual impairment, age, etc.). The kiosk screen is dynamically adjusted according to the characteristics of the user to provide efficient services. This study shows that the system communication and operation were performed in the embedded environment, and the used object detection, gait recognition, and speech recognition technologies showed accuracy of 74%, 98.9%, and 96%, respectively. The proposed technology was verified for its effectiveness by implementing a prototype, and through this, this study showed the possibility of reducing the digital gap and providing user-friendly "barrier-free kiosk" services.

Speech Activity Detection using Lip Movement Image Signals (입술 움직임 영상 선호를 이용한 음성 구간 검출)

  • Kim, Eung-Kyeu
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.11 no.4
    • /
    • pp.289-297
    • /
    • 2010
  • In this paper, A method to prevent the external acoustic noise from being misrecognized as the speech recognition object is presented in the speech activity detection process for the speech recognition. Also this paper confirmed besides the acoustic energy to the lip movement image signals. First of all, the successive images are obtained through the image camera for personal computer and the lip movement whether or not is discriminated. The next, the lip movement image signal data is stored in the shared memory and shares with the speech recognition process. In the mean time, the acoustic energy whether or not by the utterance of a speaker is verified by confirming data stored in the shared memory in the speech activity detection process which is the preprocess phase of the speech recognition. Finally, as a experimental result of linking the speech recognition processor and the image processor, it is confirmed to be normal progression to the output of the speech recognition result if face to the image camera and speak. On the other hand, it is confirmed not to the output the result of the speech recognition if does not face to the image camera and speak. Also, the initial feature values under off-line are replaced by them. Similarly, the initial template image captured while off-line is replaced with a template image captured under on-line, so the discrimination of the lip movement image tracking is raised. An image processing test bed was implemented to confirm the lip movement image tracking process visually and to analyze the related parameters on a real-time basis. As a result of linking the speech and image processing system, the interworking rate shows 99.3% in the various illumination environments.

Home Appliance Control through Speech Recognition User Interface (음성 인식 사용자 인터페이스를 통한 가전기기 제어 기법)

  • Song, Wook;Jang, Hyun-Su;Eom, Young-Ik
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2006.11a
    • /
    • pp.265-268
    • /
    • 2006
  • 유비쿼터스 컴퓨팅 환경이 확대됨에 따라, 기존의 키보드와 마우스만을 사용자 인터페이스로 주로 사용했던 방법에서 벗어나 좀 더 사용자 중심의 멀티모달 유저 인터페이스 적응이 요구되고 있다. 이에 XHTML+Voice는 음성 및 시각을 모두 제공할 수 있는 새로운 서비스 패러다임으로서 기존의 음성정보만을 제공하거나 시각적인 정보만을 제공하는 시스템과는 달리 XHTML내에 VoiceXML을 삽입함으로써 두 언어의 장점을 모두 활용할 수 있다. 본 논문에서는 VoiceXML의 이러한 장점을 살려 스마트 홈을 구성하는 여러 가전기기들의 인터페이스를 미리 템플릿으로 만들어 두어 모바일 디바이스를 통해 이것들을 제어하는 시나리오를 제안하고 구현하는 방법에 대해 실험하였다.

  • PDF

Implementation of Signboard Voice Guidance Service for Visually Impaired Person Using Virtual Beacon (가상비콘을 이용한 시각장애인 대상 간판 음성 안내 구현)

  • Lee, Yunho;Park, Kwangjung;Kwon, Soon-Kak
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.23 no.6
    • /
    • pp.1-8
    • /
    • 2018
  • In this paper, we implement the signboard voice guidance service for visually impaired person using virtual beacon. The new location to provide a location-based service can be added easily by using virtual beacons, which are locating by Wi-fi, GPS, and so on, instead of physical beacons, which are locating by physical devices. We provide the voice service for guiding information of the captured signboard for the visually impaired when he arrived at the location registered through the virtual beacon.

Design and Implementation of ICT Convergence Autonomous Driving Service Robot for the Visually Impaired (시각장애인을 위한 ICT융합 자율주행 서비스 로봇 설계 및 구현)

  • Gu, Ye-Chan;Kwon, Se-Jin;Nam, Ga-Bin;Lee, Woong-Ki
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2022.11a
    • /
    • pp.1003-1005
    • /
    • 2022
  • 코로나-19의 여파와 기술의 발전으로 비대면 거래가 증가하고 있는 추세다. 비대면 거래가 증가하면서 무인점포도 늘어나고 있고, 늘어남에 따라서 시각장애인들은 무인점포를 이용하는데 어려움을 겪고 있다. 본 논문에서는 자율주행 로봇에 음성인식 기술과 딥러닝 기술을 적용하여 시각장애인에게 도움을 줄 수 있도록 ICT융합 로봇을 구현하였다. 무인점포 뿐만이 아니라 안내가 필요한 다양한 장소에서도 적용될 수 있을 것으로 기대한다.

Situation-awareness White Cane Using a Mobile Device (모바일기기를 이용한 상황인식-흰지팡이)

  • Jeon, Dong-Hee;Jeon, Jun-Uk;Beak, Hwa-Hyeon;Moon, Mi-Kyeong
    • Journal of the Korea Society of Computer and Information
    • /
    • v.19 no.11
    • /
    • pp.167-173
    • /
    • 2014
  • A white cane is used by many people who are blind or visually impaired. The studies and developments in the field of smart electronic white cane have been progressed to improve walking safety of the visually handicapped person. In this paper, we describe a situation-awareness white cane which support and guide walking of the visually handicapped person by giving information of surrounding situation through connection of a white cane and a mobile device. There are easily accessible buttons in the situation-awareness white cane. These buttons is connected with the mobile device by blue-tooth, so by pushing these buttons, the visually handicapped person can receive sound-guidance. As a result, the visually handicapped person can walk more safely by getting more information of situation.