• Title/Summary/Keyword: Speech recognition platform

Search Result 32, Processing Time 0.03 seconds

Development of robotic hands of signbot, advanced Malaysian sign-language performing robot

  • Al-Khulaidi, Rami Ali;Akmeliawati, Rini;Azlan, Norsinnira Zainul;Bakr, Nuril Hana Abu;Fauzi, Norfatehah M.
    • Advances in robotics research
    • /
    • v.2 no.3
    • /
    • pp.183-199
    • /
    • 2018
  • This paper presents the development of a 3D printed humanoid robotic hands of SignBot, which can perform Malaysian Sign Language (MSL). The study is considered as the first attempt to ease the means of communication between the general community and the hearing-impaired individuals in Malaysia. The signed motions performed by the developed robot in this work can be done by two hands. The designed system, unlike previously conducted work, includes a speech recognition system that can feasibly integrate with the controlling platform of the robot. Furthermore, the design of the system takes into account the grammar of the MSL which differs from that of Malay spoken language. This reduces the redundancy and makes the design more efficient and effective. The robot hands are built with detailed finger joints. Micro servo motors, controlled by Arduino Mega, are also loaded to actuate the relevant joints of selected alphabetical and numerical signs as well as phrases for emergency contexts from MSL. A database for the selected signs is developed wherein the sequential movements of the servo motor arrays are stored. The results showed that the system performed well as the selected signs can be understood by hearing-impaired individuals.

AI Advisor for Response of Disaster Safety in Risk Society (위험사회 재난 안전 분야 대응을 위한 AI 조력자)

  • Lee, Yong-Hak;Kang, Yunhee;Lee, Min-Ho;Park, Seong-Ho;Kang, Myung-Ju
    • Journal of Platform Technology
    • /
    • v.8 no.3
    • /
    • pp.22-29
    • /
    • 2020
  • The 4th industrial revolution is progressing by country as a mega trend that leads various technological convergence directions in the social and economic fields from the initial simple manufacturing innovation. The epidemic of infectious diseases such as COVID-19 is shifting digital-centered non-face-to-face business from economic operation, and the use of AI and big data technology for personalized services is essential to spread online. In this paper, we analyze cases focusing on the application of artificial intelligence technology, which is a key technology for the effective implementation of the digital new deal promoted by the government, as well as the major technological characteristics of the 4th industrial revolution and describe the use cases in the field of disaster response. As a disaster response use case, AI assistants suggest appropriate countermeasures according to the status of the reporter in an emergency call. To this end, AI assistants provide speech recognition data-based analysis and disaster classification of converted text for adaptive response.

  • PDF

A Study on the Educational Uses of Smart Speaker (스마트 스피커의 교육적 활용에 관한 연구)

  • Chang, Jiyeun
    • Journal of the Korea Convergence Society
    • /
    • v.10 no.11
    • /
    • pp.33-39
    • /
    • 2019
  • Edutech, which combines education and information technology, is in the spotlight. Core technologies of the 4th Industrial Revolution have been actively used in education. Students use an AI-based learning platform to self-diagnose their needs. And get personalized training online with a cloud learning platform. Recently, a new educational medium called smart speaker that combines artificial intelligence technology and voice recognition technology has emerged and provides various educational services. The purpose of this study is to suggest a way to use smart speaker educationally to overcome the limitation of existing education. To this end, the concept and characteristics of smart speakers were analyzed, and the implications were derived by analyzing the contents provided by smart speakers. Also, the problem of using smart speaker was considered.

Trends in Neuromorphic Software Platform for Deep Neural Network (딥 뉴럴 네트워크 지원을 위한 뉴로모픽 소프트웨어 플랫폼 기술 동향)

  • Yu, Misun;Ha, Youngmok;Kim, Taeho
    • Electronics and Telecommunications Trends
    • /
    • v.33 no.4
    • /
    • pp.14-22
    • /
    • 2018
  • Deep neural networks (DNNs) are widely used in various domains such as speech and image recognition. DNN software frameworks such as Tensorflow and Caffe contributed to the popularity of DNN because of their easy programming environment. In addition, many companies are developing neuromorphic processing units (NPU) such as Tensor Processing Units (TPUs) and Graphical Processing Units (GPUs) to improve the performance of DNN processing. However, there is a large gap between NPUs and DNN software frameworks due to the lack of framework support for various NPUs. A bridge for the gap is a DNN software platform including DNN optimized compilers and DNN libraries. In this paper, we review the technical trends of DNN software platforms.

A Design and Implementation of Speech Recognition and Synthetic Application for Hearing-Impairment

  • Kim, Woo-Lin;Ham, Hye-Won;Yun, Sang-Un;Lee, Won Joo
    • Journal of the Korea Society of Computer and Information
    • /
    • v.26 no.12
    • /
    • pp.105-110
    • /
    • 2021
  • In this paper, we design and implement an Android mobile application that helps hearing impaired people communicate based on STT(Speech-to-Text) and TTS(Text-to-Speech) APIs and accelerometer sensor of a smartphone. This application provides the ability to record what the hearing-Impairment person's interlocutor is saying with a microphone, convert it to text using the STT API, and display it to the hearing-Impairment person. In addition. In addition, when a hearing-impaired person inputs a text using the TTS API, it is converted into voice and told to the interlocutor. When a hearing-impaired person shakes their smartphone, an accelerometer based background service function is provided to run the application. The application implemented in this paper provides a function that allows hearing impaired people to communicate easily with other people when communicating with others without using sign language as a video call.

Dobot-based Omok platform using Voice recognition (음성인식을 활용한 Dobot 기반 오목 플랫폼)

  • Park, Sang-Yong;Lee, Kang Hee
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2020.07a
    • /
    • pp.19-21
    • /
    • 2020
  • 해가 갈수록 여러 요인들로 인하여 장애인 인구는 증가하고 있다. 하지만 이러한 장애인들을 위한 주변기기의 발전은 미흡한 상황이며 더욱이 장애인들의 여가를 위한 놀이문화, 기술들은 더욱 발전이 더딘 상태이다. 여가 활동은 장애인의 행복에도 직관적인 영향을 끼치는 중요한 부분이다. 따라서 본 논문에서는 장애인들 중 손과 같은 신체를 움직이기 어려운 중증 지체장애인을 위한 오목 플랫폼을 제안한다. 본 논문에서 구현한 오목 플랫폼은 음성인식을 기반으로 사용자가 오목알을 착수하고 싶은 좌표를 음성으로 입력시키면 Dobot 즉 로봇암을 통하여 착수점에 오목알을 착수한다. 실험에선 Google Vocie To Text API를 Python 환경에서 사용하여 사용자의 음성입력을 받았으며 Dobot Studio의 Script에서 입력된 음성 값을 오목판에 1대1로 맵핑시켜 정확한 위치에 착수할 수 있도록 구현하였다. 본 논문의 연구 결과를 응용하면 오목에만 국한되는게 아닌 다양한 보드게임을 구현할 수 있다.

  • PDF

Design and implement of the Educational Humanoid Robot D2 for Emotional Interaction System (감성 상호작용을 갖는 교육용 휴머노이드 로봇 D2 개발)

  • Kim, Do-Woo;Chung, Ki-Chull;Park, Won-Sung
    • Proceedings of the KIEE Conference
    • /
    • 2007.07a
    • /
    • pp.1777-1778
    • /
    • 2007
  • In this paper, We design and implement a humanoid robot, With Educational purpose, which can collaborate and communicate with human. We present an affective human-robot communication system for a humanoid robot, D2, which we designed to communicate with a human through dialogue. D2 communicates with humans by understanding and expressing emotion using facial expressions, voice, gestures and posture. Interaction between a human and a robot is made possible through our affective communication framework. The framework enables a robot to catch the emotional status of the user and to respond appropriately. As a result, the robot can engage in a natural dialogue with a human. According to the aim to be interacted with a human for voice, gestures and posture, the developed Educational humanoid robot consists of upper body, two arms, wheeled mobile platform and control hardware including vision and speech capability and various control boards such as motion control boards, signal processing board proceeding several types of sensors. Using the Educational humanoid robot D2, we have presented the successful demonstrations which consist of manipulation task with two arms, tracking objects using the vision system, and communication with human by the emotional interface, the synthesized speeches, and the recognition of speech commands.

  • PDF

The Study on Automatic Speech Recognizer Utilizing Mobile Platform on Korean EFL Learners' Pronunciation Development (자동음성인식 기술을 이용한 모바일 기반 발음 교수법과 영어 학습자의 발음 향상에 관한 연구)

  • Park, A Young
    • Journal of Digital Contents Society
    • /
    • v.18 no.6
    • /
    • pp.1101-1107
    • /
    • 2017
  • This study explored the effect of ASR-based pronunciation instruction, using a mobile platform, on EFL learners' pronunciation development. Particularly, this quasi-experimental study focused on whether using mobile ASR, which provides voice-to-text feedback, can enhance the perception and production of target English consonants minimal pairs (V-B, R-L, and G-Z) of Korean EFL learners. Three intact classes of 117 Korean university students were assigned to three groups: a) ASR Group: ASR-based pronunciation instruction providing textual feedback by the mobile ASR; b) Conventional Group: conventional face-to-face pronunciation instruction providing individual oral feedback by the instructor; and the c) Hybrid Group: ASR-based pronunciation instruction plus conventional pronunciation instruction. The ANCOVA results showed that the adjusted mean score for pronunciation production post-test on the Hybrid instruction group (M=82.71, SD =3.3) was significantly higher than the Conventional group (M=62.6, SD =4.05) (p<.05).

A Study On The ASP Module Using VoiceMXL in Automatic Speech Recognition System (VoiceXML을 이용한 음성 인식시스템에서의 ASP 모듈 연구)

  • 장준식;김민석;윤재석
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2001.10a
    • /
    • pp.609-612
    • /
    • 2001
  • In this research, it has been shown that how the computer can recognize and understand spoken natural language and its symbolization using VoiceXML and Grammar Specific Language. In order for user to hear correct information, ASP Module has been revised and its effectivities has been experimented on the Voice portal airplane information system platform.

  • PDF

A Review on Deep Learning Platform for Artificial Intelligence (인공지능 딥러링 학습 플랫폼에 관한 선행연구 고찰)

  • Jin, Chan-Yong;Shin, Seong-Yoon;Nam, Soo-Tai
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2019.05a
    • /
    • pp.169-170
    • /
    • 2019
  • Lately, as artificial intelligence becomes a source of global competitiveness, the government is strategically fostering artificial intelligence that is the base technology of future new industries such as autonomous vehicles, drones, and robots. Domestic artificial intelligence research and services have been launched mainly in Naver and Kakao, but their size and level are weak compared to overseas. Recently, deep learning has been conducted in recent years while recording innovative performance in various pattern recognition fields including speech recognition and image recognition. In addition, deep running has attracted great interest from industry since its inception, and global information technology companies such as Google, Microsoft, and Samsung have successfully applied deep learning technology to commercial products and are continuing research and development. Therefore, we will look at artificial intelligence which is attracting attention based on previous research.

  • PDF