• Title/Summary/Keyword: 시각 음성인식

Search Result 130, Processing Time 0.027 seconds

Navigation App for the Blind and Tactile guide stick (시각장애인을 위한 내비게이션 App과 촉각을 이용한 방향 안내 지팡이)

  • Han, Hyo-Byung;Lee, Gi-Hyuk;Park, Keun-Joon;Beom, Hyo-Won;Kim, Ung-Seop;Seong, Ji-Ae
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2018.10a
    • /
    • pp.327-330
    • /
    • 2018
  • 우리는 본 연구를 통해 모바일을 통해 음성으로 목적지를 설정하고 사용자의 위치 정보를 바탕으로 경로 상의 다음 노드 방향을 효과적으로 계산하는 시스템을 설계하였다. 우리가 설계한 시스템은 손잡이에 달린 모터가 예상 경로방향을 가리키고 사용자는 모터 방향을 손가락의 촉각을 통해 인식함으로써 방향을 예측한다.

The Design of User Interface for Location Based Traffic Guide Signal System (위치 기반 교통 신호 안내 시스템의 사용자 인터페이스 설계)

  • Ko, Ginam;Moon, Nammee
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2011.11a
    • /
    • pp.1111-1112
    • /
    • 2011
  • 본 연구는 최근 널리 보급된 스마트 모바일 단말의 제한된 디스플레이에서 사용자에게 짧은 시간에 효율적으로 핵심 정보를 제공함으로써 사용자 편의성을 향상 시킬 수 있는 사용자 인터페이스(User Interface, UI)를 연구하기 위해 사용자 위치 기반 교통 신호 안내 시스템(Location Based Traffic Signal Guide System, LBTGS) 의 UI 설계에 관한 연구이다. 시각 정보와 음성 정보를 활용하여 짧은 시간에 사용자가 전방 교통 신호 정보를 인지하도록 하여 안전한 도로 운행에 반영할 수 있는 UI 를 효과적으로 설계함으로써 신호 위반으로 인한 교통 사고의 발생을 감소시킬 수 있다. 또한 보다 직관적으로 짧은 시간에 인식할 수 있는 UI 를 설계하여 스마트 모바일 단말의 제한된 디스플레이에서 향상된 사용자 경험을 제공할 수 있을 것으로 기대한다.

OnDot: Braille Training System for the Blind (시각장애인을 위한 점자 교육 시스템)

  • Kim, Hak-Jin;Moon, Jun-Hyeok;Song, Min-Uk;Lee, Se-Min;Kong, Ki-sok
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.20 no.6
    • /
    • pp.41-50
    • /
    • 2020
  • This paper deals with the Braille Education System which complements the shortcomings of the existing Braille Learning Products. An application dedicated to the blind is configured to perform full functions through touch gestures and voice guidance for user convenience. Braille kit is produced for educational purposes through Arduino and 3D printing. The system supports the following functions. First, the learning of the most basic braille, such as initial consonants, final consonant, vowels, abbreviations, etc. Second, the ability to check learned braille by solving step quizzes. Third, translation of braille. Through the experiment, the recognition rate of touch gestures and the accuracy of braille expression were confirmed, and in case of translation, the translation was done as intended. The system allows blind people to learn braille efficiently.

Design and Implementation of a Language Identification System for Handwriting Input Data (필기 입력데이터에 대한 언어식별 시스템의 설계 및 구현)

  • Lim, Chae-Gyun;Kim, Kyu-Ho;Lee, Ki-Young
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.10 no.1
    • /
    • pp.63-68
    • /
    • 2010
  • Recently, to accelerate the Ubiquitous generation, the input interface of the mobile machinery and tools are actively being researched. In addition with the existing interfaces such as the keyboard and curser (mouse), other subdivisions including the handwriting, voice, vision, and touch are under research for new interfaces. Especially in the case of small-sized mobile machinery and tools, there is a increasing need for an efficient input interface despite the small screens. This is because, additional installment of other devices are strictly limited due to its size. Previous studies on handwriting recognition have generally been based on either two-dimensional images or algorithms which identify handwritten data inserted through vectors. Futhermore, previous studies have only focused on how to enhance the accuracy of the handwriting recognition algorithms. However, a problem arisen is that when an actual handwriting is inserted, the user must select the classification of their characters (e.g Upper or lower case English, Hangul - Korean alphabet, numbers). To solve the given problem, the current study presents a system which distinguishes different languages by analyzing the form/shape of inserted handwritten characters. The proposed technique has treated the handwritten data as sets of vector units. By analyzing the correlation and directivity of each vector units, a more efficient language distinguishing system has been made possible.

The Character Recognition System of Mobile Camera Based Image (모바일 이미지 기반의 문자인식 시스템)

  • Park, Young-Hyun;Lee, Hyung-Jin;Baek, Joong-Hwan
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.11 no.5
    • /
    • pp.1677-1684
    • /
    • 2010
  • Recently, due to the development of mobile phone and supply of smart phone, many contents have been developed. Especially, since the small-sized cameras are equiped in mobile devices, people are interested in the image based contents development, and it also becomes important part in their practical use. Among them, the character recognition system can be widely used in the applications such as blind people guidance systems, automatic robot navigation systems, automatic video retrieval and indexing systems, automatic text translation systems. Therefore, this paper proposes a system that is able to extract text area from the natural images captured by smart phone camera. The individual characters are recognized and result is output in voice. Text areas are extracted using Adaboost algorithm and individual characters are recognized using error back propagated neural network.

A Study on the Creation of Digital Self-portrait with Intertextuality (상호텍스트성을 활용한 디지털 자화상 창작)

  • Lim, Sooyeon
    • The Journal of the Convergence on Culture Technology
    • /
    • v.8 no.1
    • /
    • pp.427-434
    • /
    • 2022
  • The purpose of this study is to create a self-portrait that provides an immersive experience that immerses the viewer into the problem of self-awareness. We propose a method to implement an interactive self-portrait by using audio and image information obtained from viewers. The viewer's voice information is converted into text and visualized. In this case, the viewer's face image is used as pixel information composing the text. Text is the result of a mixture of one's own emotions, imaginations, and intentions based on personal experiences and memories. People have different interpretations of certain texts in different ways.The proposed digital self-portrait not only reproduces the viewer's self-consciousness in the inner aspect by utilizing the intertextuality of the text, but also expands the meanings inherent in the text. Intertextuality in a broad sense refers to the totality of all knowledge that occurs between text and text, and between subject and subject. Therefore, the self-portrait expressed in text expands and derives various relationships between the viewer and the text, the viewer and the viewer, and the text and the text. In addition, this study shows that the proposed self-portrait can confirm the formativeness of text and re-create spatial and temporality in the external aspect. This dynamic self-portrait reflects the interests of viewers in real time, and has the characteristic of being updated and created.

Phonological awareness skills in terms of visual and auditory stimulus and syllable position in typically developing children (청각적, 시각적 자극제시 방법과 음절위치에 따른 일반아동의 음운인식 능력)

  • Choi, Yu Mi;Ha, Seunghee
    • Phonetics and Speech Sciences
    • /
    • v.9 no.4
    • /
    • pp.123-128
    • /
    • 2017
  • This study aims to compare the performance of syllable identification task according to auditory and visual stimuli presentation methods and syllable position. Twenty-two typically developing children (age 4-6) participated in the study. Three-syllable words were used to identify the first syllable and the final syllable in each word with auditory and visual stimuli. For the auditory stimuli presentation, the researcher presented the test word only with oral speech. For the visual stimuli presentation, the test words were presented as a picture, and asked each child to choose appropriate pictures for the task. The results showed that when tasks were presented visually, the performances of phonological awareness were significantly higher than in presenting with auditory stimuli. Also, the performances of the first syllable identification were significantly higher than those of the last syllable identification. When phonological awareness task are presented by auditory stimuli, it is necessary to go through all the steps of the speech production process. Therefore, the phonological awareness performance by auditory stimuli may be low due to the weakness of the other stages in the speech production process. When phonological awareness tasks are presented using visual picture stimuli, it can be performed directly at the phonological representation stage without going through the peripheral auditory processing, phonological recognition, and motor programming. This study suggests that phonological awareness skills can be different depending on the methods of stimulus presentation and syllable position of the tasks. The comparison of performances between visual and auditory stimulus tasks will help identify where children may show weakness and vulnerability in speech production process.

Effects of auditory and visual presentation on phonemic awareness in 5- to 6- year-old children (청각적 말소리 자극과 시각적 글자 자극 제시방법에 따른 5, 6세 일반아동의 음소인식 수행력 비교)

  • Kim, Myung-Heon;Ha, Ji-Wan
    • Phonetics and Speech Sciences
    • /
    • v.8 no.1
    • /
    • pp.71-80
    • /
    • 2016
  • The phonemic awareness tasks (phonemic synthesis, phonemic elision, phonemic segmentation) by auditory presentation and visual presentation were conducted to 40 children who are 5 and 6 years old. The scores and error types in the sub-tasks by two presentations were compared to each other. Also, the correlation between the performances of phonemic awareness sub-tasks in two presentation conditions were examined. As a result, 6-year-old group showed significantly higher phonemic awareness scores than 5-year-old group. Both group showed significantly higher scores in visual presentation than auditory presentation. While the performance under the visual presentation was significantly lower especially in the segmentation than the other two tasks, there was no significant difference among sub-tasks under the auditory presentation. 5-year-old group showed significantly more 'no response' errors than 6-year-old group and 6-year-old group showed significantly more 'phoneme substitution' and 'phoneme omission' errors than 5-year-old group. Significantly more 'phoneme omission' errors were observed in the segmentation than the elision task, and significantly more 'phoneme addition' errors were observed in elision than the synthesis task. Lastly, there are positive correlations in auditory and visual synthesis tasks, auditory and visual elision tasks, and auditory and visual segmentation tasks. Summarizing the results, children tend to depend on orthographic knowledge when acquiring the initial phonemic awareness. Therefore, the result of this research would support the position that the orthographic knowledge affects the improvement of phonemic awareness.

Analysis of Korea's Artificial Intelligence Competitiveness Based on Patent Data: Focusing on Patent Index and Topic Modeling (특허데이터 기반 한국의 인공지능 경쟁력 분석 : 특허지표 및 토픽모델링을 중심으로)

  • Lee, Hyun-Sang;Qiao, Xin;Shin, Sun-Young;Kim, Gyu-Ri;Oh, Se-Hwan
    • Informatization Policy
    • /
    • v.29 no.4
    • /
    • pp.43-66
    • /
    • 2022
  • With the development of artificial intelligence technology, competition for artificial intelligence technology patents around the world is intensifying. During the period 2000 ~ 2021, artificial intelligence technology patent applications at the US Patent and Trademark Office have been steadily increasing, and the growth rate has been steeper since the 2010s. As a result of analyzing Korea's artificial intelligence technology competitiveness through patent indices, it is evaluated that patent activity, impact, and marketability are superior in areas such as auditory intelligence and visual intelligence. However, compared to other countries, overall Korea's artificial intelligence technology patents are good in terms of activity and marketability, but somewhat inferior in technological impact. While noise canceling and voice recognition have recently decreased as topics for artificial intelligence, growth is expected in areas such as model learning optimization, smart sensors, and autonomous driving. In the case of Korea, efforts are required as there is a slight lack of patent applications in areas such as fraud detection/security and medical vision learning.

A Relevance Feedback Method Using Threshold Value and Pre-Fetching (경계 값과 pre-fetching을 이용한 적합성 피드백 기법)

  • Park Min-Su;Hwang Byung-Yeon
    • Journal of Korea Multimedia Society
    • /
    • v.7 no.9
    • /
    • pp.1312-1320
    • /
    • 2004
  • Recently, even if a lot of visual feature representations have been studied and systems have been built, there is a limit to existing content-based image retrieval mechanism in its availability. One of the limits is the gap between a user's high-level concepts and a system's low-level features. And human beings' subjectivity in perceiving similarity is excluded. Therefore, correct visual information delivery and a method that can retrieve the data efficiently are required. Relevance feedback can increase the efficiency of image retrieval because it responds of a user's information needs in multimedia retrieval. This paper proposes an efficient CBIR introducing positive and negative relevance feedback with threshold value and pre-fetching to improve the performance of conventional relevance feedback mechanisms. With this Proposed feedback strategy, we implement an image retrieval system that improves the conventional retrieval system.

  • PDF