• 제목/요약/키워드: audio and visual cues

검색결과 12건 처리시간 0.037초

The Effect of Audio and Visual Cues on Korean and Japanese EFL Learners' Perception of English Liquids

  • Chung, Hyun-Song
    • 영어어문교육
    • /
    • 제11권2호
    • /
    • pp.135-148
    • /
    • 2005
  • This paper investigated the effect of audio and visual cues on Korean and Japanese EFL learners' perception of the lateral/retroflex contrast in English. In a perception experiment, the two English consonants /l/ and /r/ were embedded in initial and medial position in nonsense words in the context of the vowels /i, a, u/. Singletons and clusters were included in the speech material. Audio and video recordings were made using a total of 108 items. The items were presented to Korean and Japanese learners of English in three conditions: audio-alone (A), visual-alone (V) and audio-visual presentation (AV). The results showed that there was no evidence of AV benefit for the perception of the /l/-/r/ contrast for either Korean or Japanese learners of English. Korean listeners showed much better identification rates of the /l/-/r/ contrast than Japanese listeners when presented in audio or audio-visual conditions.

  • PDF

한국어 화자의 영어 양순음 /b/와 순치음 /v/ 식별에서 시각 단서의 효과 (The Effect of Visual Cues in the Identification of the English Consonants /b/ and /v/ by Native Korean Speakers)

  • 김윤현;고성룡
    • 말소리와 음성과학
    • /
    • 제4권3호
    • /
    • pp.25-30
    • /
    • 2012
  • This study investigated whether native Korean listeners could use visual cues for the identification of the English consonants /b/ and /v/. Both auditory and audiovisual tokens of word minimal pairs in which the target phonemes were located in word-initial or word-medial position were used. Participants were instructed to decide which consonant they heard in $2{\times}2$ conditions: cue (audio-only, audiovisual) and location (word-initial, word-medial). Mean identification scores were significantly higher for audiovisual than audio-only condition and for word-initial than word-medial condition. Also, according to signal detection theory, sensitivity, d', and response bias, c were calculated based on both hit rates and false alarm rates. The measures showed that the higher identification rate in the audiovisual condition was related with an increase in sensitivity. There were no significant differences in response bias measures across conditions. This result suggests that native Korean speakers can use visual cues while identifying confusing non-native phonemic contrasts. Visual cues can enhance non-native speech perception.

시청각 자극의 시간적 인지 판단 (Temporal-perceptual Judgement of Visuo-Auditory Stimulation)

  • 유미;이상민;박용군;권대규;김남균
    • 한국정밀공학회지
    • /
    • 제24권1호
    • /
    • pp.101-109
    • /
    • 2007
  • In situations of spatio-temporal perception about visuo-auditory stimulus, researches propose optimal integration hypothesis that perceptual process is optimized to the interaction of the senses for the precision of perception. So, when the visual information considered generally dominant over any other sense is ambiguous, the information of the other sense like auditory stimulus influences the perceptual process in interaction with visual information. Thus, we performed two different experiments to certain the conditions of the interacting senses and influence of the condition. We consider the interaction of the visuo-auditory stimulation in the free space, the color of visual stimulus and sex difference of testee with normal people. In first experiment, 12 participants were asked to judge the change in the frequency of audio-visual stimulation using a visual flicker and auditory flutter stimulation in the free space. When auditory temporal cues were presented, the change in the frequency of the visual stimulation was associated with a perceived change in the frequency of the auditory stimulation as the results of the previous studies using headphone. In second experiment, 30 male and 30 female were asked to judge the change in the frequency of audio-visual stimulation using a color of visual flicker and auditory flutter stimulation. In the color condition using red and green. Both male and female testees showed same perceptual tendency. male and female testees showed same perceptual tendency however, in case of female, the standard deviation is larger than that of male. This results implies that audio-visual asymmetry effects are influenced by the cues of visual and auditory information, such as the orientation between auditory and visual stimulus, the color of visual stimulus.

한국인 및 일본인 영어학습자의 유음 차이 지각에 미치는 시각/청각 효과 (Korean and Japanese EFL Learners' AV Benefit for the Perception of the Liquid Contrast in English)

  • 정현성
    • 대한음성학회지:말소리
    • /
    • 제60호
    • /
    • pp.1-11
    • /
    • 2006
  • This paper investigated AV benefit of Korean and Japanese EFL learners' perception of the liquid contrast in English. In a perception experiment, the two English consonants /l/ and /r/ were embedded in initial and medial position in nonsense words in the context of the vowels /i, a, u/. Singletons and clusters were included in the speech material. Audio and video recordings were made using a total of 108 items. The items were presented to Korean and Japanese learners of English in three conditions: audio-alone (A), visual-alone (V) and audio-visual presentation (AV). The results showed that there was no evidence of AV benefit for the perception of the /l/-/r/ contrast for either Korean or Japanese learners of English. The results suggest that increasing auditory proficiency in identifying a non-native contrast is linked with an increasing proficiency in using visual cues to the contrast.

  • PDF

차량 시스템 개발 및 운전자 인자 연구를 위한 실시간 차량 시뮬레이터의 개발 (Development of a Real-Time Driving Simulator for Vehicle System Development and Human Factor Study)

  • 이승준
    • 한국자동차공학회논문집
    • /
    • 제7권7호
    • /
    • pp.250-257
    • /
    • 1999
  • Driving simulators are used effectively for human factor study, vehicle system development and other purposes by enabling to reproduce actural driving conditions in a safe and tightly controlled enviornment. Interactive simulation requries appropriate sensory and stimulus cuing to the driver . Sensory and stimulus feedback can include visual , auditory, motion, and proprioceptive cues. A fixed-base driving simulator has been developed in this study for vehicle system developmnet and human factor study . The simulator consists of improved and synergistic subsystems (a real-time vehicle simulation system, a visual/audio system and a control force loading system) based on the motion -base simulator, KMU DS-Ⅰ developed for design and evaluation of a full-scale driving simulator and for driver-vehicle interaction.

  • PDF

Real-time 3D Audio Downmixing System based on Sound Rendering for the Immersive Sound of Mobile Virtual Reality Applications

  • Hong, Dukki;Kwon, Hyuck-Joo;Kim, Cheong Ghil;Park, Woo-Chan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제12권12호
    • /
    • pp.5936-5954
    • /
    • 2018
  • Eight out of the top ten the largest technology companies in the world are involved in some way with the coming mobile VR revolution since Facebook acquired Oculus. This trend has allowed the technology related with mobile VR to achieve remarkable growth in both academic and industry. Therefore, the importance of reproducing the acoustic expression for users to experience more realistic is increasing because auditory cues can enhance the perception of the complicated surrounding environment without the visual system in VR. This paper presents a audio downmixing system for auralization based on hardware, a stage of sound rendering pipelines that can reproduce realiy-like sound but requires high computation costs. The proposed system is verified through an FPGA platform with the special focus on hardware architectural designs for low power and real-time. The results show that the proposed system on an FPGA can downmix maximum 5 sources in real-time rate (52 FPS), with 382 mW low power consumptions. Furthermore, the generated 3D sound with the proposed system was verified with satisfactory results of sound quality via the user evaluation.

스펙트럼 성형기법을 이용한 멀티미디어 콘텐츠의 명료도 향상 (Intelligibility Enhancement of Multimedia Contents Using Spectral Shaping)

  • 지유나;박영철;황영수
    • 전자공학회논문지
    • /
    • 제53권11호
    • /
    • pp.82-88
    • /
    • 2016
  • 본 논문에서는 스펙트럼 성형기법을 이용한 멀티미디어 콘텐츠 명료도 향상 알고리즘을 제안한다. 영화, 동영상과 같은 오디오-비주얼 미디어 콘텐츠에서 다이얼로그는 영상의 내용을 이해하기 위한 중요한 요소이다. 하지만 종종 영상내의 효과음, 배경음악 등과 같이 함께 믹싱 된 오디오 성분에 의해 중요한 정보를 지닌 다이얼로그의 명료도가 떨어지는 문제점이 제기되어왔다. 뿐만 아니라 멀티미디어 콘텐츠의 이용 환경이 다양해지면서 청자의 주변 환경 또한 오디오 볼륨에 영향을 미치는 요소가 된다. 본 논문에서는 이러한 문제점을 해결하기 위해 영상의 중요 단서를 담고 있는 사운드트랙의 음성 성분 명료도를 높이고자 한다. 제안된 알고리즘은 먼저 영상의 스테레오 오디오 신호에서 음성 존재 확률(Speech Presence Probability)을 이용한 소프트 마스커를 통해 다이얼로그 성분을 검출한다. 추출된 다이얼로그 성분은 스펙트럼 성형 기법을 적용하여 명료도에 중요한 영향을 미치는 고주파대역의 성분을 증폭시키는 등 음성 신호 스펙트럼의 에너지를 재분배하여 신호의 명료도를 향상 시켰다. 마지막으로 크기 정규화 과정을 통해 프로세스 전과 후의 전체 오디오의 파워를 동일하게 유지함으로써 증폭으로 인한 스피커의 오디오 포화(saturation)를 방지하였다. 실험을 통해 본 알고리즘이 동일한 오디오 볼륨에서 영상의 명료도를 향상시킴을 확인 할 수 있었다.

L2 Proficiency Effect on the Acoustic Cue-Weighting Pattern by Korean L2 Learners of English: Production and Perception of English Stops

  • Kong, Eun Jong;Yoon, In Hee
    • 말소리와 음성과학
    • /
    • 제5권4호
    • /
    • pp.81-90
    • /
    • 2013
  • This study explored how Korean L2 learners of English utilize multiple acoustic cues (VOT and F0) in perceiving and producing the English alveolar stop with a voicing contrast. Thirty-four 18-year-old high-school students participated in the study. Their English proficiency level was classified as either 'high' (HEP) or 'low' (LEP) according to high-school English level standardization. Thirty different synthesized syllables were presented in audio stimuli by combining a 6-step VOTs and a 5-step F0s. The listeners judged how close the audio stimulus was to /t/ or /d/ in L2 using a visual analogue scale. The L2 /d/ and /t/ productions collected from the 22 learners (12 HEP, 10 LEP) were acoustically analyzed by measuring VOT and F0 at the vowel onset. Results showed that LEP listeners attended to the F0 in the stimuli more sensitively than HEP listeners, suggesting that HEP listeners could inhibit less important acoustic dimensions better than LEP listeners in their L2 perception. The L2 production patterns also exhibited a group-difference between HEP and LEP in that HEP speakers utilized their VOT dimension (primary cue in L2) more effectively than LEP speakers. Taken together, the study showed that the relative cue-weighting strategies in L2 perception and production are closely related to the learner's L2 proficiency level in that more proficient learners had a better control of inhibiting and enhancing the relevant acoustic parameters.

Haptics for Human-Machine Interaction at The Johns Hopkins University

  • Okamura, Allison M.;Chang, Sung-Ouk
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2003년도 ICCAS
    • /
    • pp.2676-2681
    • /
    • 2003
  • The Haptic Exploration Laboratory at The Johns Hopkins University is currently exploring many problems related to haptics (force and tactile information) in human-machine systems. We divide our work into two main areas: virtual environments and robot-assisted manipulation systems. Our interest in virtual environments focuses on reality-based modeling, in which measurements of the static and dynamic properties of actual objects are taken in order to produce realistic virtual environments. Thus, we must develop methods for acquiring data from real objects and populating pre-defined models. We also seek to create systems that can provide active manipulation assistance to the operator through haptic, visual, and audio cues. These systems may be teleoperated systems, which allow human users to operate in environments that would normally be inaccessible due to hazards, distance, or scale. Alternatively, cooperative manipulation systems allow a user and a robot to share a tool, allowing the user to guide or override the robot directly if necessary. Haptics in human-machine systems can have many applications, such as undersea and space operations, training for pilots and surgeons, and manufacturing. We focus much of our work on medical applications.

  • PDF

신생아 중환자실 환아 달래기시 나타나는 간호사 행위 분석: 기관지폐이형성증 환아 중심으로 (Analysis of Nurses' Soothing Behaviors in Neonatal Intensive Care Unit: Focused on Babies with Bronchopulmonary Dysplasia)

  • 이유나;신현숙
    • Child Health Nursing Research
    • /
    • 제23권4호
    • /
    • pp.494-504
    • /
    • 2017
  • Purpose: The aim of this study was to analyze Neonatal Intensive Care Unit nurses' behaviors while soothing newborns with bronchopulmonary dysplasia. Methods: An observational study was used to assess nurses' soothing behaviors. Data were collected from September, 2012 to March, 2013 using an audio-video recording system. Participants were eight babies and 12 nurses caring for those babies. After obtaining parental permission, the overall process of each episode from nurses' engagement in soothing to the end of soothing was recorded. Then a researcher interviewed each participating nurse. Data from 18 episodes were transcribed as verbal and nonverbal nursing behaviors and then categorized by two researchers. Results: There were 177 observed soothing behaviors which were classified with the five sensory-based categories (tactile, oral, visual, auditory, vestibular). Most frequently observed soothing behavior was 'Gently talking' followed by 'Removing irritant', and 'Providing non-nutritive sucking'. Nurses' perceived soothing behaviors were similar to the observed soothing behaviors except for 'Gently talking'. Conclusion: Nurses used diverse and mixed soothing behaviors as well as recognizing those behaviors as essential nursing skills. Nurses' soothing behaviors identified in this study can be used to comfort babies and to enhance their developmental potential in accordance with individual characterstics or cues.