• Title/Summary/Keyword: Interaural Time Difference

Search Result 18, Processing Time 0.039 seconds

Implementation of Sound Source Location Detector (음원 위치 검출기의 구현)

  • 이종혁;김진천
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.4 no.5
    • /
    • pp.1017-1025
    • /
    • 2000
  • The human auditory system has been shown to posses remarkable abilities in the localization and tracking of sound sources. The localization is the result of processing two primary acoustics cues. These are the interaural time difference(ITD) cues and interaural intensity difference(IID) cues at the two ears. In this paper, we propose TEPILD(Time Energy Previous Integration Location Detector) model. TEPILD model is constructed with time function generator, energy function generator, previous location generator and azimuth detector. Time function generator is to process ITD and energy function generator is to process IID. Total average accuracy rate is 99.2%. These result are encouraging and show that proposed model can be applied to the sound source location detector.

  • PDF

A Study on Sweet Spot of Crosstalk Cancellation Schemes for Sound Rendering Systems (입체음향시스템을 위한 상호간접제거 기법의 유효청취범위 분석)

  • Lee, Jung-Hyuck;Jeong, Sang-Hyo;Yoo, Seung-Soo;Song, Iick-Ho;Kim, Sun-Yong
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.36 no.5C
    • /
    • pp.309-316
    • /
    • 2011
  • In this paper, equalization zone of two crosstalk cancellation (CC) schemes, which are the one based on only head related transfer function (HRTF) and the other one based on interaural intensity/time difference (ITD/IID) as well as HRTF is studied. To do this, the condition numbers and ITD/IID levels of two schemes are shown.

Salience of Envelope Interaural Time Difference of High Frequency as Spatial Feature (공간감 인자로서의 고주파 대역 포락선 양이 시간차의 유효성)

  • Seo, Jeong-Hun;Chon, Sang-Bae;Sung, Koeng-Mo
    • The Journal of the Acoustical Society of Korea
    • /
    • v.29 no.6
    • /
    • pp.381-387
    • /
    • 2010
  • Both timbral features and spatial features are important in the assessment of multichannel audio coding systems. The prediction model, extending the ITU-R Rec. BS. 1387-1 to multichannel audio coding systems, with the use of spatial features such as ITDDist (Interaural Time Difference Distortion), ILDDist (Interaural Level Difference Distortion), and IACCDist (InterAural Cross-correlation Coefficient Distortion) was proposed by Choi et al. In that model, ITDDistswere only computed for low frequency bands (below 1500Hz), and ILDDists were computed only for high frequency bands (over 2500Hz) according to classical duplex theory. However, in the high frequency range, information in temporal envelope is also important in spatial perception, especially in sound localization. A new model to compute the ITD distortions of temporal envelopes in high frequency components is introduced in this paper to investigate the role of such ITD on spatial perception quantitatively. The computed ITD distortions of temporal envelopes in high frequency components were highly correlated with perceived sound quality of multichannel audio sounds.

Sound Source Localization Method Based on Deep Neural Network (깊은 신경망 기반 음원 추적 기법)

  • Park, Hee-Mun;Jung, Jong-Dae
    • Journal of IKEEE
    • /
    • v.23 no.4
    • /
    • pp.1360-1365
    • /
    • 2019
  • In this paper, we describe a sound source localization(SSL) system which can be applied to mobile robot and automatic control systems. Usually the SSL method finds the Interaural Time Difference, the Interaural Level Difference, and uses the geometrical principle of microphone array. But here we proposed another approach based on the deep neural network to obtain the horizontal directional angle(azimuth) of the sound source. We pick up the sound source signals from the two microphones attached symmetrically on both sides of the robot to imitate the human ears. Here, we use difference of spectral distributions of sounds obtained from two microphones to train the network. We train the network with the data obtained at the multiples of 10 degrees and test with several data obtained at the random degrees. The result shows quite promising validity of our approach.

3-D Sound Image Control for two Channel Headphone (헤드폰을 이용한 3차원 음장 제어시스템)

  • 이동형;김성진;정의필;김규년;이수동
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 1999.10b
    • /
    • pp.307-309
    • /
    • 1999
  • 입체음향이란 음원의 위치에 따라 두 귀에 입력되는 신호를 제어함으로써 시각정보 없이 음상의 위치를 파악할 수 있는 음이다. 헤드폰을 이용하면 음장이 머리 내에 위치하게 됨으로써 거리를 파악하는 것이 매우 힘들다. 본 논문에서는 모노음을 이용하여, 2채널 헤드폰에서 재생할 수 있는 3차원 음을 만들기 위하여 Interaural Time Difference(ITD)와 Interaural Intensity Difference(IID)를 이용한 머리 전달함수(Head Related Transfer Function:HRTF)를 만든 결과와 측정 HRTF 자료인 KEMAR Data를 이용한 결과를 비교하였으며, 거리 효과를 효과적으로 구현하기 위하여 잔향효과를 추가하여 음장을 머리밖으로 꺼냄으로써, 보다 향상된 3차원 음상 제어 시스템을 제안하고 실험하였다.

  • PDF

Implementation of Transaural filter method for the location of listeners and loudspeakers (청취자 및 스피커의 위치에 따른 Transaural 필터 구현)

  • shin Hwang;Bhang Seungbeum;Kim Soonhyob;Cheung Wan-Sup
    • Proceedings of the Acoustical Society of Korea Conference
    • /
    • spring
    • /
    • pp.197-200
    • /
    • 2000
  • 본 논문에서는 공간에 위치한 음원으로부터 양 귀에 들리는 음향을 스피커를 이용하여 재현 하는 기술에 대한 문제점과 이를 효과적으로 구현할 수 있는 방법에 대해 소개한다. 좌우 대칭 위치의 스피커를 이용하여, 음상 정위를 실현할 때 직면한 문제점인 Cross-talk 제거와 스피커와 청취자 양 귀사이의 음향전달 모델 선정에 대한 문제점들을 본 논문에서 고찰한다. 이러한 고찰을 통하여 Transaural 필터의 모델 제시와 본 모델의 음향학적 특성을 소개한다. 인간 청각기관의 공간 감응 인자인 lID(Interaural Intensity Difference)와 ITD(Interaural Time Difference)의 개념을 이용한 Transaural 필터의 구현 방법을 제시한다. 제시한 구현방법과 영국 ISVR의 "Stereo Dipole"과의 비교와, 끝으로, 주파수에 따른 적합한 스피커간의 배치 문제에 관한 것도 언급한다.

  • PDF

CASA-based Front-end Using Two-channel Speech for the Performance Improvement of Speech Recognition in Noisy Environments (잡음환경에서의 음성인식 성능 향상을 위한 이중채널 음성의 CASA 기반 전처리 방법)

  • Park, Ji-Hun;Yoon, Jae-Sam;Kim, Hong-Kook
    • Proceedings of the IEEK Conference
    • /
    • 2007.07a
    • /
    • pp.289-290
    • /
    • 2007
  • In order to improve the performance of a speech recognition system in the presence of noise, we propose a noise robust front-end using two-channel speech signals by separating speech from noise based on the computational auditory scene analysis (CASA). The main cues for the separation are interaural time difference (ITD) and interaural level difference (ILD) between two-channel signal. As a result, we can extract 39 cepstral coefficients are extracted from separated speech components. It is shown from speech recognition experiments that proposed front-end has outperforms the ETSI front-end with single-channel speech.

  • PDF

A study on the effect of leading sound and following sound on sound localization (선행음 및 후속음이 음원의 방향지각에 미치는 영향에 관한 연구)

  • Lee, Chai-Bong
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.16 no.2
    • /
    • pp.40-43
    • /
    • 2015
  • In this paper, the effects of the leading and the following sounds with single frequency on sound localization are investigated. The sounds with different levels and ISIs(Inter Stimuli Intervals) were used. The width of test sound is 2ms, and those of the leading and the following sounds are 10ms. 1 kHz of the test sound is utilized. The arrival time difference in the subject's ears is set to be 0.5ms. The four kinds of level differences used for one ISI are 0, -10, -15, and -20dB interval. The leading sound is found to have more effect on sound localization than the following sound is. The effect of the leading sound is also found to be dependent on the value of ISI. When the value of the ISI is small, different effects affecting the sound localization are observed.

The Implementation of Real-Time Speaker Localization Using Multi-Modality (멀티모달러티를 이용한 실시간 음원추적 시스템 구현)

  • Park, Jeong-Ok;Na, Seung-You;Kim, Jin-Young
    • Proceedings of the KIEE Conference
    • /
    • 2004.11c
    • /
    • pp.459-461
    • /
    • 2004
  • This paper presents an implementation of real-time speaker localization using audio-visual information. Four channels of microphone signals are processed to detect vertical as well as horizontal speaker positions. At first short-time average magnitude difference function(AMDF) signals are used to determine whether the microphone signals are human voices or not. And then the orientation and distance information of the sound sources can be obtained through interaural time difference and interaual level differences. Finally visual information by a camera helps get finer tuning of the speaker orientation. Experimental results of the real-time localization system show that the performance improves to 99.6% compared to the rate of 88.8% when only the audio information is used.

  • PDF

Implementation of Sound Source Localization Based on Audio-visual Information for Humanoid Robots (휴모노이드 로봇을 위한 시청각 정보 기반 음원 정위 시스템 구현)

  • Park, Jeong-Ok;Na, Seung-You;Kim, Jin-Young
    • Speech Sciences
    • /
    • v.11 no.4
    • /
    • pp.29-42
    • /
    • 2004
  • This paper presents an implementation of real-time speaker localization using audio-visual information. Four channels of microphone signals are processed to detect vertical as well as horizontal speaker positions. At first short-time average magnitude difference function(AMDF) signals are used to determine whether the microphone signals are human voices or not. And then the orientation and distance information of the sound sources can be obtained through interaural time difference. Finally visual information by a camera helps get finer tuning of the angles to speaker. Experimental results of the real-time localization system show that the performance improves to 99.6% compared to the rate of 88.8% when only the audio information is used.

  • PDF