• Title/Summary/Keyword: Front-Back Confusion

Search Result 17, Processing Time 0.023 seconds

Improvement of front-back sound localization characteristics in headphone-based 3D sound generation (헤드폰 기반의 입체음향 생성에서 앞/뒤 음상정위 특성 개선)

  • 김경훈;김시호;배건성;최송인;박만호
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.8C
    • /
    • pp.1142-1148
    • /
    • 2004
  • A binaural filtering method using HRTF DB is generally used to make the headphone-based 3D sound. But it can make some confusion between front and back directions or between up and down directions due to the non-individual HRTF depending on each listener. To reduce the confusion of sound image localization, we propose a new method to boost the spectral cue by modifying HRTF spectra with spectrum difference between front and back directions. Informal listening tests show that the proposed method improves the front-back sound localization characteristics much better than the conventional methods

Improvement of front/back Sound Localization Characteristics using Psychoacoustics of Head Related Transfer Function (머리전달함수의 심리음향적 특성을 이용한 전/후 음상정위 특성 개선)

  • Koo, Kyo-Sik;Cha, Hyung-Tai
    • Journal of Broadcast Engineering
    • /
    • v.11 no.4 s.33
    • /
    • pp.448-457
    • /
    • 2006
  • HRTF DB, including the information of the sounds which is arrived to our ears, is generally used to make a 3D sound. But it can decline some three-dimensional effects by the confusion between front and back directions due to the non-individual HRTF depending on each listener. In this paper, we propose a new method to use psychoacoustic theory that reduces the confusion of sound image localization. And we make use of an excitation energy by the sense of hearing. This method is brought HRTF spectrum characteristics into relief to draw out the energy ratio about the bark band. Informal listening tests show that the proposed method improves the front-back sound localization characteristics much better than the conventional methods.

Improvement of virtual speaker localization characteristics using grouped HRTF (머리전달함수의 그룹화를 이용한 가상 스피커의 정위감 개선)

  • Seo, Bo-Kug;Cha, Hyung-Tai
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.16 no.6
    • /
    • pp.671-676
    • /
    • 2006
  • A convolution with HRTF DB and the original sound is generally used to make the method of sound image localization for virtual speaker realization. But it can decline localization by the confusion between up and down or front and back directions due to the non-individual HRTF depending on each listener. In this paper, we study a virtual speaker using a new HRTF, which is grouping the HRTF around the virtual speaker to improve localization between up and down or front and back directions. To effective HRTF grouping, we decide the location and number of HRTF using informal listening test. A performance test result of virtual speaker using the grouped HRTF shows that the proposed method improves the front-back and up-down sound localization characteristics much better than the conventional methods.

Fast 360° Sound Source Localization using Signal Energies and Partial Cross Correlation for TDOA Computation

  • Yiwere, Mariam;Rhee, Eun Joo
    • Journal of Information Technology Applications and Management
    • /
    • v.24 no.1
    • /
    • pp.157-167
    • /
    • 2017
  • This paper proposes a simple sound source localization (SSL) method based on signal energies comparison and partial cross correlation for TDOA computation. Many sound source localization methods include multiple TDOA computations in order to eliminate front-back confusion. Multiple TDOA computations however increase the methods' computation times which need to be as minimal as possible for real-time applications. Our aim in this paper is to achieve the same results of localization using fewer computations. Using three microphones, we first compare signal energies to predict which quadrant the sound source is in, and then we use partial cross correlation to estimate the TDOA value before computing the azimuth value. Also, we apply a threshold value to reinforce our prediction method. Our experimental results show that the proposed method has less computation time; spending approximately 30% less time than previous three microphone methods.

Improvement of sound localization for real 3D Sound (현실적인 3D 입체음향 구현을 위한 HRTF의 앞/뒤 음상정위 특성 개선)

  • Koo, Kyo-Sik;Han, Sang-Il;Seo, Bo-Kug;Cha, Hyung-Tai
    • Proceedings of the IEEK Conference
    • /
    • 2007.07a
    • /
    • pp.415-416
    • /
    • 2007
  • HRTF DB, including the information of the sounds which is arrived to our ears, is generally used to make a 3D sound. But it can decline some three-dimensional effects by the confusion between front and back directions due to the non-individual HRTF depending on each listener. In this paper, we propose a new method to use psychoacoustic theory that reduces the confusion of sound image localization. And we make use of an excitation energy by the sense of hearing. This method is brought HRTF spectrum characteristics into relief to draw out the energy ratio about the bark band and control low frequency band. Informal listening tests show that the proposed method improves the front-back sound localization characteristics much better than the conventional methods.

  • PDF

Perceptual Structure of Korean Consonants in High Vowel Contexts (고설 모음 환경에서 한국어 자음의 지각적 구조)

  • Bae, Moon-Jung
    • Phonetics and Speech Sciences
    • /
    • v.1 no.2
    • /
    • pp.95-103
    • /
    • 2009
  • We investigated the perceptual structure of Korean consonants by analyzing the confusion among consonants in various vowel contexts. The 36 CV syllable types combined by 18 consonants and 2 vowels (/i/ and /u/) were presented with masking noises or in degraded intensity. The confusion data were analyzed by the INDSCAL (Individual Difference Scaling), ADCLUS (Additive Clustering) and the probability of the transmitted information. The results were compared with those of a previous study with /a/ vowel context (Bae and Kim, 2002). The overall results showed that the laryngeal features-aspiration, lax and tense-are the most salient features in the perception of Korean consonant regardless of vowel contexts, but the perceptual saliency of place features varies across vowel conditions. In high vowel (front and back vowel) contexts, sibilant consonants were perceptually salient compared to in low vowel contexts. In back vowel contexts, grave (labial and velar) consonants were perceptually salient. These findings imply that place features and vowel features strongly interact in speech perception as well as in speech production. All statistical measures from our confusion data ensured that the perceptual structure of Korean consonants correspond to the hierarchical structure suggested in the feature geometry (Clements, 1991). We discuss the link between speech perception and production as the basis of phonology.

  • PDF

A TDOA Sign-Based Algorithm for Fast Sound Source Localization using an L-Shaped Microphone Array

  • Yiwere, Mariam;Rhee, Eun Joo
    • Journal of Information Technology Applications and Management
    • /
    • v.23 no.3
    • /
    • pp.87-97
    • /
    • 2016
  • This paper proposes a fast sound source localization method using a TDOA sign-based algorithm. We present an L-shaped microphone set-up which creates four major regions in the range of $0^{\circ}{\sim}360^{\circ}$ by the intersection of the positive and negative regions of the individual microphone pairs. Then, we make an initial source region prediction based on the signs of two TDOA estimates before computing the azimuth value. Also, we apply a threshold and angle comparison to tackle the existing front-back confusion problem. Our experimental results show that the proposed method is comparable in accuracy to previous three microphone array methods; however, it takes a shorter computation time because we compute only two TDOA values.

Objective and Subjective Test of a Virtual Sound Reproduction Using a Headphone (헤드폰을 이용한 가상음향 재현의 주관적, 객관적 평가)

  • 최원재;김상명
    • Proceedings of the Korean Society for Noise and Vibration Engineering Conference
    • /
    • 2003.05a
    • /
    • pp.611-616
    • /
    • 2003
  • The headphone is regarded as the most effective means for reproducing 3-dimentional virtual sound due to its channel separation property. However, there still exist several serious problems in headphone reproduction, such as, 'front-back confusion' and in-head localization'. These well-known problems are in general assessed by the subjective test that is based on human judgment. In this paper, an objective test is conducted in parallel with the subject test in order to validate the objective reproduction performance. Such a combined approach may be a more scientific and systematic approach to the reproduction performance.

  • PDF

3-channel HRTF measurement for binaural synthesis. (바이노럴 합성을 위한 3채널 HRTF 측정)

  • Lee Sin-lyul;Kim Lae-hoon;Pang Hee-suk;Sung Koeng-Mo
    • Proceedings of the Acoustical Society of Korea Conference
    • /
    • autumn
    • /
    • pp.337-340
    • /
    • 2000
  • 입체음향 생성을 위한 기존의 방법은 크게 바이노럴 녹음기법과 머리전달함수(HRTF)를 이용한 바이노럴 합성 기법으로 나눌 수 있다. 기존 2채널 더미헤드를 이용한 바이노럴 녹음기법과 바이노럴 합성기법은 표준 더미헤드를 사용함으로써 청취자 머리와의 오차로 정면 음상 정위의 어려움, "Front-back confusion", 이동 음 음상 정위 어려움 등의 문제로 실제 녹음 현장에서는 거의 사용되지 않고 있다. 본 논문에서 제안한 3채널 더미헤드 기법은 이러한 문제점을 극복할 수 있고, 특히, HRTF 합성 시 기존의 HRTF의 문제점을 극복할 수 있는 새로운 HRTF를 구축할 수 있다. 따라서 바이노럴 합성 기법이 필요한 오락, 시뮬레이터, 음장 가청화 기술(Auralization) 프로그램 등 다양한 분야에서의 적용이 가능하다.

  • PDF

Headphone-based multi-channel 3D sound generation using HRTF (HRTF를 이용한 헤드폰 기반의 다채널 입체음향 생성)

  • Kim Siho;Kim Kyunghoon;Bae Keunsung;Choi Songin;Park Manho
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.42 no.1
    • /
    • pp.71-77
    • /
    • 2005
  • In this paper we implement a headphone-based 5.1 channel 3-dimensional (3D) sound generation system using HRTF (Head Related Transfer Function). Each mono sound source in the 5.1 channel signal is localized on its virtual location by binaural filtering with corresponding HRTFs, and reverberation effect is added for spatialization. To reduce the computational burden, we reduce the number of taps in the HRTF impulse response and model the early reverberation effect with several tens of impulses extracted from the whole impulse sequences. We modified the spectrum of HRTF by weighing the difference of front-back spec01m to reduce the front-back confusion caused by non-individualized HRTF DB. In informal listening test we can confirm that the implemented 3D sound system generates live and rich 3D sound compared with simple stereo or 2 channel down mixing.