• Title/Summary/Keyword: Head-related transfer function

Search Result 101, Processing Time 0.03 seconds

Median HRIR Customization via Principal Components Analysis (주성분 분석을 이용한 HRIR 맞춤 기법)

  • Hwang, Sung-Mok;Park, Young-Jin
    • Transactions of the Korean Society for Noise and Vibration Engineering
    • /
    • v.17 no.7 s.124
    • /
    • pp.638-648
    • /
    • 2007
  • A principal components analysis of the entire median HRIRs in the CIPIC HRTF database reveals that the individual HRIRs can be adequately reconstructed by a linear combination of several orthonormal basis functions. The basis functions represent the inter-individual and inter-elevation variations in median HRIRs. There exist elevation-dependent tendencies in the weights of basis functions, and the basis functions can be ordered according to the magnitude of standard deviation of the weights at each elevation. We propose a HRIR customization method via tuning of the weights of 3 dominant basis functions corresponding to the 3 largest standard deviations at each elevation. Subjective listening test results show that both front-back reversal and vertical perception can be improved with the customized HRIRs.

An Acoustic Echo Canceler under 3-Dimensional Synthetic Stereo Environments (3차원 합성 입체음향 환경에서의 음향반향제거기)

  • 김현태;박장식
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.28 no.7A
    • /
    • pp.520-528
    • /
    • 2003
  • This paper proposes a method of implementing synthetic stereo and an acoustic echo cancellation algorithm for multiple participant conference system. Synthetic stereo is generated by HRTF and two loudspeakers. A robust adaptive algorithm for synthetic stereo echo cancellation is proposed to reduce the weight misalignment due to near-end speech signals and ambient noises. The proposed adaptive algorithm is modified version of SMAP algorithm and the coefficients of adaptive filter is updated with cross correlation of input and estimation error signal normalized with sum of the autocorrelation of input signal and the power of the estimation error signal multiplied with projection order. This is more robust to projection order and ambient noise than conventional SMAP. Computer simulation show that the proposed algorithm effectively attenuates synthetic stereo acoustic echo.

Improvement of front/back Sound Localization Characteristics using Psychoacoustics of Head Related Transfer Function (머리전달함수의 심리음향적 특성을 이용한 전/후 음상정위 특성 개선)

  • Koo, Kyo-Sik;Cha, Hyung-Tai
    • Journal of Broadcast Engineering
    • /
    • v.11 no.4 s.33
    • /
    • pp.448-457
    • /
    • 2006
  • HRTF DB, including the information of the sounds which is arrived to our ears, is generally used to make a 3D sound. But it can decline some three-dimensional effects by the confusion between front and back directions due to the non-individual HRTF depending on each listener. In this paper, we propose a new method to use psychoacoustic theory that reduces the confusion of sound image localization. And we make use of an excitation energy by the sense of hearing. This method is brought HRTF spectrum characteristics into relief to draw out the energy ratio about the bark band. Informal listening tests show that the proposed method improves the front-back sound localization characteristics much better than the conventional methods.

Modeling of distance localization using by an extended auditory parallax model (확장폭주각 모델을 이용한 음상거리정위의 모델화)

  • KIM Hae-Young;SUZUKI Yoiti;TAKANE Shouichi;SONE Toshio
    • Proceedings of the Acoustical Society of Korea Conference
    • /
    • spring
    • /
    • pp.141-146
    • /
    • 1999
  • This study aims at establishing an digital signal processing technique to control 3-D sound localization, especially focusing our eyes on the role of information provided by Head-Related Transfer Function(HRTF). In order to clarify the cues to control the auditory distance perception, two conventional models named Hirsch-Tahara model and auditory parallax model were examined. As a result, it was shown that both models have limitations to universally explain the auditory distance perception. Hence, the auditory parallax model was extended so as to apply in broader cases of auditory distance perception. The results of the experiment by simulating HRTFs based on the extented parallax model showed that the cues provided by the new model were almost sufficient to control the perception of auditory distance from an actual sound source located within about 2 m.

  • PDF

Low Dimensional Modeling and Synthesis of Head-Related Transfer Function (HRTF) Using Nonlinear Feature Extraction Methods (비선형 특징추출 기법에 의한 머리전달함수(HRTF)의 저차원 모델링 및 합성)

  • Seo, Sang-Won;Kim, Gi-Hong;Kim, Hyeon-Seok;Kim, Hyeon-Bin;Lee, Ui-Taek
    • The Transactions of the Korea Information Processing Society
    • /
    • v.7 no.5
    • /
    • pp.1361-1369
    • /
    • 2000
  • For the implementation of 3D Sound Localization system, the binaural filtering by HRTFs is generally employed. But the HRTF filter is of high order and its coefficients for all directions have to be stored, which imposes a rather large memory requirement. To cope with this, research works have centered on obtaining low dimensional HRTF representations without significant loss of information and synthesizing the original HRTF efficiently, by means of feature extraction methods for multivariate dat including PCA. In these researches, conventional linear PCA was applied to the frequency domain HRTF data and using relatively small number of principal components the original HRTFs could be synthesized in approximation. In this paper we applied neural network based nonlinear PCA model (NLPCA) and the nonlinear PLS repression model (NLPLS) for this low dimensional HRTF modeling and analyze the results in comparison with the PCA. The NLPCA that performs projection of data onto the nonlinear surfaces showed the capability of more efficient HRTF feature extraction than linear PCA and the NLPLS regression model that incorporates the direction information in feature extraction yielded more stable results in synthesizing general HRTFs not included in the model training.

  • PDF

Enhancement of Sound Image Localization on Vertical Plane for Three-Dimensional Acoustic Synthesis (3차원 음향 합성을 위한 수직면에서의 음상 정위 향상)

  • 김동현;정하영;김기만
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.3 no.3
    • /
    • pp.541-546
    • /
    • 1999
  • The head-related transfer function (HRTF), which expresses the acoustic process from the sound source to the human ears in the free field, contains critical informations which the location of the source can be traced. It also makes it possible to realize multi-dimensional acoustic system that can approximately generate non-existing sound source. The use of non-individual, common HRTF brings performance degradation in localization ability such as front-back judgment error, elevation judgment error. In this paper, we have reduced the error on vertical plane by increasing the spectral notch level. The performance of the proposed method was Proved through subjective test that it is Possible to improve the ability to locate stationary/moving source.

  • PDF

Improvement of virtual speaker localization characteristics using grouped HRTF (머리전달함수의 그룹화를 이용한 가상 스피커의 정위감 개선)

  • Seo, Bo-Kug;Cha, Hyung-Tai
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.16 no.6
    • /
    • pp.671-676
    • /
    • 2006
  • A convolution with HRTF DB and the original sound is generally used to make the method of sound image localization for virtual speaker realization. But it can decline localization by the confusion between up and down or front and back directions due to the non-individual HRTF depending on each listener. In this paper, we study a virtual speaker using a new HRTF, which is grouping the HRTF around the virtual speaker to improve localization between up and down or front and back directions. To effective HRTF grouping, we decide the location and number of HRTF using informal listening test. A performance test result of virtual speaker using the grouped HRTF shows that the proposed method improves the front-back and up-down sound localization characteristics much better than the conventional methods.

Modeling of Distance Localization by Using an Extended Auditory Parallax Model (확장된 음향적 시차 모델을 이용한 음상 거리정위의 모델화)

  • 김해영
    • The Journal of the Acoustical Society of Korea
    • /
    • v.23 no.1
    • /
    • pp.30-39
    • /
    • 2004
  • This study aims at establishing a digital signal processing technique to control 3-D sound localization, especially focusing our ores on the role of information provided by Head-Related Transfer Function (HRTF). In order to clarify the cues to control the auditory distance perception, two conventional models named Hirsch-Tahara model and auditory parallax model were examined. As a result, it was shown that both models have limitations to universally explain the auditory distance perception. Hence, the auditory parallax model was extended so as to apply in broader cases of auditory distance perception. The results of the experiment by simulating HRTFs based on the extended parallax model showed that the cues provided by the new model were almost sufficient to control the perception of auditory distance from an actual sound source located within about 2m.

The Spatial Equalizer$^{(R)}$

  • Kim, Yang-Han;Choe, Jeong-U
    • Broadcasting and Media Magazine
    • /
    • v.16 no.4
    • /
    • pp.31-45
    • /
    • 2011
  • 사용자가 원하는 3D 사운드 혹은 소리의 공간감을 원하는 대로 재현할 수 있는 오디오 시스템은 오랜 기간 동안 인류가 가지고 싶었던 꿈의 기계였다. 그러나 과연 개인 혹은 사용자가 원하는 3D 사운드라는 것이 무엇이며 어떻게 정의하여야 하는지는 명확하지 않다. 이것은 매우 주관적인 개념일 뿐만 아니라 개인에 따라 다를 수 있으며, 그 평가에 대한 객관적인 방법 또한 존재하지 않는다. 관련된 연구를 살펴보면, 원하는 소리의 파동 전파 자체를 시공간 상에서 물리적으로 재현하는 WFS(Wave Field Synthesis)나 Ambisonics, 또는 머리전달함수(HRTF: Head Related Transfer Function)를 기반으로 한 많은 연구들이 있다. 이렇게 재현된 음장(sound field)을 보면 이들이 인지되고 평가되는 등의 객관화를 위하여는 청취 환경에 따라 그 특성이 바뀌고 동일한 환경에서도 청취자에 따라 다르게 인지되는 근본적인 문제점을 가지고 있다. 음장 재현 방법의 이러한 근본적인 문제는 놀랍게도 과거의 스테레오 시스템에서 볼 수 있는 밸런스 노브(balance knob)로부터 그 해결의 실마리를 찾을 수 있다. 밸런스 노브는 보편적인 최적의 소리를 찾는 대신에 청취자가 원하는 음향 효과를 얻을 때까지 직접적으로 소리를 청취하고, 스스로 조절하여 평가할 수 있는 매개체의 역할을 수행한다. 만일 밸런스 노브와 같이 청취자가 원하는 3D 사운드를 스스로 평가하고 조절하기 위한 방법을 마련할 수 있다면? 즉, 청취자가 시공간적으로 원하는 3D 사운드를 실시간으로 청취하고 변화시킬 수 있는 인터페이스를 구현할 수 있다면? 과연 그러한 것이 어떻게 가능할 수 있는지 체계적인 검토가 이루어질 수 있다면 매우 좋을 것이다. 본 고는 이러한 것을 가능케 할 수 있는 즉, 청취자가 자유 자재로 원하는 음장을 형성할 수 있는 렌더링 기법 및 즉각적인 피드백이 가능한 인터페이스를 소개하고 있다. 인터페이스는 현재까지 오디오 시스템에서 주로 사용되는 주파수 이퀄라이져(frequency equalizer)와 매우 유사한 특징이 있다. 이러한 점을 감안하여 "Spatial Equalizer$^{(R)}$"라는 이름을 붙여 보았다. Spatial Equalizer$^{(R)}$는 공간 상에 하나의 점 또는 다수의 점으로 표시되는 가상 음원을 사용자가 조종하여 원 소리의 공간감을 제어할 수 있도록 구성되어 있다. 공간 상에 다수의 점 음원들의 위치를 변화시키거나 크기를 변화시킴으로써 청취자가 원하는 공간감을 구현할 수 있도록 하고 있다. 중요한 것은 종전의 이퀄라이져와 같이 Spatial Equalizer$^{(R)}$에 의해 형성되는 음장이 어떤 객관적인 척도에 의해서 평가되는 대신 사용자에 의해 직접 주관적으로 평가되고, 선택된다는 점이다.

Research for Characteristics of Sound Localization at Monaural System Using Acoustic Energy (청각에너지를 이용한 모노럴 시스템에서의 음상 정위 특성 연구)

  • Koo, Kyo-Sik;Cha, Hyung-Tai
    • The Journal of the Acoustical Society of Korea
    • /
    • v.30 no.4
    • /
    • pp.181-189
    • /
    • 2011
  • According to developments of digital signal processing, 3D sound come into focus on multimedia systems. Many studies on 3d sound have proposed lots of clues to create realistic sounds. But these clues are only focused on binaural systems which two ears are normal. If we make the 3d sound using those clues at monaural systems, the performance goes down dramatically. In order to use the clues for monaural systems, we have studies algorithms such as duplex theory. In duplex theory, the sounds that we listen are affected by human's body, pinna and shoulder. So, we can enhance sound localization performances using its characteristics. In this paper, we propose a new method to use psychoacoustic theory that creates realistic 3D audio at monaural systems. To improve 3d sound, we calculate the excitation energy rates of each symmetric HRTF and extract the weights in each bark range. Finally, they are applied to emphasize the characteristics related to each direction. Informal listening tests show that the proposed method improves sound localization performances much better than the conventional methods.