• Title/Summary/Keyword: 음상정위

Search Result 78, Processing Time 0.02 seconds

A study on sound source segregation of frequency domain binaural model with reflection (반사음이 존재하는 양귀 모델의 음원분리에 관한 연구)

  • Lee, Chai-Bong
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.15 no.3
    • /
    • pp.91-96
    • /
    • 2014
  • For Sound source direction and separation method, Frequency Domain Binaural Model(FDBM) shows low computational cost and high performance for sound source separation. This method performs sound source orientation and separation by obtaining the Interaural Phase Difference(IPD) and Interaural Level Difference(ILD) in frequency domain. But the problem of reflection occurs in practical environment. To reduce this reflection, a method to simulate the sound localization of a direct sound, to detect the initial arriving sound, to check the direction of the sound, and to separate the sound is presented. Simulation results show that the direction is estimated to lie close within 10% from the sound source and, in the presence of the reflection, the level of the separation of the sound source is improved by higher Coherence and PESQ(Perceptual Evaluation of Speech Quality) and by lower directional damping than those of the existing FDBM. In case of no reflection, the degree of separation was low.

Improvement of Head Related Transfer Function to Create Realistic 3D Sound (현실감있는 입체음향 생성을 위한 머리전달함수의 개선)

  • Koo, Kyo-Sik;Cha, Hyung-Tai
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.18 no.3
    • /
    • pp.381-386
    • /
    • 2008
  • Virtual 3D audio methods that create 3D sound effects are researched highly for multimedia devices using 2 speakers or headphone. The most typical method to create 3D effects is a technology through use of head related transfer function (HRTF) which contains the information that sound arrives from a sound source to the ears of the listener. But it can decline some 3D effects by cone of confusion between front and back directions due to the non-individual HRTF depending on each listener. In this paper, we propose a new method to use psychoacoustic theory that creates realistic 3D audio. In order to improve 3D sound, we calculate the excitation energy of each symmetric HRTF and extract the ratio of energy of each bark range. Informal listening tests show that the proposed method improves the front-bach sound localization characteristics much better than the conventional methods.

Modeling of distance localization using by an extended auditory parallax model (확장폭주각 모델을 이용한 음상거리정위의 모델화)

  • KIM Hae-Young;SUZUKI Yoiti;TAKANE Shouichi;SONE Toshio
    • Proceedings of the Acoustical Society of Korea Conference
    • /
    • spring
    • /
    • pp.141-146
    • /
    • 1999
  • This study aims at establishing an digital signal processing technique to control 3-D sound localization, especially focusing our eyes on the role of information provided by Head-Related Transfer Function(HRTF). In order to clarify the cues to control the auditory distance perception, two conventional models named Hirsch-Tahara model and auditory parallax model were examined. As a result, it was shown that both models have limitations to universally explain the auditory distance perception. Hence, the auditory parallax model was extended so as to apply in broader cases of auditory distance perception. The results of the experiment by simulating HRTFs based on the extented parallax model showed that the cues provided by the new model were almost sufficient to control the perception of auditory distance from an actual sound source located within about 2 m.

  • PDF

The Design of Object-based 3D Audio Broadcasting System (객체기반 3차원 오디오 방송 시스템 설계)

  • 강경옥;장대영;서정일;정대권
    • The Journal of the Acoustical Society of Korea
    • /
    • v.22 no.7
    • /
    • pp.592-602
    • /
    • 2003
  • This paper aims to describe the basic structure of novel object-based 3D audio broadcasting system To overcome current uni-directional audio broadcasting services, the object-based 3D audio broadcasting system is designed for providing the ability to interact with important audio objects as well as realistic 3D effects based on the MPEG-4 standard. The system is composed of 6 sub-modules. The audio input module collects the background sound object, which is recored by 3D microphone, and audio objects, which are recorded by monaural microphone or extracted through source separation method. The sound scene authoring module edits the 3D information of audio objects such as acoustical characteristics, location, directivity and etc. It also defines the final sound scene with a 3D background sound, which is intended to be delievered to a receiving terminal by producer. The encoder module encodes scene descriptors and audio objects for effective transmission. The decoder module extracts scene descriptors and audio objects from decoding received bistreams. The sound scene composition module reconstructs the 3D sound scene with scene descriptors and audio objects. The 3D sound renderer module maximizes the 3D sound effects through adapting the final sound to the listner's acoustical environments. It also receives the user's controls on audio objects and sends them to the scene composition module for changing the sound scene.

A Tracking of Head Movement for Stereophonic 3-D Sound (스테레오 입체음향을 위한 머리 움직임 추정)

  • Kim Hyun-Tae;Lee Kwang-Eui;Park Jang-Sik
    • Journal of Korea Multimedia Society
    • /
    • v.8 no.11
    • /
    • pp.1421-1431
    • /
    • 2005
  • There are two methods in 3-D sound reproduction: a surround system, like 3.1 channel method and a binaural system using 2-channel method. The binaural system utilizes the sound localization principle of a human using two ears. Generally, a crosstalk between each channel of 2-channel loudspeaker system should be canceled to produce a natural 3-D sound. To solve this problem, it is necessary to trace a head movement. In this paper, we propose a new algorithm to correctly trace the head movement of a listener. The Proposed algorithm is based on the detection of face and eye. The face detection uses the intensity of an image and the position of eyes is detected by a mathematical morphology. When the head of the listener moves, length of borderline between face area and eyes may change. We use this information to the tracking of head movement. A computer simulation results show That head movement is effectively estimated within +10 margin of error using the proposed algorithm.

  • PDF

IIR Filter Design of HRTF for Real-Time Implementation of 3D Sound by Synthetic Stereo Method (합성 스테레오 방식 3차원 입체음향의 실시간 구현을 위한 머리전달 함수의 IIR 필터 설계)

  • Park Jang-Sik;Kim Hyun-Tae
    • The Journal of the Korea Contents Association
    • /
    • v.5 no.6
    • /
    • pp.74-86
    • /
    • 2005
  • In this paper, we proposed an algorithm for the approximation of high order FIR filters by low order IIR filters to efficient implementing two channel 3-D surround sound systems using Head-related transfer functions(HRTFs). The algorithm is based on a concept of the balanced model reduction. The binaural sounds using the approximated HRTFs are reproduced by headphone, and serves as a cue of sound image localization. HRTFs of dummy-head are approximated from 512-order FIR filters by 32-order IIR filters and compare with each other. .Experiment of sound image are carried out for 10 participants. We perform the experiment based on computer simulation and hardware experiment with TMS320C32. The results of the experiments show that the localization using the approximated HRTFs is the same accuracy as the case of FIR filters that simulate the HRTFs.

  • PDF

A Study for economic improvement of sound image localization and dead zone using computer simulations (컴퓨터 시뮬레이션을 이용한 음의 사각지역 및 음상의 경제적 개선방안 연구)

  • Ko, Eun-Ji;Lee, Hyun-Soo;Lee, Kyung-Ryang;Kim, Seong-Kweon
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.6 no.5
    • /
    • pp.703-708
    • /
    • 2011
  • In this paper, most of the church except for a large church has accommodated many audience to install a balcony floor in a small space. therefore, dead zone and dislocated sound image localization is made due to this under-balcony seats. This paper propose that the problems of dead zone and dislocated sound image localization could be solved using computer simulation in the view of practical side. The economical computer simulation tool, Mapp online that can be found easily was used to the specified church. Installation a sub speaker for dead zone and -10 dB power control of the sub speaker to main speaker power for dislocated sound image localization was proposed. Computer simulation result shows that the value of definition for area was improved from "Normal" to "Very Good" which means about 52% improvement.

Salience of Envelope Interaural Time Difference of High Frequency as Spatial Feature (공간감 인자로서의 고주파 대역 포락선 양이 시간차의 유효성)

  • Seo, Jeong-Hun;Chon, Sang-Bae;Sung, Koeng-Mo
    • The Journal of the Acoustical Society of Korea
    • /
    • v.29 no.6
    • /
    • pp.381-387
    • /
    • 2010
  • Both timbral features and spatial features are important in the assessment of multichannel audio coding systems. The prediction model, extending the ITU-R Rec. BS. 1387-1 to multichannel audio coding systems, with the use of spatial features such as ITDDist (Interaural Time Difference Distortion), ILDDist (Interaural Level Difference Distortion), and IACCDist (InterAural Cross-correlation Coefficient Distortion) was proposed by Choi et al. In that model, ITDDistswere only computed for low frequency bands (below 1500Hz), and ILDDists were computed only for high frequency bands (over 2500Hz) according to classical duplex theory. However, in the high frequency range, information in temporal envelope is also important in spatial perception, especially in sound localization. A new model to compute the ITD distortions of temporal envelopes in high frequency components is introduced in this paper to investigate the role of such ITD on spatial perception quantitatively. The computed ITD distortions of temporal envelopes in high frequency components were highly correlated with perceived sound quality of multichannel audio sounds.