• Title/Summary/Keyword: Sound Interpolation

Search Result 21, Processing Time 0.018 seconds

A Sound Interpolation Method Using Deep Neural Network for Virtual Reality Sound (가상현실 음향을 위한 심층신경망 기반 사운드 보간 기법)

  • Choi, Jaegyu;Choi, Seung Ho
    • Journal of Broadcast Engineering
    • /
    • v.24 no.2
    • /
    • pp.227-233
    • /
    • 2019
  • In this paper, we propose a deep neural network-based sound interpolation method for realizing virtual reality sound. Through this method, sound between two points is generated by using acoustic signals obtained from two points. Sound interpolation can be performed by statistical methods such as arithmetic mean or geometric mean, but this is insufficient to reflect actual nonlinear acoustic characteristics. In order to solve this problem, in this study, the sound interpolation is performed by training the deep neural network based on the acoustic signals of the two points and the target point, and the experimental results show that the deep neural network-based sound interpolation method is superior to the statistical methods.

A Relevant Distortion Criterion for Interpolation of the Head-Related Transfer Functions (머리 전달 함수의 보간에 적합한 왜곡 척도)

  • Lee, Ki-Seung;Lee, Seok-Pil
    • The Journal of the Acoustical Society of Korea
    • /
    • v.28 no.2
    • /
    • pp.85-95
    • /
    • 2009
  • In the binaural synthesis environments, wide varieties of the head-related transfer functions (HRTFs) that have measured with a various direction would be desirable to obtain the accurate and various spatial sound images. To reduce the size' of HRTFs, interpolation has been often employed, where the HRTF for any direction is obtained by a limited number of the representative HRTFs. In this paper, we study on the distortion measures for interpolation, which has an important role in interpolation. With lhe various objective distortion metrics, the differences between the interpolated and the measured HRTFs were computed. These were then compared and analyzed with the results from the listening tests. From the results, the objective distortion measures were selected, that reflected the perceptual differences in spatial sound image. This measure was employed in a practical interpolation technique. We applied the proposed method to four kinds of an HRTF set, measured from three human heads and one mannequin. As a result, the Mel-frequency cepstral distortion was shown to be a good predictor for the differences in spatial sound location, when three HRTF measured from human, and the time-domain signal to distortion ratio revealed good prediction results for the entire four HRTF sets.

HRTF Interpolation Using a Spherical Head Model (원형 머리 모델을 이용한 머리 전달 함수의 보간)

  • Lee, Ki-Seung;Lee, Seok-Pil
    • The Journal of the Acoustical Society of Korea
    • /
    • v.27 no.7
    • /
    • pp.333-341
    • /
    • 2008
  • In this paper, a new interpolation model for the head related transfer function (HRTF) was proposed. In the method herein, we assume that the impulse response of the HRTF for each azimuth angle is given by linear interpolation of the time-delayed neighboring impulse responses of HRTFs. The time delay of the HRTF for each azimuth angle is given by sum of the sound wave propagation time from the ears to the sound source, which can be estimated by using azimuth angle, the physical shape of the underlying head and the distance between the head and sound source, and the refinement time yielding the minimum mean square error. Moreover, in the proposed model, the interpolation intervals were not fixed but varied, which were determined by minimizing the total number of HRTFs while the synthesized signals have no perceptual difference from the original signals in terms of sound location. To validate the usefulness of the proposed interpolation model, the proposed model was applied to the several HRTFs that were obtained from one dummy-head and three human heads. We used the HRTFs that have 5 degree azimuth angle resolution at 0 degree elevation (horizontal plane). The experimental results showed that using only $30\sim40%$ of the original HRTFs were sufficient for producing the signals that have no audible differences from the original ones in terms of sound location.

The study of sound source synthesis IC to realize the virtual engine sound of a car powered by electricity without an engine (엔진 없이 전기로 구동되는 자동차의 가상 엔진 음 구현을 위한 음원합성 IC에 관한 연구)

  • Koo, Jae-Eul;Hong, Jae-Gyu;Song, Young-Woog;Lee, Gi-Chang
    • The Journal of the Acoustical Society of Korea
    • /
    • v.40 no.6
    • /
    • pp.571-577
    • /
    • 2021
  • This study is a study on System On Chip (SOC) that implements virtual engine sound in electric vehicles without engines, and realizes vivid engine sound by combining Adaptive Difference PCM (ADPCM) method and frequency modulation method for satisfaction of driver's needs and safety of pedestrians. In addition, by proposing an electronic sound synthesis algorithm applying Musical Instrument Didital Interface (MIDI), an engine sound synthesis method and a constitutive model of an engine sound generation system are presented. In order to satisfy both drivers and pedestrians, this study uses Controller Area Network (CAN) communication to receive information such as Revolution Per Minute (RPM), vehicle speed, accelerator pedal depressed amount, torque, etc., transmitted according to the driver's driving habits, and then modulates the frequency according to the appropriate preset parameters We implemented an interaction algorithm that accurately reflects the intention of the system and driver by using interpolation for the system, ADPCM algorithm for reducing the amount of information, and MIDI format information for making engine sound easier.

HRTF Improvement of Smooth Moving Sound Effect (자연스러운 이동음 효과를 위한 머리전달함수 개선)

  • Lee, Jung-Kyoung;Seo, Bo-Kug;Cha, Hyung-Tai
    • Proceedings of the IEEK Conference
    • /
    • 2007.07a
    • /
    • pp.427-428
    • /
    • 2007
  • In this paper, we propose the method of generating smooth moving sound in a two-channel based 3D sound technique. HRTF is generally used to make the moving sound effect in frequency domain. But, the moving sound using the interpolation has many problems to be resolved as there are only discrete measured point of HRTF database. Therefore, HRTF of the moving sound channel makes effective moving sound by the improved grouping. Informal listening tests show that the proposed improves the effective moving sound much better than the conventional methods.

  • PDF

An Improvement of Strain Measuring Technique by using the B-spline Surface Interpolation Method (3차원 곡면 내삽법을 이용한 자동차 박판 부품의 변형율 측정법 개선)

  • 김종봉;양동열
    • Transactions of the Korean Society of Automotive Engineers
    • /
    • v.6 no.4
    • /
    • pp.141-150
    • /
    • 1998
  • The measurement of strains in stamped sheet metal is essential to the design and manufacture of sound sheet metal products. The measured strains can also be used in verifying the reliability of the computer analysis such as finite element analysis. In most engineering applications, strains are measured from the deformed square grids or deformed circular grids in comparison with the initial undeformed grids. In such a case, however, strains are averaged in each grid and the localized strain in a region smaller than a grid size can not be measured. In the present study, the B-spline surface interpolation technique is introduced in order to measure the strains more exactly and effectively. The strains calculated by using the surface interpolation technique are compared with the strains calculated from the three-noded grids as well as with the finite element analysis.

  • PDF

Implementation of Text-to-Audio Visual Speech Synthesis Using Key Frames of Face Images (키프레임 얼굴영상을 이용한 시청각음성합성 시스템 구현)

  • Kim MyoungGon;Kim JinYoung;Baek SeongJoon
    • MALSORI
    • /
    • no.43
    • /
    • pp.73-88
    • /
    • 2002
  • In this paper, for natural facial synthesis, lip-synch algorithm based on key-frame method using RBF(radial bases function) is presented. For lips synthesizing, we make viseme range parameters from phoneme and its duration information that come out from the text-to-speech(TTS) system. And we extract viseme information from Av DB that coincides in each phoneme. We apply dominance function to reflect coarticulation phenomenon, and apply bilinear interpolation to reduce calculation time. At the next time lip-synch is performed by playing the synthesized images obtained by interpolation between each phonemes and the speech sound of TTS.

  • PDF

Development of Range-Dependent Ray Model for Sonar Simulator (소나 시뮬레이터용 거리 종속 음선 모델 개발)

  • Jung, Young-Cheol;Lee, Keunhwa;Seong, Woojae;Kim, Hyoung-Rok
    • The Journal of the Acoustical Society of Korea
    • /
    • v.33 no.3
    • /
    • pp.163-173
    • /
    • 2014
  • Sound propagation algorithm for a sonar simulator is required to run in real-time and should be able to model the range and depth dependence of the Korean ocean environments. Ray model satisfies these requirements and we developed an algorithm for range-dependent ocean environments. In this algorithm, we considered depth-dependence of sound speed through rays based on a rectangular cell method and layer method. Range-dependence of sound speed was implemented based on a split-step method in the range direction. Eigen-ray is calculated through an interpolation of ray bundles and Gaussian interpolation function was used. The received time signal of sonar was simulated by Fourier transform of eigen-ray solution in the frequency domain. Finally, for the verification of proposed algorithm, we compared the results of transmission loss with other validated models such as BELLHOP, SNUPE, KRAKEN and OASES, for the Pekeris waveguide, wedge, and deep ocean environments. As a result, we obtained satisfactory agreements among them.

Regeneration of the Retarded Time Vector for Enhancing the Precision of Acoustic Pyrometry (온도장 측정 정밀도 향상을 위한 시간 지연 벡터의 재형성)

  • Kim, Tae-Kyoon;Ih, Jeong-Guon
    • The Journal of the Acoustical Society of Korea
    • /
    • v.33 no.2
    • /
    • pp.118-125
    • /
    • 2014
  • An approximation of speed of sound in the measurement plane is essential for the inverse estimation of temperature. To this end, an inverse problem relating the measured retarded time data in between set of sensors and actuators array located on the wall is formulated. The involved transfer matrix and its coefficient vectors approximate speed of sound of the measurement plane by using the radial basis function with finite number of interpolation points deployed inside the target field. Then, the temperature field can be reconstructed by using spatial interpolation technique, which can achieve high spatial resolution with proper number of interpolation points. A large number of retarded time data of acoustic paths in between sensors and arrays are needed to obtain accurate reconstruction result. However, the shortage of interpolation points due to practical limitations can cause the decrease of spatial resolution and deterioration of the reconstruction result. In this works, a regeneration for obtaining the additional retarded time data for an arbitrary acoustic path is suggested to overcome the shortage of interpolation points. By applying the regeneration technique, many interpolation points can be deployed inside the field by increasing the number of retarded time data. As a simulation example, two rectangular duct sections having arbitrary temperature distribution are reconstructed by two different data set: measured data only, combination of measured and regenerated data. The result shows a decrease in reconstruction error by 15 % by combining the original and regenerated retarded time data.