• Title/Summary/Keyword: human auditory model

Search Result 39, Processing Time 0.023 seconds

CONCERT HALL ACOUSTICS - Physics, Physiology and Psychology fusing Music and Hall - (콘서트홀 음향 - 음악과 홀을 융합시키는 물리학, 생리학, 심리학 -)

  • 안도요이찌
    • Proceedings of the Acoustical Society of Korea Conference
    • /
    • 1992.06a
    • /
    • pp.3-8
    • /
    • 1992
  • The theory of subjective preference with temporal and spatial factors which include sound signals arriving at both ears is described. Then, auditory evoked potentials which may relate to a primitive subjective response namely subjective preference are discussed. According to such fundamental phenomena, a workable model of human auditory-brain system is proposed. For eample, important subjective attributes, such as loudness, coloration, threshold of preception of a reflection and echo distrubance as well as subjective preference in relation to the initial time delay gap between the direct sound and the first reflection, and the subsequent reverberation time are well described by the autocorrelation function of source signals. Speech clarity, subjective diffuseness as well as subjective preference are related to the magnitude of inter-aural crosscorrelation function (IACC). Even the caktail party effects may be eplained by spatialization of human brain, i.e., independence of temporal and spatial factors.

  • PDF

The Design of Vector Processor for MDCT/IMDCT of MPEG-II AAC (MPEG-II AAC의 MDCT/IMDCT를 위한 벡터 프로세서 설계)

  • 이강현
    • Proceedings of the IEEK Conference
    • /
    • 1999.06a
    • /
    • pp.329-332
    • /
    • 1999
  • Currently, the most important technology is compression methods in the multimedia society. In audio compression, the method using human auditory nervous property is used. This method using psychoacoustical model is applied to perceptual audio coding, because human's audibility is limited. MPEG-II AAC(Advanced Audio Coding) is the most advanced coding scheme that is of benefit to high quality audio coding. The compression ratio is 1.4 times compared with MPEG-I layer-III. In this paper, the vector processor for MDCT/IMDCT(Modified Discrete Cosine Transform /Inverse Modified Discrete Cosine Transform) of MPEG-II AAC is designed.

  • PDF

Speech Recognition Performance Improvement using Gamma-tone Feature Extraction Acoustic Model (감마톤 특징 추출 음향 모델을 이용한 음성 인식 성능 향상)

  • Ahn, Chan-Shik;Choi, Ki-Ho
    • Journal of Digital Convergence
    • /
    • v.11 no.7
    • /
    • pp.209-214
    • /
    • 2013
  • Improve the recognition performance of speech recognition systems as a method for recognizing human listening skills were incorporated into the system. In noisy environments by separating the speech signal and noise, select the desired speech signal. but In terms of practical performance of speech recognition systems are factors. According to recognized environmental changes due to noise speech detection is not accurate and learning model does not match. In this paper, to improve the speech recognition feature extraction using gamma tone and learning model using acoustic model was proposed. The proposed method the feature extraction using auditory scene analysis for human auditory perception was reflected In the process of learning models for recognition. For performance evaluation in noisy environments, -10dB, -5dB noise in the signal was performed to remove 3.12dB, 2.04dB SNR improvement in performance was confirmed.

Implementation a Physical Ear Model for Determinating Location of the Microphone of Fully Implantable Middle Ear Hearing Device (완전 이식형 인공중이용 마이크로폰의 위치 결정을 위한 물리적 귀 모델의 구현)

  • Kim, D.W.;Seong, K.W.;Lim, H.K.;Kim, M.W.;Jung, E.S.;Lee, J.W.;Lee, M.W.;Lee, J.H.;Kim, M.N.;Cho, J.H.
    • Journal of rehabilitation welfare engineering & assistive technology
    • /
    • v.2 no.1
    • /
    • pp.27-33
    • /
    • 2009
  • Generally, implantable microphone has been implanted in the temporal bone for implantable middle ear hearing devices (IMEHDs). In this case, the microphone's membrane can be damaged and can be generated biological noise. In order to overcome the these problems, the location of implanted microphone should be changed. As an alternative, the microphone can be implanted in the external auditory canal. However, the sound emission can be produced because of vibration transducer toward reverse direction from the tympanic membrane to the external auditory canal. In this paper, an amount of the emitted sound is measured using a probe microphone as the changing the position of microphone in the external auditory canal of a physical ear model, which is similar to acoustical and vibratory properties of the human ear. Through the measured value, the location of the microphone was assumed in the external auditory canal. According to the analysis, the microphone input sound can be decreased when microphone position become more distance from the tympanic membrane in the auditory canal. However, the external auditory canal is not appropriated to implantable microphone position, because sound emission is not completely eliminated.

  • PDF

Implementation of Mutual Conversion System between Body Movement and Visual·Auditory Information (신체 움직임-시·청각 정보 상호변환 시스템의 구현)

  • Bae, Myung-Jin;Kim, Sung-Ill
    • Journal of IKEEE
    • /
    • v.22 no.2
    • /
    • pp.362-368
    • /
    • 2018
  • This paper has implemented a mutual conversion system that mutually converts between body motion signals and both visual and auditory signals. The present study is based on intentional synesthesia that can be perceived by learning. The Euler's angle was used in body movements as the output of a wearable armband(Myo). As a muscle sense, roll, pitch and yaw signals were used in this study. As visual and auditory signals, MIDI(Musical Instrument Digital Interface) signals and HSI(Hue, Saturation, Intensity) color model were used respectively. The method of mutual conversion between body motion signals and both visual and auditory signals made it easy to infer by applying one-to-one correspondence. Simulation results showed that input motion signals were compared with output simulation ones using ROS(Root Operation System) and Gazebo which is a 3D simulation tool, to enable the mutual conversion between body motion information and both visual and auditory information.

Speech Processing System Using a Noise Reduction Neural Network Based on FFT Spectrums

  • Choi, Jae-Seung
    • Journal of information and communication convergence engineering
    • /
    • v.10 no.2
    • /
    • pp.162-167
    • /
    • 2012
  • This paper proposes a speech processing system based on a model of the human auditory system and a noise reduction neural network with fast Fourier transform (FFT) amplitude and phase spectrums for noise reduction under background noise environments. The proposed system reduces noise signals by using the proposed neural network based on FFT amplitude spectrums and phase spectrums, then implements auditory processing frame by frame after detecting voiced and transitional sections for each frame. The results of the proposed system are compared with the results of a conventional spectral subtraction method and minimum mean-square error log-spectral amplitude estimator at different noise levels. The effectiveness of the proposed system is experimentally confirmed based on measuring the signal-to-noise ratio (SNR). In this experiment, the maximal improvement in the output SNR values with the proposed method is approximately 11.5 dB better for car noise, and 11.0 dB better for street noise, when compared with a conventional spectral subtraction method.

Design of a Noise Generator for Tinnitus Retraining Therapy Using Auditory Model (청각 모델을 이용한 이명 재훈련 치료용 잡음 발생기의 설계)

  • 이규동;이윤정;김필운;조진호;장용민;이상흔;김명남
    • Journal of Biomedical Engineering Research
    • /
    • v.25 no.5
    • /
    • pp.369-376
    • /
    • 2004
  • The tinnitus retraining therapy(TRT) is an effective method for treating tinnitus patients. This therapy use the white noise to stimulate auditory cells for a wide frequency range. In this paper, the small white noise generator using the thermal noise is proposed. And frequency response controller which can compensate the frequency response changed by the human outer and middle ear system is proposed. We can know that proposed system is more proper type on a purpose of the tinnitus retraining therapy comparing with conventional white noise generator.

An evaluation of the effects of VDT tasks on multiple resources processing in working menory using MD, PD method (MD, PD법을 이용한 VDT 직무의 단기기억 다중자원처리에의 영향평가)

  • 윤철호;노병옥
    • Journal of the Ergonomics Society of Korea
    • /
    • v.16 no.1
    • /
    • pp.85-96
    • /
    • 1997
  • This article reviews the effects of VDT tasks on multiple resources for processing and storage in short-term working memory. MD and PD method were introduced toevaluate the modalities (auditory-visual) in the multiple resources model. The subjects conducted 2 sessions of 50 minites VDT tasks. Before, between and after VDT tasks, MD, PD task performance scores and CFF(critical flicker frequency0 values were measured. The review suggested that the modalities of human information processing in working memory were affected by VDT tasks with different task contents.

  • PDF

Digital Watermarking Using Psychoacoustic Model

  • Poomdaeng, S.;Toomnark, S.;Amornraksa, T.
    • Proceedings of the IEEK Conference
    • /
    • 2002.07b
    • /
    • pp.872-875
    • /
    • 2002
  • A digital watermarking technique applying psychoacoustic model for audio signal is proposed in this paper. In the watermarking scheme, the pseudo-random bit stream used as a watermark signal is embedded into the audio signal in both speech and music. The strength of the embedded signal is subject to the human auditory system in such a way that the disturbances on host audio signal are beyond the sensing of human ears. The experimental results show that the quality of the watermarked audio signal, in term of signal to noise ratio, can be improved up to 3.2 dB.

  • PDF

Speech Feature Extraction based on Spikegram for Phoneme Recognition (음소 인식을 위한 스파이크그램 기반의 음성 특성 추출 기술)

  • Han, Seokhyeon;Kim, Jaewon;An, Soonho;Shin, Seonghyeon;Park, Hochong
    • Journal of Broadcast Engineering
    • /
    • v.24 no.5
    • /
    • pp.735-742
    • /
    • 2019
  • In this paper, we propose a method of extracting speech features for phoneme recognition based on spikegram. The Fourier-transform-based features are widely used in phoneme recognition, but they are not extracted in a biologically plausible way and cannot have high temporal resolution due to the frame-based operation. For better phoneme recognition, therefore, it is desirable to have a new method of extracting speech features, which analyzes speech signal in high temporal resolution following the model of human auditory system. In this paper, we analyze speech signal based on a spikegram that models feature extraction and transmission in auditory system, and then propose a method of feature extraction from the spikegram for phoneme recognition. We evaluate the performance of proposed features by using a DNN-based phoneme recognizer and confirm that the proposed features provide better performance than the Fourier-transform-based features for short-length phonemes. From this result, we can verify the feasibility of new speech features extracted based on auditory model for phoneme recognition.