• Title/Summary/Keyword: acoustic features

Search Result 329, Processing Time 0.026 seconds

Application of Acoustic Emission for Assessing Deterioration in Reinforced Concrete Beams (철근 콘크리트 빔의 노화도 평가를 위한 음향방출 기술의 응용)

  • Yoon, Dong-Jin;Park, Phi-Lip;Lee, Seung-Seok
    • Journal of the Korean Society for Nondestructive Testing
    • /
    • v.20 no.4
    • /
    • pp.276-284
    • /
    • 2000
  • The acoustic emission (AE) behavior of reinforced concrete beams tested under flexural loading was investigated to characterize and identify the source of damage. This research was aimed at identifying the characteristic AE response associated with micro-crack development, localized crack propagation, corrosion, and debonding of the reinforcing steel. Concrete beams were prepared to isolate the damage mechanisms by using plain, notched-plain, reinforced, and corroded-reinforced specimens. The beams were tested using four-point cyclic step-loading. The AE response was analyzed to obtain key parameters such as the time history of AE events, the total number and rate of AE events, and the characteristic features of the waveform. Initial analysis of the AE signal has shown that a clear difference in the AE response is observed depending on the source of the damage. The Felicity ratio exhibited a correlation with the overall damage level, while the number of AE events during unloading can be an effective criterion to estimate the level of corrosion distress in reinforced concrete structures. Consequently, AE measurement characterization appears to provide a promising approach for estimating the level of deterioration in reinforced concrete structure.

  • PDF

Investigating an Automatic Method in Summarizing a Video Speech Using User-Assigned Tags (이용자 태그를 활용한 비디오 스피치 요약의 자동 생성 연구)

  • Kim, Hyun-Hee
    • Journal of the Korean Society for Library and Information Science
    • /
    • v.46 no.1
    • /
    • pp.163-181
    • /
    • 2012
  • We investigated how useful video tags were in summarizing video speech and how valuable positional information was for speech summarization. Furthermore, we examined the similarity among sentences selected for a speech summary to reduce its redundancy. Based on such analysis results, we then designed and evaluated a method for automatically summarizing speech transcripts using a modified Maximum Marginal Relevance model. This model did not only reduce redundancy but it also enabled the use of social tags, title words, and sentence positional information. Finally, we compared the proposed method to the Extractor system in which key sentences of a video speech were chosen using the frequency and location information of speech content words. Results showed that the precision and recall rates of the proposed method were higher than those of the Extractor system, although there was no significant difference in the recall rates.

A New Temporal Filtering Method for Improved Automatic Lipreading (향상된 자동 독순을 위한 새로운 시간영역 필터링 기법)

  • Lee, Jong-Seok;Park, Cheol-Hoon
    • The KIPS Transactions:PartB
    • /
    • v.15B no.2
    • /
    • pp.123-130
    • /
    • 2008
  • Automatic lipreading is to recognize speech by observing the movement of a speaker's lips. It has received attention recently as a method of complementing performance degradation of acoustic speech recognition in acoustically noisy environments. One of the important issues in automatic lipreading is to define and extract salient features from the recorded images. In this paper, we propose a feature extraction method by using a new filtering technique for obtaining improved recognition performance. The proposed method eliminates frequency components which are too slow or too fast compared to the relevant speech information by applying a band-pass filter to the temporal trajectory of each pixel in the images containing the lip region and, then, features are extracted by principal component analysis. We show that the proposed method produces improved performance in both clean and visually noisy conditions via speaker-independent recognition experiments.

Constructive music creation: the process and effectiveness of sampling in computer-based electronic music production (구성적 음악 창작: 컴퓨터 기반 전자적 음악 프로덕션 상에서 샘플링의 과정과 효과)

  • Han, Jinseung
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2009.05a
    • /
    • pp.127-134
    • /
    • 2009
  • In spite of controversial debates on aesthetic issues of computer-generated electronic music, rapid advancement of music technologies in the past decade have resulted proliferation of using virtual software synthesizers and samplers in music composition. Computer-based music production platform has become not only a norm among some of contemporary music composers but also vital apparatus for their compositional process. There are two imperative parts of this compositional process involving sampling in computer-based music production, which are commercially available sample libraries that include pre-recorded audio samples, and music production software that processes them. The purpose of this study is to investigate the process and effectiveness of reconstructive compositional process utilizing distinctive features of sampling on computer music production software. This study addresses issues such as: the definition of audio sampling, how sampling is incorporated in compositional process, and what features of music production software are particularly effective in various musical expressions. The result of this study will hopefully accommodate and fulfill the needs of electronic and acoustic musicians' creativeness.

  • PDF

AN ACOUSTIC ANALYSIS ON THE PRONUNCIATION OF KOREAN VOWELS IN PATIENT WITH CLASS III MALOCCLUSION (III급 부정교합 환자의 한국어 모음 발음에 관한 음향학적 분석)

  • Kim, Young-Ho;Yoo, Hyun-Ji;Kim, Whi-Young;Hong, Jong-Rak
    • Journal of the Korean Association of Oral and Maxillofacial Surgeons
    • /
    • v.35 no.4
    • /
    • pp.221-228
    • /
    • 2009
  • The purpose of the study was to investigate the characteristics of the pronunciation of Korean vowels in patients with class III malocclusion. 11 adult male patients with class III malocclusion(mean ages 22.3 years) and four adult males with normal occlusion(mean ages 26.5 years) were selected for the analysis of eight Korean monophthongs /ㅣ, ㅔ, ㅐ, ㅏ, ㅓ, ㅗ, ㅡ, ㅜ/. The values and relationships of F1, F2 and F3 were derived from the stable section of target vowel in each sentence, and the analysis using formant plots and vowel triangles' distance and area was conducted to find the features of two groups' vowel distributions. Consequently, it was identified that the pronunciation of males patients with class III malocclusion showed high values of F1 in the low vowels, high values of F2 in the back vowels, and remarkably low position of /ㅏ/. The vowel triangle suggested that the triangle areas of male patients with class III malocclusion were shown wider vertically and narrower horizontally than those of males with normal occlusion. These characteristics could reflect the structural features of class III malocclusion such as the prognathic mandible, low tongue position, and advancement of back position of the tongue.

The Contribution of Prosody to the Foreign Accent of Chinese Talkers' English Speech

  • Liu, Xing;Lee, Joo-Kyeong
    • Phonetics and Speech Sciences
    • /
    • v.4 no.3
    • /
    • pp.59-73
    • /
    • 2012
  • This study attempts to investigate the contribution of prosody to the foreign accent in Chinese speakers' English production by examining the synthesized speech of crossing native and non-native talkers' prosody and segments. For the stimuli of the foreign accent ratings, we transplanted gender-matched native speakers' prosody onto non-native talkers' segments and vice versa, utilizing the TD-PSOLA algorithm. Eight English native listeners participated in judging foreign accent and comprehensibility of the transplanted stimuli. Results showed that the synthesized stimuli were perceived as stronger foreign accent regardless of speakers' proficiency when English speakers' prosody was crossed with Chinese speakers' segments. This suggests that segments contribute more than prosody to native listeners' evaluation of foreign accent. When transplanted with English speakers' segments, Chinese speakers' prosody showed a difference in duration rather than pitch between high and low proficiency such that stronger foreign accent was detected when low proficient Chinese speakers' duration was crossed with English speakers' segments. This indicated that prosody, more specifically duration, plays a role though the prosodic role is not overall as significant as segments. According to the post acoustic analysis, the temporal features contributing to making the duration parameter prominent as opposed to pitch were found out to be speaking rate, pause duration and pause frequency. Finally, foreign accent and comprehensibility showed no significant correlation such that native listeners had no difficulty listening to highly foreign accented speech.

Acoustic analysis of English lexical stress produced by Korean, Japanese and Taiwanese-Chinese speakers

  • Jung, Ye-Jee;Rhee, Seok-Chae
    • Phonetics and Speech Sciences
    • /
    • v.10 no.1
    • /
    • pp.15-22
    • /
    • 2018
  • Stressed vowels in English are usually produced using longer duration, higher pitch, and greater intensity than unstressed vowels. However, many English as a foreign language (EFL) learners have difficulty producing English lexical stress because their mother tongues do not have such features. In order to investigate if certain non-native English speakers (Korean, Japanese, and Taiwanese-Chinese native speakers) are able to produce English lexical stress in a native-like manner, speech samples were extracted from the L2 learners' corpus known as AESOP (the Asian English Speech cOrpus Project). Sixteen disyllabic words were analyzed in terms of the ratio of duration, pitch, and intensity. The results demonstrate that non-native English speakers are able to produce English stress in a similar way to native English speakers, and all speakers (both native and non-native) show a tendency to use duration as the strongest cue in producing stress. The results also show that the duration ratio of native English speakers was significantly higher than that of non-native speakers, indicating that native speakers produce a bigger difference in duration between stressed and unstressed vowels.

CHROMOSPHERIC MAGNETIC RECONNECTION ON THE SUN

  • CHAE JONGCHUL;CHOI BYUNG-Kyu;PARK MIN-JU
    • Journal of The Korean Astronomical Society
    • /
    • v.35 no.1
    • /
    • pp.59-65
    • /
    • 2002
  • Solar observations support that magnetic reconnect ion ubiquitously occurs in the chromosphere as well as in the corona. It is now widely accepted that coronal magnetic reconnect ion is fast reconnect ion of the Petschek type, and is the main driver of solar flares. On the other hand, it has been thought that the traditional Sweet-Parker model may describe chromospheric reconnect ion without difficulty, since the electric conductivity in the chromoshphere is much lower than that in the corona. However, recent observations of cancelling magnetic features have suggested that chromospheric reconnect ion might proceed at a faster rate than the Sweet-Parker model predicts. We have applied the Sweet-Parker model and Petschek model to a well-observed cancelling magnetic feature. As a result, we found that the inflow speed of the Sweet-Parker reconnect ion is too small to explain the observed converging speed of the feature. On the other hand, the inflow speeds and outflow speeds of the Petschek reconnect ion are well compatible with observations. Moreover, we found that the Sweet-Parker type current sheet is subject to the ion-acoustic instability in the chromosphere, implying the Petschek mechanism may operate there. Our results strongly suggest that chromospheric reconnect ion is of the Petschek type.

On the Beam Focusing Behavior of Time Reversed Ultrasonic Arrays Using a Multi-Gaussian Beam Model

  • Jeong, Hyun-Jo;Lee, Jeong-Sik;Jeong, Yon-Ho;Bae, Sung-Min
    • Journal of the Korean Society for Nondestructive Testing
    • /
    • v.28 no.6
    • /
    • pp.531-537
    • /
    • 2008
  • One of the fundamental features of time reversal acoustic (TRA) techniques is the ability to focus the propagating ultrasonic beam to a specific point within the test material. Therefore, it is important to understand the focusing properties of a TR device in many applications including nondestructive testing. In this paper, we employ an analytical scheme for the analysis of TR beam focusing in a homogeneous medium. More specifically, a nonparaxial multi-Gaussian beam (NMGB) model is used to simulate the focusing behavior of array transducers composed of multiple rectangular elements. The NMGB model is found to generate accurate beam fields beyond the nonparaxial region. Two different simulation cases are considered here for the focal points specified on and off from the central axis of the array transducer. The simulation results show that the focal spot size increases with increasing focal length and focal angle. Furthermore, the maximum velocity amplitude does not always coincide with the specified focal point. Simulation results for the off-axis focusing cases do demonstrate the accurate steering capability of the TR focusing.

Modified Mel Frequency Cepstral Coefficient for Korean Children's Speech Recognition (한국어 유아 음성인식을 위한 수정된 Mel 주파수 캡스트럼)

  • Yoo, Jae-Kwon;Lee, Kyoung-Mi
    • The Journal of the Korea Contents Association
    • /
    • v.13 no.3
    • /
    • pp.1-8
    • /
    • 2013
  • This paper proposes a new feature extraction algorithm to improve children's speech recognition in Korean. The proposed feature extraction algorithm combines three methods. The first method is on the vocal tract length normalization to compensate acoustic features because the vocal tract length in children is shorter than in adults. The second method is to use the uniform bandwidth because children's voice is centered on high spectral regions. Finally, the proposed algorithm uses a smoothing filter for a robust speech recognizer in real environments. This paper shows the new feature extraction algorithm improves the children's speech recognition performance.