• Title/Summary/Keyword: Filter bank method

Search Result 169, Processing Time 0.023 seconds

Speech synthesis using acoustic Doppler signal (초음파 도플러 신호를 이용한 음성 합성)

  • Lee, Ki-Seung
    • The Journal of the Acoustical Society of Korea
    • /
    • v.35 no.2
    • /
    • pp.134-142
    • /
    • 2016
  • In this paper, a method synthesizing speech signal using the 40 kHz ultrasonic signals reflected from the articulatory muscles was introduced and performance was evaluated. When the ultrasound signals are radiated to articulating face, the Doppler effects caused by movements of lips, jaw, and chin observed. The signals that have different frequencies from that of the transmitted signals are found in the received signals. These ADS (Acoustic-Doppler Signals) were used for estimating of the speech parameters in this study. Prior to synthesizing speech signal, a quantitative correlation analysis between ADS and speech signals was carried out on each frequency bin. According to the results, the feasibility of the ADS-based speech synthesis was validated. ADS-to-speech transformation was achieved by the joint Gaussian mixture model-based conversion rules. The experimental results from the 5 subjects showed that filter bank energy and LPC (Linear Predictive Coefficient) cepstrum coefficients are the optimal features for ADS, and speech, respectively. In the subjective evaluation where synthesized speech signals were obtained using the excitation sources extracted from original speech signals, it was confirmed that the ADS-to-speech conversion method yielded 72.2 % average recognition rates.

A ProstateSegmentationofTRUS ImageusingSupport VectorsandSnake-likeContour (서포트 벡터와 뱀형상 윤곽선을 이용한 TRUS 영상의 전립선 분할)

  • Park, Jae Heung;Se, Yeong Geon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.17 no.12
    • /
    • pp.101-109
    • /
    • 2012
  • In many diagnostic and treatment procedures for prostate disease accurate detection of prostate boundaries in transrectal ultrasound(TRUS) images is required. This is a challenging and difficult task due to weak prostate boundaries, speckle noise and the short range of gray levels. In this paper a method for automatic prostate segmentation inTRUS images using support vectors and snake-like contour is presented. This method involves preprocessing, extracting Gabor feature, training, and prostate segmentation. Gabor filter bank for extracting the texture features has been implemented. A support vector machine(SVM) for training step has been used to get each feature of prostate and nonprostate. The boundary of prostate is extracted by the snake-like contour algorithm. The results showed that this new algorithm extracted the prostate boundary with less than 9.3% relative to boundary provided manually by experts.

A Study on the Damages of Head Works by the Storm Flood in the Area of Cheong Ju and Boeun -Emphasis onFactors Influenced on the Disasters and their Countermeasures- (淸州 및 報恩地方의 頭首工洪水災害에 關한 調査硏究(II) -災害原因 및 對策方案을 中心으로-)

  • Nam, Seong-Woo;Kim, Choul-Kee
    • Magazine of the Korean Society of Agricultural Engineers
    • /
    • v.24 no.2
    • /
    • pp.49-55
    • /
    • 1982
  • The purpose of this study is to classify the factors influenced on the damages of head works suffered from the storm flood occurred on July 22 1980 in both Musim and Bochong rivers and to find out an integral counter measures against the causes influenced on the disaster of head works in the engineering aspect of planning, design, construction and maintenance. In this survey, number of samples was taken 25 head Works, and the counter measures against the causes of their disasters summarized was as follows, 1. In the aspect of planning a. As the flood water level after the establishment of head works is more increased than the level before setting of head works owing to having more gentle slope of river bed between the head works than nature slope of river bed. Number of head works should be reduced for the appropriate annexation of them b. In the place where head works is established on the curved point of levee, the destruction of levee becomes severe by the strong deflective current. Therefore the setting of head works on the curved point should be kept off as long as possible and in case of unavoidable circumstances the construction method such as reinforced concrete wall or stone wall filed with concrete and anchored bank revetments should be considered. 2. In the aspect of design a. As scoring phenomena at up stream is serious around the weir Where the concentration of strong current is present in such a place, up stream apron having impermeability should be designed to resist and prevent scoring. b. As the length of apron and protected bed is too short to prevent scoring as down stream bed, the design length should be taken somewhat more than the calculated value, but in the case the calculated length becomes too long to be profitable, a device of water cushion should be considered. c. The structure of protected river bed should be improved to make stone mesh bags fixed to apron and to have vinyl mattress laid on river bed together with the improvement for increasing the stability of stone mesh bags and preventing the sucked sand from the river bed. d. As the shortage of cut-off length, especialy in case of the cutoffs conneting both shore sides of river makes the cause of destruction of embankment and weir body, the culculation of cut-off length should be taken enough length based on seepage length. 3. In the aspect of design and constructions a. The overturing destruction of weir by piping action was based on the jet water through cracks at the construction and expansion joints. therefore the expansion joint should be designed and constructed with the insertion of water proof plate and asphalt filling, and the construction joint, with concaved shape structure and steel reinforcement. b. As the wrong design and construction of the weep holes on apron will cause water piping and weir destruction, the design and construction of filter based on the rule of filter should be kept for weep holes. c. The wrong design and construction of bank revetment caused the severe destruction of levee and weir body resulting from scoring and impulse by strong current and formation of water route behind the revetment. Therefore bank revetment should be designod and constructed with stone wall filled with concrete and anchored, or reinforced concrete wall to prevent the formation of water flow route behind the wall and to resist against the scoring and impulse of strong stream. 4. In the aspect of maintenance When the damaged parts occurred at head works the authorities and farmers concerned should find and mend them as soon as possible with mutual cooperation, and on the other hand public citizen should be guided for good use of public property.

  • PDF

An Efficient Algorithm for Improving Calculation Complexity of the MDCT/IMDCT (MDCT/IMDCT의 계산 복잡도를 개선하기 위한 효율적인 알고리즘)

  • 조양기;이원표;김희석
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.40 no.6
    • /
    • pp.106-113
    • /
    • 2003
  • The modified discrete cosine transform (MDCT) and inverse MDCT (IMDCT) are employed in subband/transform coding schemes as the analysis/synthesis filter bank based on time domain aliasing cancellation (TDAC). And the MDCT and IMDCT are the most computational intensive operations in layer III of the MPEG audio coding standard. In this paper, we propose a new efficient algorithm for the MDCT/IMDCT computation in various audio coding systems. It is based on the MDCT/IMDCT computation algorithm using the discrete cosine transforms (DCTs), and It employs two discrete cosine transform of type II (DCT-II) to compute the MDCT/IMDCT In addition, it takes advantage of ability in calculating the MDCT/IMDCT computation, where the length of a data block Is divisible by 4. The Proposed algorithm in this paper requires less calculation complexity than the existing method does. Also, it can be implemented by the parallel structure, therefore its structure is particularly suitable for VLSI realization

Design and Implementation of Wideband Ultra-Fast High Precision Frequency Synthesizer for ELINT Equipment (ELINT 장비용 광대역 초고속 고정밀 주파수 합성기 설계 및 구현)

  • Lee, Kyu-Song;Jeon, Kye-Ik;Oh, Seung-Hyeub
    • The Journal of Korean Institute of Electromagnetic Engineering and Science
    • /
    • v.20 no.11
    • /
    • pp.1178-1185
    • /
    • 2009
  • In this paper, a wideband ultra-high speed & high purity discrete frequency synthesizer having minimum 2.5 MHz step size was proposed. To achieve fast and wideband operation, discrete frequencies were synthesized by mixing of 3 different pre-synthesized 16 frequencies made from fixed PLL and frequency dividers. Frequencies with discrete 2.5 MHz step were produced in 710~1,610 MHz. The measured hopping response time was 350 nsec average, output level was 21.5 dBm average with 2.65 dB flatness, spurious and harmonics level were suppressed below -60 dBc, and phase noise was -94 dBc/Hz@100 Hz. Also, a new measurement method for synthesizer response time was described.

Performance Improvement of Cardiac Disorder Classification Based on Automatic Segmentation and Extreme Learning Machine (자동 분할과 ELM을 이용한 심장질환 분류 성능 개선)

  • Kwak, Chul;Kwon, Oh-Wook
    • The Journal of the Acoustical Society of Korea
    • /
    • v.28 no.1
    • /
    • pp.32-43
    • /
    • 2009
  • In this paper, we improve the performance of cardiac disorder classification by continuous heart sound signals using automatic segmentation and extreme learning machine (ELM). The accuracy of the conventional cardiac disorder classification systems degrades because murmurs and click sounds contained in the abnormal heart sound signals cause incorrect or missing starting points of the first (S1) and the second heart pulses (S2) in the automatic segmentation stage, In order to reduce the performance degradation due to segmentation errors, we find the positions of the S1 and S2 pulses, modify them using the time difference of S1 or S2, and extract a single period of heart sound signals. We then obtain a feature vector consisting of the mel-scaled filter bank energy coefficients and the envelope of uniform-sized sub-segments from the single-period heart sound signals. To classify the heart disorders, we use ELM with a single hidden layer. In cardiac disorder classification experiments with 9 cardiac disorder categories, the proposed method shows the classification accuracy of 81.6% and achieves the highest classification accuracy among ELM, multi-layer perceptron (MLP), support vector machine (SVM), and hidden Markov model (HMM).

Using noise filtering and sufficient dimension reduction method on unstructured economic data (노이즈 필터링과 충분차원축소를 이용한 비정형 경제 데이터 활용에 대한 연구)

  • Jae Keun Yoo;Yujin Park;Beomseok Seo
    • The Korean Journal of Applied Statistics
    • /
    • v.37 no.2
    • /
    • pp.119-138
    • /
    • 2024
  • Text indicators are increasingly valuable in economic forecasting, but are often hindered by noise and high dimensionality. This study aims to explore post-processing techniques, specifically noise filtering and dimensionality reduction, to normalize text indicators and enhance their utility through empirical analysis. Predictive target variables for the empirical analysis include monthly leading index cyclical variations, BSI (business survey index) All industry sales performance, BSI All industry sales outlook, as well as quarterly real GDP SA (seasonally adjusted) growth rate and real GDP YoY (year-on-year) growth rate. This study explores the Hodrick and Prescott filter, which is widely used in econometrics for noise filtering, and employs sufficient dimension reduction, a nonparametric dimensionality reduction methodology, in conjunction with unstructured text data. The analysis results reveal that noise filtering of text indicators significantly improves predictive accuracy for both monthly and quarterly variables, particularly when the dataset is large. Moreover, this study demonstrated that applying dimensionality reduction further enhances predictive performance. These findings imply that post-processing techniques, such as noise filtering and dimensionality reduction, are crucial for enhancing the utility of text indicators and can contribute to improving the accuracy of economic forecasts.

Compression Sensing Technique for Efficient Structural Health Monitoring - Focusing on Optimization of CAFB and Shaking Table Test Using Kobe Seismic Waveforms (효율적인 SHM을 위한 압축센싱 기술 - Kobe 지진파형을 이용한 CAFB의 최적화 및 지진응답실험 중심으로)

  • Heo, Gwang-Hee;Lee, Chin-Ok;Seo, Sang-Gu;Jeong, Yu-Seung;Jeon, Joon-Ryong
    • Journal of the Korea institute for structural maintenance and inspection
    • /
    • v.24 no.2
    • /
    • pp.23-32
    • /
    • 2020
  • The compression sensing technology, CAFB, was developed to obtain the raw signal of the target structure by compressing it into a signal of the intended frequency range. At this point, for compression sensing, the CAFB can be optimized for various reference signals depending on the desired frequency range of the target structure. In addition, optimized CAFB should be able to efficiently compress the effective structural answers of the target structure even in sudden/dangerous conditions such as earthquakes. In this paper, the targeted frequency range for efficient structural integrity monitoring of relatively flexible structures was set below 10Hz, and the optimization method of CAFB for this purpose and the seismic response performance of CAFB in seismic conditions were evaluated experimentally. To this end, in this paper, CAFB was first optimized using Kobe seismic waveform, and embedded it in its own wireless IDAQ system. In addition, seismic response tests were conducted on two span bridges using Kobe seismic waveform. Finally, using an IDAQ system with built-in CAFB, the seismic response of the two-span bridge was wirelessly obtained, and the compression signal obtained was cross-referenced with the raw signal. From the results of the experiment, the compression signal showed excellent response performance and data compression effects in relation to the raw signal, and CAFB was able to effectively compress and sensitize the effective structural response of the structure even in seismic situations. Finally, in this paper, the optimization method of CAFB was presented to suit the intended frequency range (less than 10Hz), and CAFB proved to be an economical and efficient data compression sensing technology for instrumentation-monitoring of seismic conditions.

Feasibility of Deep Learning Algorithms for Binary Classification Problems (이진 분류문제에서의 딥러닝 알고리즘의 활용 가능성 평가)

  • Kim, Kitae;Lee, Bomi;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.95-108
    • /
    • 2017
  • Recently, AlphaGo which is Bakuk (Go) artificial intelligence program by Google DeepMind, had a huge victory against Lee Sedol. Many people thought that machines would not be able to win a man in Go games because the number of paths to make a one move is more than the number of atoms in the universe unlike chess, but the result was the opposite to what people predicted. After the match, artificial intelligence technology was focused as a core technology of the fourth industrial revolution and attracted attentions from various application domains. Especially, deep learning technique have been attracted as a core artificial intelligence technology used in the AlphaGo algorithm. The deep learning technique is already being applied to many problems. Especially, it shows good performance in image recognition field. In addition, it shows good performance in high dimensional data area such as voice, image and natural language, which was difficult to get good performance using existing machine learning techniques. However, in contrast, it is difficult to find deep leaning researches on traditional business data and structured data analysis. In this study, we tried to find out whether the deep learning techniques have been studied so far can be used not only for the recognition of high dimensional data but also for the binary classification problem of traditional business data analysis such as customer churn analysis, marketing response prediction, and default prediction. And we compare the performance of the deep learning techniques with that of traditional artificial neural network models. The experimental data in the paper is the telemarketing response data of a bank in Portugal. It has input variables such as age, occupation, loan status, and the number of previous telemarketing and has a binary target variable that records whether the customer intends to open an account or not. In this study, to evaluate the possibility of utilization of deep learning algorithms and techniques in binary classification problem, we compared the performance of various models using CNN, LSTM algorithm and dropout, which are widely used algorithms and techniques in deep learning, with that of MLP models which is a traditional artificial neural network model. However, since all the network design alternatives can not be tested due to the nature of the artificial neural network, the experiment was conducted based on restricted settings on the number of hidden layers, the number of neurons in the hidden layer, the number of output data (filters), and the application conditions of the dropout technique. The F1 Score was used to evaluate the performance of models to show how well the models work to classify the interesting class instead of the overall accuracy. The detail methods for applying each deep learning technique in the experiment is as follows. The CNN algorithm is a method that reads adjacent values from a specific value and recognizes the features, but it does not matter how close the distance of each business data field is because each field is usually independent. In this experiment, we set the filter size of the CNN algorithm as the number of fields to learn the whole characteristics of the data at once, and added a hidden layer to make decision based on the additional features. For the model having two LSTM layers, the input direction of the second layer is put in reversed position with first layer in order to reduce the influence from the position of each field. In the case of the dropout technique, we set the neurons to disappear with a probability of 0.5 for each hidden layer. The experimental results show that the predicted model with the highest F1 score was the CNN model using the dropout technique, and the next best model was the MLP model with two hidden layers using the dropout technique. In this study, we were able to get some findings as the experiment had proceeded. First, models using dropout techniques have a slightly more conservative prediction than those without dropout techniques, and it generally shows better performance in classification. Second, CNN models show better classification performance than MLP models. This is interesting because it has shown good performance in binary classification problems which it rarely have been applied to, as well as in the fields where it's effectiveness has been proven. Third, the LSTM algorithm seems to be unsuitable for binary classification problems because the training time is too long compared to the performance improvement. From these results, we can confirm that some of the deep learning algorithms can be applied to solve business binary classification problems.