The Reduction or computation in MLLR Framework using PCA or ICA for Speaker Adaptation

화자적응에서 PCA 또는 ICA를 이용한 MLLR알고리즘 연산량 감소

  • 김지운 (인하대학교 전자공학과 DSP Lab) ;
  • 정재호 (인하대학교 전자공학과 DSP Lab)
  • Published : 2003.08.01

Abstract

We discuss how to reduce the number of inverse matrix and its dimensions requested in MLLR framework for speaker adaptation. To find a smaller set of variables with less redundancy, we adapt PCA (principal component analysis) and ICA (independent component analysis) that would give as good a representation as possible. The amount of additional computation when PCA or ICA is applied is as small as it can be disregarded. 10 components for ICA and 12 components for PCA represent similar performance with 36 components for ordinary MLLR framework. If dimension of SI model parameter is n, the amount of computation of inverse matrix in MLLR is proportioned to O(n⁴). So, compared with ordinary MLLR, the amount of total computation requested in speaker adaptation is reduced by about 1/81 in MLLR with PCA and 1/167 in MLLR with ICA.

본 논문은 화자 적응시 화자 독립 모델의 차수를 줄이고 MLLR (Maximum Likelihood Linear Regression) 알고리즘에서 요구되는 역행렬 횟수를 줄이는 방법을 제안한다. 주성분분석 (PCA: principal components analysis)과 독립성분분석 (ICA: independent components analysis)을 통해 모델 혼합성분 (mixture component)들간의 상관관계를 줄임으로서 모델의 차수를 감소하였다. 주성분분석 및 독립성분분석에 요구되는 추가 연산량은 화자 독립 모델을 훈련할 때 추가함으로써 화자 적응시에 추가되는 연산량은 극히 미소하다. 36차의 HMM 파라메타 차수를 PCA는 12차, ICA는 10차로 감소하였을 때 기존의 MLLR 적응방법과 유사한 단어 인식률을 나타내었다. 즉, 모델 파라미터의 차수를 n이라고 할 때 기존의 MLLR알고리즘에서 역행열 연산에서 요구되는 연산량은 O(n⁴)에 비례하므로 PCA는 1/81, ICA는 1/167만큼 연산량을 감소하였다.

Keywords

References

  1. PhD Thesis, Univ. of Cambridge Improved acoustic modelling for HMMs using linear transforms C.J.Leggetter
  2. IEEE Trans. on Signal Processing v.39 no.4 A study on speaker adaptation of the parameters of continuous density hidden Markov models C.H.Lee;C.H.Lin;B.H.Juang https://doi.org/10.1109/78.80902
  3. IEEE Trans. on Speech and Audio Processing v.9 no.4 Joint maximum a posteriori adaptation of transformation and HMM parameters O.Siohan;C.Chesta;C.H.Lee https://doi.org/10.1109/89.917687
  4. IEEE Trans. on Speech and Audio Processing v.7 no.6 Online hierarchical transformation of hidden Markov models for speech recognition J.T.Chien https://doi.org/10.1109/89.799691
  5. Independent Component Analysis A.Hyvarinen;J.Karhunen;E.Oja
  6. Journal of the Royal Statistical Society v.Series B no.61 Probabilistic principal component analysis M.E.Tipping;C.M.Bishop
  7. The Eleventh British Machine Vision Conference Probabilistic PCS and ICA subspace mixture models for image segmentation D.Ridder;J.Kittler;R.P.W.Duin
  8. IEEE Trans. on Speech Audio Processing v.4 A maximum-likelihood approach to stochastic matching for robust speech recognition A.Sankar;C.H.Lee https://doi.org/10.1109/89.496215
  9. Proc. 5th Eur. Conf. Speech Communication and Technology v.4 On-line adaptaion of hidden Markov models using incremental estimation algorithms V.Digalakis