Reduction of Dimension of HMM parameters in MLLR Framework for Speaker Adaptation

화자적응시스템을 위한 MLLR 알고리즘 연산량 감소

  • 김지운 (인하대학교 전자공학과 DSP Lab.) ;
  • 정재호 (인하대학교 전자공학과 DSP Lab.)
  • Published : 2003.05.01

Abstract

We discuss how to reduce the number of inverse matrix and its dimensions requested in MLLR framework for speaker adaptation. To find a smaller set of variables with less redundancy, we employ PCA(principal component analysis) and ICA(independent component analysis) that would give as good a representation as possible. The amount of additional computation when PCA or ICA is applied is as small as it can be disregarded. The dimension of HMM parameters is reduced to about 1/3 ~ 2/7 dimensions of SI(speaker independent) model parameter with which speech recognition system represents word recognition rate as much as ordinary MLLR framework. If dimension of SI model parameter is n, the amount of computation of inverse matrix in MLLR is proportioned to O($n^4$). So, compared with ordinary MLLR, the amount of total computation requested in speaker adaptation is reduced to about 1/80~1/150.

Keywords