Paper: | SP-P2.2 | ||
Session: | Speaker Adaptation | ||
Time: | Tuesday, May 18, 13:00 - 15:00 | ||
Presentation: | Poster | ||
Topic: | Speech Processing: Adaptation/Normalization | ||
Title: | ADAPTIVE TRAINING USING STRUCTURED TRANSFORMS | ||
Authors: | Kai Yu; Cambridge University | ||
Mark J. F. Gales; Cambridge University | |||
Abstract: | Adaptive training is an important approach to train speech recognition systems on found, non-homogeneous, data. Standard adaptive training employs a single transform to represent unwanted acoustic variability for an utterance. A canonical model representing only the inherent speech variability may then be trained given this set of transforms. For found data there are commonly multiple acoustic factors affecting the speech signal. This paper investigates the use of multiple forms of transformations, structured transforms (ST), to represent the complex non-speech variabilities in an adaptive training framework. Two forms of transform are considered, cluster mean interpolation and constrained MLLR. Re-estimation formulae for estimating the canonical model using both maximum likelihood and minimum phone error training are presented. Experiments to compare ST to standard adaptive training schemes were performed on a conversational telephone speech task. ST were found to significantly reduce the word error rate. | ||
Back |
Home -||-
Organizing Committee -||-
Technical Committee -||-
Technical Program -||-
Plenaries
Paper Submission -||-
Special Sessions -||-
ITT -||-
Paper Review -||-
Exhibits -||-
Tutorials
Information -||-
Registration -||-
Travel Insurance -||-
Housing -||-
Workshops