๐”– Bobbio Scriptorium
โœฆ   LIBER   โœฆ

Discounted likelihood linear regression for rapid speaker adaptation

โœ Scribed by Asela Gunawardana; William Byrne


Book ID
102566853
Publisher
Elsevier Science
Year
2001
Tongue
English
Weight
273 KB
Volume
15
Category
Article
ISSN
0885-2308

No coin nor oath required. For personal study only.

โœฆ Synopsis


The widely used maximum likelihood linear regression speaker adaptation procedure suffers from overtraining when used for rapid adaptation tasks in which the amount of adaptation data is severely limited. This is a well known difficulty associated with the expectation maximization algorithm. We use an information geometric analysis of the expectation maximization algorithm as an alternating minimization of a Kullback-Leibler-type divergence to see the cause of this difficulty, and propose a more robust discounted likelihood estimation procedure. This gives rise to a discounted likelihood linear regression procedure, which is a variant of maximum likelihood linear regression suited for small adaptation sets. Our procedure is evaluated on an unsupervised rapid adaptation task defined on the Switchboard conversational telephone speech corpus, where our proposed procedure improves word error rate by 1.6% (absolute) with as little as 5 seconds of adaptation data, which is a situation in which maximum likelihood linear regression overtrains in the first iteration of adaptation. We compare several realizations of discounted likelihood linear regression with maximum likelihood linear regression and other simple maximum likelihood linear regression variants, and discuss issues that arise in implementing our discounted likelihood procedures.


๐Ÿ“œ SIMILAR VOLUMES


Maximum likelihood linear regression for
โœ C.J. Leggetter; P.C. Woodland ๐Ÿ“‚ Article ๐Ÿ“… 1995 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 148 KB

A method of speaker adaptation for continuous density hidden Markov models (HMMs) is presented. An initial speaker-independent system is adapted to improve the modelling of a new speaker by updating the HMM parameters. Statistics are gathered from the available adaptation data and used to calculate

Empirical Likelihood Confidence Interval
โœ S.X. Chen ๐Ÿ“‚ Article ๐Ÿ“… 1994 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 527 KB

Nonparametric versions of Wilks' theorem are proved for empirical likelihood estimators of slope and mean parameters for a simple linear regression model. They enable us to construct empirical likelihood confidence intervals for these parameters. The coverage errors of these confidence intervals are