Abstract for clarkson_icslp98

In Proceedings 5th International Conference on Spoken Language Processing, Sydney, Australia, 1998


P.R. Clarkson and A.J. Robinson

December 1998

Adaptive language models have consistently been shown to lead to a significant reduction in language model perplexity compared to the equivalent static trigram model on many data sets. When these language models have been applied to speech recognition, however, they have seldom resulted in a corresponding reduction in word error rate. This paper will investigate some of the possible reasons for this apparent discrepancy, and will explore the circumstances under which adaptive language models can be useful. We will concentrate on cache-based and mixture-based models and their use on the Broadcast News task.

(ftp:) clarkson_icslp98.ps.gz (http:) clarkson_icslp98.ps.gz
PDF (automatically generated from original PostScript document - may be badly aliased on screen):
  (ftp:) clarkson_icslp98.pdf | (http:) clarkson_icslp98.pdf

If you have difficulty viewing files that end '.gz', which are gzip compressed, then you may be able to find tools to uncompress them at the gzip web site.

If you have difficulty viewing files that are in PostScript, (ending '.ps' or '.ps.gz'), then you may be able to find tools to view them at the gsview web site.

We have attempted to provide automatically generated PDF copies of documents for which only PostScript versions have previously been available. These are clearly marked in the database - due to the nature of the automatic conversion process, they are likely to be badly aliased when viewed at default resolution on screen by acroread.