Abstract for niesler_icslp96

Proc. ICSLP '96.


T.R. Niesler and P.C. Woodland

October 1996

A language model combining word-based and category-based n-grams within a backoff framework is presented. Word n-grams conveniently capture sequential relations between particular words, while the category-model, which is based on part-of-speech classifications and allows ambiguous category membership, is able to generalise to unseen word sequences and therefore appropriate in backoff situations. Experiments on the LOB, Switchboard and WSJ0 corpora demonstrate that the technique greatly improves language model perplexities for sparse training sets, and offers significantly improved complexity versus performance tradeoffs when compared with standard trigram models.

(ftp:) niesler_icslp96.ps.Z (http:) niesler_icslp96.ps.Z
PDF (automatically generated from original PostScript document - may be badly aliased on screen):
  (ftp:) niesler_icslp96.pdf | (http:) niesler_icslp96.pdf

If you have difficulty viewing files that end '.gz', which are gzip compressed, then you may be able to find tools to uncompress them at the gzip web site.

If you have difficulty viewing files that are in PostScript, (ending '.ps' or '.ps.gz'), then you may be able to find tools to view them at the gsview web site.

We have attempted to provide automatically generated PDF copies of documents for which only PostScript versions have previously been available. These are clearly marked in the database - due to the nature of the automatic conversion process, they are likely to be badly aliased when viewed at default resolution on screen by acroread.