Abstract for senior_fbrnn

Proc. NIPS '95

FORWARD-BACKWARD RETRAINING OF RECURRENT NEURAL NETWORKS

Andrew Senior and Tony Robinson

November 1995

This paper describes the training of a recurrent neural network as the letter posterior probability estimator for a hidden Markov model, off-line handwriting recognition system. The network estimates posterior distributions for each of a series of frames representing sections of a handwritten word. The supervised training algorithm, backpropagation through time, requires target outputs to be provided for each frame. Three methods for deriving these targets are presented. A novel method based upon the forward-backward algorithm is found to result in the recognizer with the lowest error rate.


(ftp:) senior_fbrnn.ps.gz (http:) senior_fbrnn.ps.gz
PDF (automatically generated from original PostScript document - may be badly aliased on screen):
  (ftp:) senior_fbrnn.pdf | (http:) senior_fbrnn.pdf

If you have difficulty viewing files that end '.gz', which are gzip compressed, then you may be able to find tools to uncompress them at the gzip web site.

If you have difficulty viewing files that are in PostScript, (ending '.ps' or '.ps.gz'), then you may be able to find tools to view them at the gsview web site.

We have attempted to provide automatically generated PDF copies of documents for which only PostScript versions have previously been available. These are clearly marked in the database - due to the nature of the automatic conversion process, they are likely to be badly aliased when viewed at default resolution on screen by acroread.