Compression Algorithm Of Trigram Language Models Based On Maximum Likelihood Estimation

Description

In this paper we propose an algorithm for reducing the size of back-off N-gram models, with less affecting its performance than the traditional cutoff method. The algorithm is based on the Maximum Likelihood (ML) estimation and realizes an N-gram language model with a given number of N-gram probability parameters that minimize the training set perplexity. To confirm the effectiveness of our algorithm, we apply it to trigram and bigram models, and the experiments in terms of perplexity and word error rate in a dictation system are carried out.

Journal

Citations (2)*help

See more

Details 詳細情報について

Report a problem

Back to top