N-gramモデルのエントロピーに基づくパラメータ削減に関する検討
Bibliographic Information
- Other Title
-
- N gram モデル ノ エントロピー ニ モトヅク パラメータ サクゲン ニ カンスル ケントウ
- A Study on Entropy-based Compression Algorithms for N-gram Parameters
Search this article
Abstract
Large vocabulary continuous speech recognition (LVCSR), which is simply called as dictation, is an essential technology for the realization of voice typing and interface between human being and a computer in various conditions. An LVCSR system reduces search space using language models, where statistical N-gram models are generally used. However, they need a huge number of parameters that grow exponentially with N and the vocabulary size. Especially in the task with large vocabulary (from a few thousand of words to several ten thousands of words), their huge memory requirement results in the system implementation difficulty. In this paper we compare algorithms for reducing the number of parameters of an N-gram model. Preliminary experiments on the augmentation of our compression algorithm to deal with (N-1)-gram are carried out.
Journal
-
- 情報処理学会論文誌
-
情報処理学会論文誌 42 (2), 327-333, 2001-02
情報処理学会
- Tweet
Keywords
Details 詳細情報について
-
- CRID
- 1050014359400009728
-
- NII Article ID
- 110002725740
-
- NII Book ID
- AA12317677
-
- ISSN
- 18827837
- 18827764
- 03875806
-
- HANDLE
- 10061/7764
-
- NDL BIB ID
- 5667528
-
- Text Lang
- ja
-
- Article Type
- journal article
-
- Data Source
-
- IRDB
- NDL
- CiNii Articles