Abstract
Many of the most effective compression methods involve complicated models. Unfortunately, as model complexity increases, so does the cost of storing the model itself. This paper examines a method to reduce the amount of storage needed to represent a Markov model with an extended alphabet, by applying a clustering scheme that brings together similar states. Experiments run on a variety of large natural language texts show that much of the overhead of storing the model can be saved at the cost of a very small loss of compression efficiency.
Original language | English |
---|---|
Pages (from-to) | 745-760 |
Number of pages | 16 |
Journal | Information Processing and Management |
Volume | 33 |
Issue number | 6 |
DOIs | |
State | Published - Nov 1997 |
Bibliographical note
Funding Information:* The work of the first author (AB) was supported, in part, by NSF Grant IRI-9307895-A01. The author gratefully acknowledges this support. We also wish to acknowledge support given by the Academy of Finland to TR, t To whom all correspondence should be addressed: tel: (773) 702-8268, fax: (773) 702-9861, [email protected]; [email protected], and [email protected]