Codes based on inaccurate source probabilities
- 1 May 1971
- journal article
- Published by Institute of Electrical and Electronics Engineers (IEEE) in IEEE Transactions on Information Theory
- Vol. 17 (3) , 304-314
- https://doi.org/10.1109/tit.1971.1054638
Abstract
Information theory obtains efficient codes by encoding messages in large blocks. The code design requires block probabilities that are often hard to measure accurately. This paper studies the effect of inaccuracies in the block probabilities and gives coding procedures that anticipate some of the worst errors. For an efficient code, the mean numberdof digits per letter must be kept small. In some cases the expected value ofdcan be related to the size of the sample on which probability estimates are based. To underestimate badly the probability of a common letter or block is usually a serious error. To ensure against this possibility, some coding procedures are given that avoid extremely long codewords. These codes provide a worthwhile insurance but are still very efficient if the probability estimates happen to be correct.Keywords
This publication has 7 references indexed in Scilit:
- Minimum-redundancy coding for the discrete noiseless channelIEEE Transactions on Information Theory, 1961
- Variable-Length Binary EncodingsBell System Technical Journal, 1959
- On a Statistical Estimate for the Entropy of a Sequence of Independent Random VariablesTheory of Probability and Its Applications, 1959
- Two inequalities implied by unique decipherabilityIEEE Transactions on Information Theory, 1956
- A Method for the Construction of Minimum-Redundancy CodesProceedings of the IRE, 1952
- A Mathematical Theory of CommunicationBell System Technical Journal, 1948
- Relativ Frequency of English Speech SoundsPublished by Harvard University Press ,1923