Entropy and the Consistent Estimation of Joint Distributions
Marton, Katalin ; Shields, Paul C.
Ann. Probab., Tome 22 (1994) no. 4, p. 960-977 / Harvested from Project Euclid
The $k$th-order joint distribution for an ergodic finite-alphabet process can be estimated from a sample path of length $n$ by sliding a window of length $k$ along the sample path and counting frequencies of $k$-blocks. In this paper the problem of consistent estimation when $k = k(n)$ grows as a function of $n$ is addressed. It is shown that the variational distance between the true $k(n)$-block distribution and the empirical $k(n)$-block distribution goes to 0 almost surely for the class of weak Bernoulli processes, provided $k(n) \leq (\log n)/(H + \epsilon)$, where $H$ is the entropy of the process. The weak Bernoulli class includes the i.i.d. processes, the aperiodic Markov chains and functions thereof and the aperiodic renewal processes. A similar result is also shown to hold for functions of irreducible Markov chains. This work sharpens prior results obtained for more general classes of processes by Ornstein and Weiss and by Ornstein and Shields, which used the $\bar{d}$-distance rather than the variational distance.
Publié le : 1994-04-14
Classification:  Empirical distribution,  entropy,  weak Bernoulli processes,  28D20,  60J05,  62B20,  60G10,  94A17
@article{1176988736,
     author = {Marton, Katalin and Shields, Paul C.},
     title = {Entropy and the Consistent Estimation of Joint Distributions},
     journal = {Ann. Probab.},
     volume = {22},
     number = {4},
     year = {1994},
     pages = { 960-977},
     language = {en},
     url = {http://dml.mathdoc.fr/item/1176988736}
}
Marton, Katalin; Shields, Paul C. Entropy and the Consistent Estimation of Joint Distributions. Ann. Probab., Tome 22 (1994) no. 4, pp.  960-977. http://gdmltest.u-ga.fr/item/1176988736/