Given a discrete-time stationary ergodic stochastic process X {\displaystyle X} on the probability space ( Ω , B , p ) {\displaystyle (\Omega ,B,p)} , the asymptotic equipartition property is an assertion that, almost surely, − 1 n log p ( X 1 , X 2 , … , X n ) → H ( X ) as n → ∞ {\displaystyle -{\frac {1}{n}}\log p(X_{1},X_{2},\dots ,X_{n})\to H(X)\quad {\text{ as }}\quad n\to \infty } where H ( X ) {\displaystyle H(X)} or simply H {\displaystyle H} denotes the entropy rate of X {\displaystyle X} , which must exist for all discrete-time stationary processes including the ergodic ones. The asymptotic equipartition property is proved for finite-valued (i.e. | Ω | < ∞ {\displaystyle |\Omega |<\infty } ) stationary ergodic stochastic processes in the Shannon–McMillan–Breiman theorem using the ergodic theory and for any i.i.d. sources directly using the law of large numbers in both the discrete-valued case (where H {\displaystyle H} is simply the entropy of a symbol) and the continuous-valued case (where H {\displaystyle H} is the differential entropy instead). The definition of the asymptotic equipartition property can also be extended for certain classes of continuous-time stochastic processes for which a typical set exists for long enough observation time. The convergence is proven almost sure in all cases.
Given X {\displaystyle X} is an i.i.d. source which may take values in the alphabet X {\displaystyle {\mathcal {X}}} , its time series X 1 , … , X n {\displaystyle X_{1},\ldots ,X_{n}} is i.i.d. with entropy H ( X ) {\displaystyle H(X)} . The weak law of large numbers gives the asymptotic equipartition property with convergence in probability, lim n → ∞ Pr [ | − 1 n log p ( X 1 , X 2 , … , X n ) − H ( X ) | > ε ] = 0 ∀ ε > 0. {\displaystyle \lim _{n\to \infty }\Pr \left[\left|-{\frac {1}{n}}\log p(X_{1},X_{2},\ldots ,X_{n})-H(X)\right|>\varepsilon \right]=0\qquad \forall \varepsilon >0.} since the entropy is equal to the expectation of 1 − 1 n log p ( X 1 , X 2 , … , X n ) . {\displaystyle -{\frac {1}{n}}\log p(X_{1},X_{2},\ldots ,X_{n}).}
The strong law of large numbers asserts the stronger almost sure convergence, Pr [ lim n → ∞ − 1 n log p ( X 1 , X 2 , … , X n ) = H ( X ) ] = 1. {\displaystyle \Pr \left[\lim _{n\to \infty }-{\frac {1}{n}}\log p(X_{1},X_{2},\ldots ,X_{n})=H(X)\right]=1.} Convergence in the sense of L1 asserts an even stronger E [ | lim n → ∞ − 1 n log p ( X 1 , X 2 , … , X n ) − H ( X ) | ] = 0 {\displaystyle \mathbb {E} \left[\left|\lim _{n\to \infty }-{\frac {1}{n}}\log p(X_{1},X_{2},\ldots ,X_{n})-H(X)\right|\right]=0}
Consider a finite-valued sample space Ω {\displaystyle \Omega } , i.e. | Ω | < ∞ {\displaystyle |\Omega |<\infty } , for the discrete-time stationary ergodic process X := { X n } {\displaystyle X:=\{X_{n}\}} defined on the probability space ( Ω , B , p ) {\displaystyle (\Omega ,B,p)} . The Shannon–McMillan–Breiman theorem, due to Claude Shannon, Brockway McMillan, and Leo Breiman, states that we have convergence in the sense of L1.2 Chung Kai-lai generalized this to the case where X {\displaystyle X} may take value in a set of countable infinity, provided that the entropy rate is still finite.3
The assumptions of stationarity/ergodicity/identical distribution of random variables is not essential for the asymptotic equipartition property to hold. Indeed, as is quite clear intuitively, the asymptotic equipartition property requires only some form of the law of large numbers to hold, which is fairly general. However, the expression needs to be suitably generalized, and the conditions need to be formulated precisely.
Consider a source that produces independent symbols, possibly with different output statistics at each instant, for which the statistics of the process are known completely, that is, the marginal distribution of the process seen at each time instant is known. The joint distribution is just the product of marginals. Then, under the condition (which can be relaxed) that V a r [ log p ( X i ) ] < M {\displaystyle \mathrm {Var} [\log p(X_{i})]<M} for all i, for some M > 0, the following holds (AEP): lim n → ∞ Pr [ | − 1 n log p ( X 1 , X 2 , … , X n ) − H ¯ n ( X ) | < ε ] = 1 ∀ ε > 0 {\displaystyle \lim _{n\to \infty }\Pr \left[\,\left|-{\frac {1}{n}}\log p(X_{1},X_{2},\ldots ,X_{n})-{\overline {H}}_{n}(X)\right|<\varepsilon \right]=1\qquad \forall \varepsilon >0} where H ¯ n ( X ) = 1 n H ( X 1 , X 2 , … , X n ) {\displaystyle {\overline {H}}_{n}(X)={\frac {1}{n}}H(X_{1},X_{2},\ldots ,X_{n})}
The proof follows from a simple application of Markov's inequality (applied to second moment of log ( p ( X i ) ) {\displaystyle \log(p(X_{i}))} . Pr [ | − 1 n log p ( X 1 , X 2 , … , X n ) − H ¯ ( X ) | > ε ] ≤ 1 n 2 ε 2 V a r [ ∑ i = 1 n ( log ( p ( X i ) ) 2 ] ≤ M n ε 2 → 0 as n → ∞ {\displaystyle {\begin{aligned}\Pr \left[\left|-{\frac {1}{n}}\log p(X_{1},X_{2},\ldots ,X_{n})-{\overline {H}}(X)\right|>\varepsilon \right]&\leq {\frac {1}{n^{2}\varepsilon ^{2}}}\mathrm {Var} \left[\sum _{i=1}^{n}\left(\log(p(X_{i})\right)^{2}\right]\\&\leq {\frac {M}{n\varepsilon ^{2}}}\to 0{\text{ as }}n\to \infty \end{aligned}}}
It is obvious that the proof holds if any moment E [ | log p ( X i ) | r ] {\displaystyle \mathrm {E} \left[|\log p(X_{i})|^{r}\right]} is uniformly bounded for r > 1 (again by Markov's inequality applied to r-th moment). Q.E.D.
Even this condition is not necessary, but given a non-stationary random process, it should not be difficult to test whether the asymptotic equipartition property holds using the above method.
The asymptotic equipartition property for non-stationary discrete-time independent process leads us to (among other results) the source coding theorem for non-stationary source (with independent output symbols) and noisy-channel coding theorem for non-stationary memoryless channels.
T {\textstyle T} is a measure-preserving map on the probability space Ω {\textstyle \Omega } .
If P {\textstyle P} is a finite or countable partition of Ω {\textstyle \Omega } , then its entropy is H ( P ) := − ∑ p ∈ P μ ( p ) ln μ ( p ) {\displaystyle H(P):=-\sum _{p\in P}\mu (p)\ln \mu (p)} with the convention that 0 ln 0 = 0 {\displaystyle 0\ln 0=0} .
We only consider partitions with finite entropy: H ( P ) < ∞ {\textstyle H(P)<\infty } .
If P {\textstyle P} is a finite or countable partition of Ω {\textstyle \Omega } , then we construct a sequence of partitions by iterating the map: P ( n ) := P ∨ T − 1 P ∨ ⋯ ∨ T − ( n − 1 ) P {\displaystyle P^{(n)}:=P\vee T^{-1}P\vee \dots \vee T^{-(n-1)}P} where P ∨ Q {\textstyle P\vee Q} is the least upper bound partition, that is, the least refined partition that refines both P {\textstyle P} and Q {\textstyle Q} : P ∨ Q := { p ∩ q : p ∈ P , q ∈ Q } {\displaystyle P\vee Q:=\{p\cap q:p\in P,q\in Q\}} Write P ( x ) {\textstyle P(x)} to be the set in P {\textstyle P} where x {\textstyle x} falls in. So, for example, P ( n ) ( x ) {\textstyle P^{(n)}(x)} is the n {\textstyle n} -letter initial segment of the ( P , T ) {\textstyle (P,T)} name of x {\textstyle x} .
Write I P ( x ) {\textstyle I_{P}(x)} to be the information (in units of nats) about x {\textstyle x} we can recover, if we know which element in the partition P {\textstyle P} that x {\textstyle x} falls in: I P := − ln μ ( P ( x ) ) {\displaystyle I_{P}:=-\ln \mu (P(x))} Similarly, the conditional information of partition P {\textstyle P} , conditional on partition Q {\textstyle Q} , about x {\textstyle x} , is I P | Q ( x ) := − ln P ∨ Q ( x ) Q ( x ) {\displaystyle I_{P|Q}(x):=-\ln {\frac {P\vee Q(x)}{Q(x)}}} h T ( P ) {\textstyle h_{T}(P)} is the Kolmogorov-Sinai entropy h T ( P ) := lim n 1 n H ( P ( n ) ) = lim n E x ∼ μ [ 1 n I P ( n ) ( x ) ] {\displaystyle h_{T}(P):=\lim _{n}{\frac {1}{n}}H(P^{(n)})=\lim _{n}E_{x\sim \mu }\left[{\frac {1}{n}}I_{P^{(n)}}(x)\right]} In other words, by definition, there is a convergence in expectation. The SMB theorem states that when T {\textstyle T} is ergodic, there is convergence in L1.5
Theorem (ergodic case)—If T {\textstyle T} is ergodic, then x ↦ 1 n I P ( n ) ( x ) {\displaystyle x\mapsto {\frac {1}{n}}I_{P^{(n)}}(x)} converges in L1 to the constant function x ↦ h T ( P ) {\textstyle x\mapsto h_{T}(P)} .
In other words, E x ∼ μ [ | lim n 1 n I P ( n ) ( x ) − h T ( P ) | ] = 0 {\displaystyle E_{x\sim \mu }\left[\left|\lim _{n}{\frac {1}{n}}I_{P^{(n)}}(x)-h_{T}(P)\right|\right]=0}
In particular, since L1 convergence implies almost sure convergence, h T ( P ) = lim n 1 n I P ( n ) ( x ) {\displaystyle h_{T}(P)=\lim _{n}{\frac {1}{n}}I_{P^{(n)}}(x)} with probability 1.
Corollary (entropy equipartition property)— ∀ ϵ > 0 , ∃ N , ∀ n ≥ N {\textstyle \forall \epsilon >0,\exists N,\forall n\geq N} , we can partition the partition ∨ k = 0 n − 1 T − k P {\textstyle \vee _{k=0}^{n-1}T^{-k}P} into two parts, the “good” part G {\textstyle G} and the “bad” part B {\textstyle B} .
The bad part is small: ∑ b ∈ B μ ( b ) < ϵ {\displaystyle \sum _{b\in B}\mu (b)<\epsilon }
The good part is almost equipartitioned according to entropy: ∀ g ∈ G , − 1 n ln μ ( g ) ∈ h T ( P ) ± ϵ {\displaystyle \forall g\in G,\quad -{\frac {1}{n}}\ln \mu (g)\in h_{T}(P)\pm \epsilon }
If T {\textstyle T} is not necessarily ergodic, then the underlying probability space would be split up into multiple subsets, each invariant under T {\textstyle T} . In this case, we still have L1 convergence to some function, but that function is no longer a constant function.6
Theorem (general case)—Let I {\textstyle {\mathcal {I}}} be the sigma-algebra generated by all T {\textstyle T} -invariant measurable subsets of Ω {\textstyle \Omega } , - x ↦ 1 n I P ( n ) ( x ) {\displaystyle x\mapsto {\frac {1}{n}}I_{P^{(n)}}(x)} converges in L1 to x ↦ E [ lim n I P | ∨ k = 1 n T − k P | I ] {\displaystyle x\mapsto E\left[\lim _{n}I_{P|\vee _{k=1}^{n}T^{-k}P}{\big |}\;{\mathcal {I}}\right]}
When T {\textstyle T} is ergodic, I {\textstyle {\mathcal {I}}} is trivial, and so the function x ↦ E [ lim n I P | ∨ k = 1 n T − k P | I ] {\displaystyle x\mapsto E\left[\lim _{n}I_{P|\vee _{k=1}^{n}T^{-k}P}{\big |}\;{\mathcal {I}}\right]} simplifies into the constant function x ↦ E [ lim n I P | ∨ k = 1 n T − k P ] {\textstyle x\mapsto E\left[\lim _{n}I_{P|\vee _{k=1}^{n}T^{-k}P}\right]} , which by definition, equals lim n H ( P | ∨ k = 1 n T − k P ) {\textstyle \lim _{n}H(P|\vee _{k=1}^{n}T^{-k}P)} , which equals h T ( P ) {\textstyle h_{T}(P)} by a proposition.
Discrete-time functions can be interpolated to continuous-time functions. If such interpolation f is measurable, we may define the continuous-time stationary process accordingly as X ~ := f ∘ X {\displaystyle {\tilde {X}}:=f\circ X} . If the asymptotic equipartition property holds for the discrete-time process, as in the i.i.d. or finite-valued stationary ergodic cases shown above, it automatically holds for the continuous-time stationary process derived from it by some measurable interpolation. i.e. − 1 n log p ( X ~ 0 τ ) → H ( X ) {\displaystyle -{\frac {1}{n}}\log p({\tilde {X}}_{0}^{\tau })\to H(X)} where n corresponds to the degree of freedom in time τ. nH(X)/τ and H(X) are the entropy per unit time and per degree of freedom respectively, defined by Shannon.
An important class of such continuous-time stationary process is the bandlimited stationary ergodic process with the sample space being a subset of the continuous L 2 {\displaystyle {\mathcal {L}}_{2}} functions. The asymptotic equipartition property holds if the process is white, in which case the time samples are i.i.d., or there exists T > 1/2W, where W is the nominal bandwidth, such that the T-spaced time samples take values in a finite set, in which case we have the discrete-time finite-valued stationary ergodic process.
Any time-invariant operations also preserves the asymptotic equipartition property, stationarity and ergodicity and we may easily turn a stationary process to non-stationary without losing the asymptotic equipartition property by nulling out a finite number of time samples in the process.
A category theoretic definition for the equipartition property is given by Gromov.7 Given a sequence of Cartesian powers P N = P × ⋯ × P {\displaystyle P^{N}=P\times \cdots \times P} of a measure space P, this sequence admits an asymptotically equivalent sequence HN of homogeneous measure spaces (i.e. all sets have the same measure; all morphisms are invariant under the group of automorphisms, and thus factor as a morphism to the terminal object).
The above requires a definition of asymptotic equivalence. This is given in terms of a distance function, giving how much an injective correspondence differs from an isomorphism. An injective correspondence π : P → Q {\displaystyle \pi :P\to Q} is a partially defined map that is a bijection; that is, it is a bijection between a subset P ′ ⊂ P {\displaystyle P'\subset P} and Q ′ ⊂ Q {\displaystyle Q'\subset Q} . Then define | P − Q | π = | P ∖ P ′ | + | Q ∖ Q ′ | , {\displaystyle |P-Q|_{\pi }=|P\setminus P'|+|Q\setminus Q'|,} where |S| denotes the measure of a set S. In what follows, the measure of P and Q are taken to be 1, so that the measure spaces are probability spaces. This distance | P − Q | π {\displaystyle |P-Q|_{\pi }} is commonly known as the earth mover's distance or Wasserstein metric.
Similarly, define | log P : Q | π = sup p ∈ P ′ | log p − log π ( p ) | log min ( | set ( P ′ ) | , | set ( Q ′ ) | ) . {\displaystyle |\log P:Q|_{\pi }={\frac {\sup _{p\in P'}|\log p-\log \pi (p)|}{\log \min \left(|\operatorname {set} (P')|,|\operatorname {set} (Q')|\right)}}.} with | set ( P ) | {\displaystyle |\operatorname {set} (P)|} taken to be the counting measure on P. Thus, this definition requires that P be a finite measure space. Finally, let dist π ( P , Q ) = | P − Q | π + | log P : Q | π . {\displaystyle {\text{dist}}_{\pi }(P,Q)=|P-Q|_{\pi }+|\log P:Q|_{\pi }.}
A sequence of injective correspondences π N : P N → Q N {\displaystyle \pi _{N}:P_{N}\to Q_{N}} are then asymptotically equivalent when dist π N ( P N , Q N ) → 0 as N → ∞ . {\displaystyle {\text{dist}}_{\pi _{N}}(P_{N},Q_{N})\to 0\quad {\text{ as }}\quad N\to \infty .}
Given a homogenous space sequence HN that is asymptotically equivalent to PN, the entropy H(P) of P may be taken as H ( P ) = lim N → ∞ 1 N | set ( H N ) | . {\displaystyle H(P)=\lim _{N\to \infty }{\frac {1}{N}}|\operatorname {set} (H_{N})|.}
Cover & Thomas (1991), p. 51. - Cover, Thomas M.; Thomas, Joy A. (1991). Elements of Information Theory (first ed.). Hoboken, New Jersey: Wiley. ISBN 978-0-471-24195-9. ↩
Hawkins, Jane M. (2021). Ergodic dynamics: from basic theory to applications. Graduate texts in mathematics. Cham, Switzerland: Springer. p. 204. ISBN 978-3-030-59241-7. 978-3-030-59241-7 ↩
Algoet, Paul H.; Cover, Thomas M. (1988). "A Sandwich Proof of the Shannon–McMillan–Breiman Theorem" (PDF). The Annals of Probability. 16 (2): 899–909. doi:10.1214/aop/1176991794. JSTOR 2243846. Archived from the original (PDF) on 2016-12-06. /wiki/Thomas_M._Cover ↩
Petersen, Karl E. (1983). "6.2. The Shannon–McMillan–Breiman Theorem". Ergodic Theory. Cambridge Studies in Advanced Mathematics. Cambridge: Cambridge University Press. ISBN 978-0-521-38997-6. 978-0-521-38997-6 ↩
Pollicott, Mark; Yuri, Michiko (1998). "12.4. The Shannon–McMillan–Brieman theorem". Dynamical Systems and Ergodic Theory. London Mathematical Society Student Texts. Cambridge: Cambridge University Press. ISBN 978-0-521-57294-1. 978-0-521-57294-1 ↩
Misha Gromov (2013). "In a Search for a Structure, Part 1: On Entropy". (See page 5, where the equipartition property is called the 'Bernoulli approximation theorem'.) /wiki/Misha_Gromov ↩