Entropy #
This file imports the entropy definitions and basic API from PFR.ForMathlib.Entropy.Basic.
The main definitions live in the ProbabilityTheory namespace:
It also provides notations such as H[X], H[X ; μ], H[X | Y], and I[X : Y].
Entropy is bounded by the logarithm of any natural-number upper bound on the alphabet size.
Entropy is bounded by c * log 2 when the alphabet has size at most 2 ^ c.
Mutual information is bounded by the entropy of the left variable.
Mutual information is bounded by the entropy of the right variable.
Mutual information is unchanged by an injective recoding of the right variable.
If the left variable is almost surely constant, its mutual information with any finite variable is zero.
If the right variable is almost surely constant, its mutual information with any finite variable is zero.
If the left variable is almost surely constant, its conditional mutual information with any finite variable is zero.
If the right variable is almost surely constant, its conditional mutual information with any finite variable is zero.
For finite alphabets, independence of two random variables follows from factorization on singleton fibers.
Conditional mutual information is bounded by the left conditional entropy.
Conditional mutual information is bounded by the right conditional entropy.
Conditional mutual information is bounded by the entropy of the left variable.
Conditional mutual information is bounded by the entropy of the right variable.
Conditional data processing where the left-side postprocessing may depend on the conditioning value.
Conditional data processing where the right-side postprocessing may depend on the conditioning value.
Adding a deterministic function of the conditioning variable to the conditioning variable does not change conditional mutual information.
Conditioning additionally on a deterministic function of the left variable and the existing conditioning data cannot increase conditional mutual information.
Conditioning additionally on a deterministic function of the right variable and the existing conditioning data cannot increase conditional mutual information.
If a coarse conditioning variable is a deterministic function of a finer one, and the finer
conditioning variable carries no conditional information about X beyond the coarse variable,
then the finer conditioning can only increase I[X : Y | -].
Chain rule for conditional mutual information, splitting a pair on the left.
Chain rule for conditional mutual information, splitting a pair on the right.
Adding an extra right-side variable cannot decrease conditional mutual information.
Chain rule for conditional mutual information against a finite boolean vector, exposing coordinates from left to right.
Conditioning on a larger event and then on a smaller event is the same as conditioning directly on the smaller event.
Real masses obey the same reweighting identity for nested conditioning events.
If the conditioning variable is a pair (K, Z), conditional mutual information is the
average over the fibers of K of the conditional mutual information given Z.
If an event is determined by the conditioning variable, then the contribution of the conditional mutual information on that event is bounded by the original conditional mutual information. This is the information-theoretic reweighting used in Claim 6.21.
Chain rule for mutual information, splitting a pair on the right.
Chain-rule rearrangement: if conditioning on Y does not increase the dependence between
X and W, then conditioning on W cannot increase the information that Y has about X
beyond the unconditioned I[X : Y].
Conditional entropy is unchanged when both variables are replaced by almost-everywhere equal variables.
Conditional mutual information is unchanged when the two measured variables are replaced by almost-everywhere equal variables and the conditioning variable is unchanged.
Conditional mutual information is unchanged when all three variables are replaced by almost-everywhere equal variables.
Conditional mutual information is determined by the joint law of (X, Y, Z).
Conditional mutual information is unchanged by injective recodings of the right variable and the conditioning variable.
If (X, Y) and (X', Y') have the same joint law, then conditioning on the same measurable
event in the Y/Y' coordinate preserves the law of X/X'. This is the heterogeneous-codomain
version of ProbabilityTheory.IdentDistrib.cond.