Conditional relative information and its axiomatizations#
In this post, we will study the conditional form of relative information. We will also look at how conditional relative information can be axiomatized and extended to non-real-valued measures.
This post is a continuation from our series on spiking networks, path integrals and motivic information.
What is conditional relative information?#
Suppose we have two random variables
Let
where
We then define the conditional relative information to be
In the case where the corresponding densities are well-defined, we have
which is the relative information to
Whats is the chain rule for conditional relative information?#
In statistics and machine learning, we often think of
To uncover the truth, it makes strategic sense to study different facets
Therefore, to get a good model of
How do we derive conditional entropy from conditional relative information?#
Just as the entropy of a random variable
Given random variables
the conditional relative information of
According to the chain rule of conditional relative information,
By the definition of entropy in our introduction, the first and third terms are the entropies
Is there an axiomatization of conditional entropy?#
As described in our previous post, we allow the total measures of
We start with an axiomatizations of conditional entropy with the hope of deriving axiomatizations of conditional relative information. I like the following categorical view of conditional entropy [BFL11]. I’ve taken the liberty of rewriting it in our notations.
Given a measured space
In this case, the conditional entropy of
where
Given two measured spaces
Let
Functoriality.
Homogeneity.
Additivity.
Continuity.
is continuous
Then,
Given a classical conditional entropy
The nice thing about the above categorical axiomatization of conditional entropy is that it fits into the view where the objects of study are spaces
Is there an axiomatization of conditional relative information?#
We prefer to work with conditional relative information rather than conditional entropy. Its axiomatization should tell us how it behaves with respect to products and coproducts of the measures being compared.
Our axioms. Note the addition of the product rule. I’m not sure if the product axiom can be derived from the others when the state spaces are not finite. Perhaps it will follow from continuity and the fact that the limit of coproducts is the product of limits.
Functoriality.
Homogeneity.
Coproduct.
Product.
Continuity.
is continuous.
Here,
The axioms for conditional entropy follow immediately from these axioms for conditional relative information, because we can write
where