Upper and Lower Mutual Information Bounds
Recently Foster et al (2020) introduced a couple of tractable bounds that can be used to estimate the expected information gain (EIG) of a design policy (a mapping from past designs and observations to the next design). These include the sequential Prior Contrastive Estimate (sPCE) lower bound:
where
We also have the corresponding sequential Nested Monte Carlo (sNMC) upper bound:
where the only difference is that we don’t include
This might lead you to wonder if we could use the upper bound instead, which does not suffer the same limitation. Intuitively it doesn’t make much sense to maximise a quantity by maximising an upper bound, because we might just be making the bound arbitrarily loose. But it turns out there is a monotonicity property connecting the sPCE and sNMC.
Let
Theorem 1 the sPCE lower bound and sNMC upper bound induce the same total order on policy space:
Proof to simplify the following analysis, we introduce the notation:
plugging this back into the difference between lower bounds gives us:
Now we subtract
Looking at the difference between upper bounds, we have that
Now, we note that if
Similarly, we can note that if
Q.E.D.
So what does this all mean? In theory, we can maximise the sNMC upper bound, and this indirectly maximise the sPCE lower bound, which we hope will drive up the actual mutual information. In fact, the two optimisations are in some sense equivalent, as increasing either bound must increase the other by a predictable, deterministic amount. In practice, my experience is that both approaches perform equally well.
References:
Foster, A., Ivanova, D. R., Malik, I., & Rainforth, T. “Deep adaptive design: Amortizing sequential bayesian experimental design.” International Conference on Machine Learning. PMLR, 2021.