Natural Latents: A Mathematical Foothold on Translatability and Intersubjectivity

17 min read

In John Wentworth and David Lorell’s paper “Natural Latents”, they aim to propose mathematical conditions under which translation between agents is guaranteed to be possible, and claim that these are the most general conditions under which translatability is guaranteed.

You can see their paper on arXiv or on the AI Alignment Forum, and read precursor posts here and here.

This paper has become a central part of my undergraduate thesis work, and as part of that work I present my own rendition of the math in their paper, with the aim of being maximally legible to readers with diverse mathematical backgrounds. Conceptual discussions are mostly outside of the scope of this post for now—we will focus on the definitions, theorems, and proofs.

Contents

  1. Mediators, Redunds, and Natural Latents
  2. Theorem: Mediator Determines Redund
  3. Probabilistic Generative Models: Observables and Latents
  4. Theorem: Guaranteed Translatability
  5. Appendix: Elementary Information Theory

If you are not intimately familiar with entropy, mutual information, and total correlation, I recommend reading the Appendix before proceeding. The main text assumes familiarity with these concepts.

Mediators, Redunds, and Natural Latents

Definition: Mediator

We say that a variable mediates over a sequence of discrete random variables when are approximately conditionally independent given .

Intuitively, this means that once we know , learning the value of any doesn’t significantly change our uncertainty about the other s. All the shared information between the variables flows through .

Formally, we require that the conditional total correlation is less than some constant :

Z is a mediator over X

is an exact mediator if ; an approximate mediator up to error otherwise. We normally assume mediators are approximate, so we just say “mediator.” We think of as small, but our bounds are global, so there are no actual assumptions needed.

Definition: Redund

When a random variable is a function of another random variable , we know the value takes as soon as we know . Therefore the new information that we get is exactly zero:

When this only holds approximately, the conditional entropy is small, and we say that is approximately a function of or approximately determined by :

If is approximately a function of each element of a sequence , then we say it is a redund over . This means that all the information contained in is also contained in each .

Z is a redund over X

is an exact redund if ; an approximate redund up to error otherwise. We normally assume redunds are approximate, so we just say “redund.” Again, we think of as small, but our bounds are global.

Definition: Natural Latent

is a natural latent over up to errors and iff it satisfies both the mediation and redundancy conditions up to those errors:

Λ is a natural latent over X

Theorem: Mediator Determines Redund

Let be a sequence of discrete random variables , and assume that for some random variables and and for some constants and the following inequalities hold:

Then:

Proof

Pick some distinct where . By the definition of conditional mutual information,

is what we want to bound, so we put it on one side:

We bound the mutual information term using monotonicity:

Now we can use the chain rule for mutual information to break apart the right hand side:

Substituting, we get an upper bound on :

Now we will bound each of these terms individually.

First, by monotonicity we have

And by the nonnegativity of mutual information we have

Which by assumption is bounded by .

Second, bounding by the entropy of its second argument gives

By monotonicity,

Which by assumption is bounded by .

Third, simply by monotonicity and the redundancy assumption,

So we have bounded each term, and substituting gives

Probabilistic Generative Models: Observables and Latents

In this section, we formalize how agents represent the world using probabilistic generative models. This framework will allow us to precisely state and prove the guaranteed translatability theorem.

If you don’t need the measure-theoretic details of how observables and latents are formally constructed from sensory data, you can skim or skip this section and jump directly to the Guaranteed Translatability theorem. The key intuition is that agents break up raw sensory data into chunks called observables and posit hidden variables (latents) to model these features.

We assume that a Bayesian agent learns a probabilistic generative model (PGM) which it uses to make predictions about the world. Traditionally, a PGM represents an agent as having a joint distribution over a set of hidden latent variables and observables . Here we will begin with a single random variable, and show how first multiple observable random variables and then latent variables are introduced. Throughout this section, we assume all random variables are discrete.

Note that we will reuse some symbols from the previous section, and while we try to choose them in an intuive way, they have new meanings here.

Sensory Data

We define the random variable as follows. Let be a probability space. Then let

be a measurable function, where:

  • is a sample space of outcomes.
  • is the -algebra on .
  • is the probability measure.
  • is a countable set representing all possible sensory data.
  • is the power set of , denoted .
  • The distribution of is a Probability Mass Function (PMF) for .

The agent can have a predictive model encoded as a distribution .

Observables

Given sensory data , we can define a sequence of observable random variables . We denote this sequence as and say that the random variables are measurable functions of :

where each is a countable set. The -th observable random variable is then .

The joint observable vector is where .

Each has a well-defined PMF derived from :

Latents

Now we assume that in order to find and the joint distribution , the agent uses a latent variable model.

The agent posits discrete latent random variables , where each is a countable set.

We define the agent’s model using the joint PMF . Since is a deterministic function of , the joint distribution of all variables is a degenerate extension of the distribution over and .

For any specific outcomes , and , the joint probabilities satisfy:

Including in the joint distribution is technically redundant because adds no new information once is known (though may contain information not captured by ). However, we include to keep the relationship between raw data, features, and latents explicit.

Motivation of Latent Variables

Modeling directly is usually difficult because sensory data is high-dimensional and complex. However, it is often easier to use a generative process where we sample a simple latent factor from a prior , and we generate data conditioned on that factor using .

We have:

  • Observable space: The countable set (and by extension ).
  • Latent space: The countable set .
  • Random variables: Discrete variables and .

From these, we define the joint probability mass function (PMF):

The marginal distributions are obtained by summing over the other variable (Law of Total Probability):

We can factor the joint distribution into a prior and a likelihood (conditional probability):

This factorization allows us to express the complex data distribution as a mixture of (hopefully) simpler distributions:

Theorem: Guaranteed Translatability

This theorem establishes when translation between two agents’ latent variables is guaranteed to be possible. Two agents observe the world and build their own models with their own latent variables. If they agree on certain shared observables and their latent variables approximately satisfy the natural latent conditions (mediation and redundancy) then there exists an approximate mapping between the latent variables in their models.

The forward direction of this theorem shows that one agent’s mediators can determine another’s redunds, and the reverse direction shows that in the special case of two observables, being determined by all mediators implies being a redund.

Agreement on Shared Observables

We now consider two Bayesian agents, () and (), who learn probabilistic generative models and , respectively. derives observables . derives observables . These need not be the same family — their feature maps and may differ.

Some of ′s and ′s observables may be essentially the same, in the sense that they are functions of each other. We call these shared observables and denote them by .

Specifically, we define subsequences and of length , such that for each , there is a bijection:

These variables correspond to the same feature of the data (e.g., ).

Important: The set need not contain all observables that are shared between the agents. We only require that each element of is indeed shared—that is, for each variable we include in and , there exists the bijection described above. The agents may have additional shared observables beyond those in , but our analysis focuses on a particular subset of shared observables.

While the variables may be linked in reality, the agents’ models of these variables may differ. We say that and agree on the observables in if their marginal distributions over are identical (under the mapping ).

Formally, for all and for all outcomes :

Since both agents assign the same probability mass to corresponding outcomes, the entropy of the shared observables is identical for both agents:

Consequently, and are interchangeable in entropy expressions, and we will often omit the superscripts (writing simply ) when the distinction is not needed.

Statement

Assume and agree on observables as defined above. Let be any mediator in ′s model:

Then, for a redund over in ′s model, is approximately determined by . Formally, if is a redund up to error :

Then:

Furthermore, when there are exactly two shared observables and , the implication works in reverse as well. If is determined by any mediator in ′s model up to error , then must be a redund up to error .

Formally, if for all such that , we have for some constant , then:

Proof

We are given that and agree on observables and that ′s latent mediates over up to error :

Forward Direction

Assume ′s latent is a redund over her shared observables :

Since and agree on observables, the redundancy bound holds for ′s observables as well:

Now we apply the Mediator Determines Redund Theorem to the system .

  • is the mediator over (from B).
  • is the redund over (from D).

Substituting the bounds directly into the theorem gives:

Reverse Direction

We restrict ourselves to the case . We assume that for any mediator in ′s model (where ), is determined by that mediator up to error :

We must show that is a redund.

In the 2-variable case, the observables mediate themselves. If chooses , the conditional total correlation is zero:

Since is a valid mediator (with ), our assumption requires that it determines :

By symmetry, is also a valid mediator, so:

Using the agreement on observables assumption to map back to ′s variables, we have:

So is a redund over up to error .

Corollary: Bi-directional Translation with Natural Latents

The theorem above establishes one-way translatability: if Bob has a mediator and Alice has a redund, then Bob’s latent determines Alice’s. However, when both agents use approximately natural latents, we get an approximate bijection between their latents.

Statement: Suppose Alice and Bob agree on shared observables , and both and are natural latents over their respective shared observables. That is:

is both a mediator over (with ) and a redund over (with for all )

is both a mediator over (with ) and a redund over (with for all )

Then both conditional entropies are small:

Proof: Apply the forward direction of the Guaranteed Translatability theorem twice.

Alice → Bob: Since mediates and is a redund over (hence over by agreement), we have

Bob → Alice: Since mediates and is a redund over (hence over by agreement), we have

Appendix: Elementary Information Theory

We will giving a few elementary definitions and identities from information theory. We will always deal with discrete random variables, as this avoids continuous pathologies and is arguably the correct level of description for the real world anyway. This will assume very minimal prior knowledge, but for an authoritative source refer to Elements of Information Theory 2nd Edition by Cover and Thomas.

So is a discrete random variable which takes values , and we write A joint distribution takes values and

All of what follows can be expressed in terms of discrete random variables with probability mass functions.

Entropy

The entropy of a random variable is the average information we get from sampling it.

Since probabilities are between zero and one, we have , so we multiple by to get a nonnegative number. Each term is the probability of outcome , while is the information content (or “surprise”) of that outcome. Rare events have high information content, and common events have low information content. Entropy is the probability-weighted average of these information contents.

The joint entropy measures uncertainty about multiple variables. The conditional entropy is the average information we get from sampling when we already know the value of .

Mutual Information

The mutual information between two random variables measures how much information they share, or equivalently, how much knowing one tells you about the other.

We give two equivalent forms:

The first form shows mutual information as “total information minus joint information”—the extent to which and are redundant. The second form shows it as “reduction in uncertainty”—how much knowing reduces our uncertainty about .

Conditional mutual information measures the information and share, given that we already know :

Again, the first form captures redundancy given , while the second form captures how much reduces our uncertainty about beyond what already tells us.

Chain rule for mutual information decomposes mutual information about multiple variables into successive contributions:

This says: the information shares with both and (given ) equals the information shares with (given ), plus the additional information shares with beyond what already provided (given both and ). We use this identity in our proof of the Mediator Determines Redund theorem.

Mutual information is bounded by entropy of either argument. Since entropy is nonnegative, , so from the second form we have:

This states that knowing cannot tell you more about than the total uncertainty you had about in the first place (given ). This bound appears in our proof of Mediator Determines Redund.

Total Correlation

Total correlation measures the total dependence among a set of variables—how much we learn by observing them together versus independently.

We give two equivalent forms for both the unconditional and conditional cases.

Unconditional total correlation:

where .

The first form is a direct generalization of the two-variable mutual information: “sum of individual entropies minus joint entropy,” measuring how much information is double-counted when we sum the individual entropies. The second form expresses this as a sum of successive dependencies: how much each variable shares with all preceding variables.

Conditional total correlation:

The conditional case measures the same redundancy, but with everything measured relative to already knowing .