With braiding
With duals for objects
category with duals (list of them)
dualizable object (what they have)
ribbon category, a.k.a. tortile category
With duals for morphisms
With traces
Closed structure
Special sorts of products
Semisimplicity
Morphisms
Internal monoids
Examples
Theorems
In higher category theory
One of the main purposes of probability theory, and of related fields such as statistics and information theory, is to make predictions in situations of uncertainty.
Suppose that we are interested a quantity , whose value we don’t know exactly (for example, a random variable), and which we cannot observe directly. Suppose that we have another quantity, , which we also don’t know exactly, but which we can observe (for example, through an experiment). We might now wonder: can observing give us information about , and reduce its uncertainty? Viewing the unknown quantities and as having hidden knowledge or hidden information, one might ask, how much of this hidden information is shared between and , so that observing uncovers information about as well?
This form of dependence between and is called stochastic dependence, and is one of the most important concepts both in probability theory, and, due to its conceptual nature, in most categorical approaches to probability theory.
Very roughly, two unknown quantities are stochastically independent when learning the value of one of them does not change what one should expect about the other.
For instance, suppose that and are random variables. If observing changes the distribution one assigns to , then and are stochastically dependent. If observing does not change the distribution of , then and are stochastically independent.
For a concrete example, let be the result of throwing one die, and let be the result of throwing another die. If the dice are thrown separately and do not influence each other, then learning the value of does not change the probabilities assigned to . If we learn that , the probability that is still . Thus and are independent.
By contrast, let be the result of throwing a die, and let be the event that the result is even. Then and are dependent: learning that has occurred changes the possible values of from
to
Thus observing gives information about .
A more geometric example is given by choosing a point uniformly at random in the unit disk, and letting and be its two coordinates. Each coordinate separately has a symmetric distribution, but the two coordinates are not independent: if is close to , then must be close to . The dependence is not caused by either coordinate determining the other exactly, but by the joint constraint
Often one is interested not in whether two quantities are independent absolutely, but whether they are independent relative to some background information.
A standard example is the following. Let say whether it is raining, let say whether the grass in Alice’s garden is wet, and let say whether the grass in Bob’s garden is wet. Marginally, and are dependent: if Alice’s grass is wet, this is evidence that it is raining, and hence evidence that Bob’s grass is wet. But after conditioning on whether it is raining, the remaining dependence may disappear. In that case and are conditionally independent given .
Thus conditional independence does not mean that and are independent in the usual sense. Rather, it means that their dependence is completely mediated by the conditioning variable .
Conversely, conditioning can also create dependence. For example, suppose that two independent causes and may each lead to a common effect . If one conditions on the effect having occurred, then learning that occurred can make less likely, since the effect has already been explained by . This phenomenon is sometimes called explaining away?. Thus conditional independence and ordinary independence are logically distinct notions.
Let be a probability space, and let be events, i.e. measurable subsets of . We say that and are independent if and only if
i.e. if the joint probability is the product of the probabilities.
More generally, if and are random variables or random elements, one says that and are independent if and only if all the events they induce are independent, i.e. for every and ,
Equivalently, one can form the joint random variable and form the joint distribution on . We have that and are independent as random variables if and only if for every and ,
If we denote the marginal distributions of by and , the independence condition reads
meaning that the joint distribution is the product of its marginals:
In this form, stochastic dependence is the obstruction to reconstructing the joint law from its two marginals. The marginal laws and describe the two random quantities separately, while the joint law contains in addition their possible dependence structure.
For a finite family of random variables
mutual independence means that the joint distribution factors as the product of its marginals:
Equivalently, for all measurable subsets ,
The infinitary equivalent can be obtained by means of the Kolmogorov extension theorem.
Let be a sub--algebra, thought of as encoding some background information. Events are conditionally independent given if
More generally, random elements and are conditionally independent given if, for all and ,
almost surely.
If is another random element, one says that and are conditionally independent given if they are conditionally independent given the sub--algebra generated by :
This is often written
When regular conditional probabilities? exist, conditional independence can be expressed by saying that the conditional joint law of over factors as the product of the corresponding conditional marginal laws:
for -almost every .
Equivalently, the joint law of admits the factorization
Denote by the Giry monad on Meas, sending a measurable space to the measurable space of probability measures on .
The coordinate projections
induce maps
and hence the marginals
The Giry monad is compatible with products by the usual product-measure construction: there is a natural map
In these terms, a joint distribution is independent precisely if it is obtained from its marginals by this product-measure map: .
Equivalently, if random elements , are defined on a probability space , then and are independent precisely when
Thus, from the point of view of the Giry monad, stochastic dependence is exactly the failure of a state of the product object to be the product of its marginal states.
More generally, for a finite family of measurable spaces , there is a product-measure map
A joint distribution
is independent precisely if it is obtained from its marginals by this map:
For an infinite family , the corresponding statement is subtler. One would like to say that an independent joint distribution on
is determined by its finite-dimensional marginals
and that these finite-dimensional marginals factor as
The existence of a measure on the infinite product with these prescribed finite-dimensional marginals is the content of the Kolmogorov extension theorem, under suitable hypotheses on the measurable spaces. Thus, for infinite families, independence is naturally formulated in terms of compatible finite-dimensional product distributions.
In particular, a family of random elements
is independent if every finite subfamily is independent, equivalently if for every finite subset the joint law of factors as the product of its marginals:
(See the analogous section in Joint and marginal probability.)
(For now see Markov category - Stochastic independence)
(…)
Kenta Cho, Bart Jacobs, Disintegration and Bayesian Inversion via String Diagrams, Mathematical Structures of Computer Science 29, 2019. (arXiv:1709.00322)
Tobias Fritz, A synthetic approach to Markov kernels, conditional independence and theorems on sufficient statistics, Advances of Mathematics 370, 2020. (arXiv:1908.07021)
Alex Simpson, Equivalence and Conditional Independence in Atomic Sheaf Logic (arXiv:2405.11073)
Last revised on May 11, 2026 at 10:40:54. See the history of this page for a list of all contributions to it.