Compositional Imprecise Probability
Liell-Cock, Staton ยท POPL 2025 ยท
arXiv:2405.09391
Prereqs: ๐ Fritz 2020 (Markov categories). 5 min.
When you don't know the exact probability โ only that it lies in some set of distributions โ you have imprecise probability. Liell-Cock shows these form a Markov category (nLab) via the Para construction, and that graded monads organize the imprecision levels.
Imprecise probability โ sets of distributions
A precise Markov kernel maps an input to one distribution. An imprecise kernel maps to a set of distributions โ a credal set. You know the output is distributed according to one of these distributions, but you don't know which one. This is stronger than a single distribution (you commit less) and weaker than full ignorance (you commit something).
The Para construction
The Para construction takes a Markov category C and builds a new category Para(C) where morphisms have a hidden parameter. A morphism in Para(C) from X to Y is a morphism P ร X โ Y in C, where P is the parameter space. Imprecise kernels arise when you existentially quantify over the parameter: "there exists some p in P such that..."
Confidence: Simplified. Real Para construction works over arbitrary Markov categories with tensor products for parameters. Same idea: hidden parameter generates a family.
Lower and upper expectations
With a credal set, you compute bounds on expectations. The lower expectation is the minimum expected value over all distributions in the set. The upper expectation is the maximum. These brackets tell you the range of possible answers โ how uncertain you are.
Graded monads โ imprecision levels
Different amounts of imprecision form a graded monad. The grade tracks how much imprecision you have. Grade 0 = precise (one distribution). Higher grades = wider credal sets. Composition accumulates imprecision โ the grade adds up, just like effect grades in ๐ Gaboardi 2021.
The bridge โ graded monads = Markov categories
Liell-Cock's main result: the Kleisli category of a graded monad on a Markov category is itself a Markov category. Imprecise probability inherits all the structure โ copy, discard, composition โ from the precise case. You can do everything Fritz does in FinStoch, but with credal sets instead of single distributions.
Confidence: Simplified. Real composition requires careful treatment of the parameter space tensor. Same compositional structure.
Notation reference
| Paper | Scheme | Meaning |
|---|---|---|
| Para(C) | (make-para params kernel) | Para construction |
| credal set | (make-credal d1 d2 ...) | Set of distributions |
| Eฬฒ[f] | (lower-exp credal f) | Lower expectation |
| Eฬ [f] | (upper-exp credal f) | Upper expectation |
| T_e | # credal set at grade e | Graded monad at imprecision level e |
Neighbors
Other paper pages
- ๐ Fritz 2020 โ the Markov category that Para generalizes
- ๐ Gaboardi 2021 โ graded monads for effects (same grading idea)
- ๐ Fritz, Perrone 2021 โ support is the maximally imprecise shadow
- ๐ Capucci 2021 โ Para construction appears in cybernetics too
Foundations (Wikipedia)
Translation notes
All examples use finite lists of finite distributions as credal sets. The paper works with convex sets of probability measures over measurable spaces, the Para construction on arbitrary Markov categories, and graded monads indexed by a quantale of imprecision levels. For example: the lower/upper expectation example on this page iterates over three coin distributions. In the paper, the same construction applies to credal sets of Gaussian measures over โโฟ โ where the lower expectation is an optimization problem over an infinite-dimensional convex set. The bracket structure (min/max over the set) is identical. The measure theory and optimization are not.
Read the paper. Start at ยง3 for the Para construction, ยง5 for graded monads and Markov categories.
Framework connection: Imprecise probability via the Para construction generalizes the Natural Framework's ambient category from precise to credal-set-valued stages. (
Ambient Category)