โ† back to index

Compositional Imprecise Probability

Liell-Cock, Staton ยท POPL 2025 ยท arxiv arXiv:2405.09391

Prereqs: ๐Ÿž Fritz 2020 (Markov categories). 5 min.

When you don't know the exact probability โ€” only that it lies in some set of distributions โ€” you have imprecise probability. Liell-Cock shows these form a Markov category (nLab) via the Para construction, and that graded monads organize the imprecision levels.

Imprecise probability โ€” sets of distributions

A precise Markov kernel maps an input to one distribution. An imprecise kernel maps to a set of distributions โ€” a credal set. You know the output is distributed according to one of these distributions, but you don't know which one. This is stronger than a single distribution (you commit less) and weaker than full ignorance (you commit something).

Scheme

The Para construction

The Para construction takes a Markov category C and builds a new category Para(C) where morphisms have a hidden parameter. A morphism in Para(C) from X to Y is a morphism P ร— X โ†’ Y in C, where P is the parameter space. Imprecise kernels arise when you existentially quantify over the parameter: "there exists some p in P such that..."

Scheme

Confidence: Simplified. Real Para construction works over arbitrary Markov categories with tensor products for parameters. Same idea: hidden parameter generates a family.

Lower and upper expectations

With a credal set, you compute bounds on expectations. The lower expectation is the minimum expected value over all distributions in the set. The upper expectation is the maximum. These brackets tell you the range of possible answers โ€” how uncertain you are.

Scheme

Graded monads โ€” imprecision levels

Different amounts of imprecision form a graded monad. The grade tracks how much imprecision you have. Grade 0 = precise (one distribution). Higher grades = wider credal sets. Composition accumulates imprecision โ€” the grade adds up, just like effect grades in ๐Ÿž Gaboardi 2021.

Scheme

The bridge โ€” graded monads = Markov categories

Liell-Cock's main result: the Kleisli category of a graded monad on a Markov category is itself a Markov category. Imprecise probability inherits all the structure โ€” copy, discard, composition โ€” from the precise case. You can do everything Fritz does in FinStoch, but with credal sets instead of single distributions.

Graded Monad T_r : C → C grade r accumulates T_r ; T_s = T_(r+s) Markov Cat Kl(T_r) copy • discard compose • condition Kleisli graded imprecision inherits full Markov structure
Scheme

Confidence: Simplified. Real composition requires careful treatment of the parameter space tensor. Same compositional structure.

Notation reference

Paper Scheme Meaning
Para(C)(make-para params kernel)Para construction
credal set(make-credal d1 d2 ...)Set of distributions
Eฬฒ[f](lower-exp credal f)Lower expectation
Eฬ…[f](upper-exp credal f)Upper expectation
T_e# credal set at grade eGraded monad at imprecision level e
Neighbors

Other paper pages

Foundations (Wikipedia)

Translation notes

All examples use finite lists of finite distributions as credal sets. The paper works with convex sets of probability measures over measurable spaces, the Para construction on arbitrary Markov categories, and graded monads indexed by a quantale of imprecision levels. For example: the lower/upper expectation example on this page iterates over three coin distributions. In the paper, the same construction applies to credal sets of Gaussian measures over โ„โฟ โ€” where the lower expectation is an optimization problem over an infinite-dimensional convex set. The bracket structure (min/max over the set) is identical. The measure theory and optimization are not.

Ready for the real thing? arxiv Read the paper. Start at ยง3 for the Para construction, ยง5 for graded monads and Markov categories.

Framework connection: Imprecise probability via the Para construction generalizes the Natural Framework's ambient category from precise to credal-set-valued stages. (jkAmbient Category)