Uncertainty Modelling Robust Optimisation

First introduced 8 May 2026

Ambiguity Set

"What we observe is not nature itself, but nature exposed to our method of questioning." — Werner Heisenberg, Physics and Philosophy (1958)

Instead of betting on one assumed probability distribution, collect all distributions consistent with what the data shows and optimise for the worst one in that set.

Why it's needed

Every optimisation model that involves uncertainty makes a choice about how to represent that uncertainty. The most common choice is to assume a specific probability distribution for the uncertain quantity — a normal distribution for demand, a lognormal for asset returns, a Poisson for arrivals. This choice feels reasonable because it can be estimated from data. But it creates a hidden fragility: the model is optimised for the assumed distribution, and if the true distribution turns out to be different, the solution can perform badly on real data.

There are two obvious ways to respond, and both fail in characteristic ways:

✗ ASSUME ONE DISTRIBUTION

Fit a normal or Poisson to your historical data. Optimise. When the true distribution has fatter tails or a different shape, your "optimal" plan runs out of stock or wastes capacity in the worst periods.

OR
✗ ASSUME WORST-CASE PARAMETERS

Classical robust optimisation hedges against the worst-case parameter value (e.g., demand is always at its maximum). The resulting plan is often far too conservative — it over-stocks everywhere to guard against a scenario that never occurs.

Both approaches fail because they force a commitment — either to a specific distribution that may be wrong, or to a worst-case scenario that is too extreme. The hard problem is: how do you make a good decision when you know the distribution is uncertain, but you also know it is not as bad as the absolute worst case?

What it does

An ambiguity set, written as 𝔻, is a set of probability distributions — not a set of outcomes or parameter values, but a set of entire probability laws. The defining property of every distribution in 𝔻 is that it is consistent with what the data shows. Typically, "consistent" means matching observed moments (the sample mean, the sample variance, or both), though other consistency criteria are possible.

Once 𝔻 is constructed, the optimisation problem becomes:

Find the decision x that minimises the worst-case expected cost over all distributions in 𝔻.

Formally: minx maxP ∈ 𝔻 EP[cost(x, ξ)], where ξ is the uncertain quantity and P is a distribution over ξ.

The key pieces and their roles:

Why the fine between min and max matters. The outer min is over decisions (your choice). The inner max is over distributions (the adversary's choice). The adversary goes second — after seeing your decision x, it picks the worst distribution. This is why the solution is robust: it holds up under the worst-case distribution that could arise given what you decided.
AMBIGUITY SET 𝔻 all distributions consistent with observed moments thin-tailed medium worst case (adversary picks this) optimiser hedges here
All three distributions fit the observed moments and are plausible members of 𝔻. The adversary selects the worst-case (fat-tailed) distribution; the optimiser finds the decision that minimises cost even under that distribution.

Core idea

In the notation introduced above:

The optimisation problem minx maxP ∈ 𝔻 EP[cost(x, ξ)] is a min-max programme. In the vocabulary above: the decision-maker minimises expected cost; the adversary maximises it by choosing the worst distribution in 𝔻; the result is the decision that performs best even when the distribution is as bad as 𝔻 permits.

How 𝔻 is typically constructed. The most common approach is moment-matching: include every distribution whose mean lies within a confidence interval of the sample mean, and whose variance lies within a confidence interval of the sample variance. This gives a set that is tight enough to be useful (not all distributions) but loose enough to hedge against distributional shift (not just the fitted distribution).

The key mathematical result that makes DRO tractable is that, for common ambiguity set definitions (moment-based, Wasserstein-ball-based, likelihood-region-based), the inner maximisation over distributions can be solved in closed form and the resulting expression substituted back into the outer minimisation. This reformulation typically yields a second-order cone programme (SOCP — a convex optimisation problem with constraints defined by second-order cones, which generalise linear constraints to include quadratic terms in a tractable form) or a semidefinite programme (a convex optimisation where the feasible set is defined by requiring a matrix to be positive semidefinite, meaning all its eigenvalues are non-negative), both of which standard solvers can handle.

Concrete example — pasta sauce manufacturing

Setup: A pasta sauce factory must decide how many jars to produce each week. Weekly demand fluctuates. The analyst has 18 months of sales data, enough to compute a sample mean of 2,400 jars and a sample variance, but not enough to confidently rule out a right-skewed distribution, a bimodal one, or a long-tailed one.

Without the ambiguity set: The analyst fits a normal distribution (mean 2,400, variance as estimated) and optimises the production plan against it. The plan is efficient — it minimises expected overstock and shortage costs. But if the true distribution has a fat right tail (occasionally very high demand weeks), the plan will run short of stock in those weeks more often than expected, because the normal model underweighted that tail.

With the ambiguity set: The analyst constructs 𝔻 as the set of all distributions with mean in [2,300, 2,500] and variance in [σ² − ε, σ² + ε] for some tolerance ε based on the sample size. This set includes the fitted normal, but also the right-skewed alternatives and the fat-tailed variants. The plan is then optimised against the worst-case distribution in this set. The resulting production quantity is higher than the normal-distribution optimum — it overproduces slightly to guard against the fat-tailed scenarios — but it performs reliably across all plausible distributions.

The symbols, for the record: x = production quantity (the decision); ξ = actual weekly demand (the uncertain quantity); cost(x, ξ) = overstock cost if x > ξ, shortage cost if x < ξ; 𝔻 = all distributions consistent with the observed moment bounds. The solution solves minx maxP ∈ 𝔻 EP[cost(x, ξ)].

Common misreads

Misread 1: "An ambiguity set is the same as a scenario set."

A scenario set in stochastic programming is a finite list of specific outcomes with assigned probabilities — for example, "demand is 2,000 with probability 30%, 2,400 with probability 50%, 2,800 with probability 20%." An ambiguity set is a collection of entire probability distributions, not individual scenarios. The two operate at different levels: a scenario set is a discrete approximation of one distribution; an ambiguity set is a region containing many distributions. Stochastic programming optimises under one assumed distribution (perhaps approximated by scenarios); DRO optimises under the worst distribution within 𝔻. They are not interchangeable concepts.

Misread 2: "More data always shrinks the ambiguity set."

This holds when the ambiguity set is defined by moment-matching constraints estimated from data, since more data tightens moment estimates. But ambiguity sets can also be defined by Wasserstein balls (regions around the empirical distribution in a transport-cost metric — a way of measuring how different two distributions are by the cost of moving probability mass from one to the other), likelihood regions, or shape constraints, some of which do not shrink monotonically with sample size. The relationship between data volume and set size depends entirely on how the set is constructed.

Misread 3: "Distributionally robust optimisation is just worst-case optimisation."

Classical robust optimisation (minimax over parameter values) and DRO (minimax over distributions) are related but not identical. Minimax over parameter values asks: what is the worst individual outcome in the uncertainty set, and what decision is best against that? DRO asks: what is the worst expected cost under any distribution in 𝔻, and what decision is best against that? The difference is the expectation operator: DRO still averages over outcomes under each distribution; it only minimax-es over which distribution is used for that average. This makes DRO typically less conservative than parameter-level worst-case thinking, because an extreme distribution is not the same as an extreme scenario.

Where this shows up in practice

Supply Chain

Demand at each customer node is uncertain and the true distribution is estimated from limited data. Ambiguity sets defined by moment conditions give replenishment and routing policies that hold up across plausible distribution shapes.

Energy

Renewable generation (wind, solar) has uncertain output whose distribution shifts with season and weather pattern. DRO dispatch and unit commitment plans remain feasible across the ambiguity set without overfitting to a fixed historical distribution.

Finance

Asset return distributions are non-stationary. Portfolio optimisation over an ambiguity set of return distributions avoids treating a sample covariance matrix as the ground truth, yielding portfolios more robust to distributional shift.

Healthcare

Patient arrival rates vary by day, season, and external events. Staff scheduling models use ambiguity sets to produce rosters that are robust to distributional shift without over-staffing for extreme scenarios.

The first question to ask about any model claiming distributional robustness: what is the ambiguity set, and how was it constructed from data?

Related concepts

The Scarf Principle (distributionally robust newsvendor, 1958)

When the demand distribution is unknown but its mean and variance are observed, the minimax stocking quantity protects against the worst-case distribution consistent with those moments — and that worst-case distribution is not the one you fitted.

Back to DO Radar, 8 May 2026