Epistemic Uncertainty and Aleatory Uncertainty
Theory and Physics
Two Types of Uncertainty
Professor, what's the difference between epistemic uncertainty and aleatory uncertainty? They're both "uncertainty," right?
Good question. Roughly speaking, aleatory uncertainty is the inherent variability of nature itself; it cannot be eliminated no matter how much data you collect. On the other hand, epistemic uncertainty arises from a lack of knowledge and can be reduced by improving models or adding more data.
Can you give me some concrete examples?
Let's consider a car crash test. Even if you crash the same car under the same conditions, the yield stress of the steel sheet will vary slightly from lot to lot. This is natural variability—in other words, aleatory uncertainty. Even if you run the test a million times, this variability will never become zero.
On the other hand, the friction coefficient used in a crash simulation might be taken directly from literature values. We simply assume $\mu = 0.15$ because we don't know the true value, but we could get closer to it with precise experiments. This is epistemic uncertainty.
I see! So the key point is whether it can be reduced or not!
Exactly. Another important point is that the international standard for V&V (Verification and Validation), ASME V&V 20, requires that the two be clearly distinguished and quantified separately. If you mix them up, you won't know where there's room for improvement.
| Characteristic | Aleatory | Epistemic |
|---|---|---|
| Other Names | Irreducible uncertainty, Statistical uncertainty | Reducible uncertainty, Systematic uncertainty |
| Cause | Natural variability (probabilistic variation) | Lack of knowledge/data/model |
| Reducibility | No — inherently random | Yes — reducible by model improvement/data addition |
| Mathematical Representation | Probability distribution $f_X(x)$ | Interval $[a, b]$, belief function, P-box |
| CAE Example | Lot-to-lot variation in material strength, wind load fluctuation | Turbulence model error, unknown boundary conditions |
| Propagation Method | Monte Carlo, PCE | Interval analysis, Dempster-Shafer |
Mathematical Representation of Aleatory Uncertainty
How do you handle aleatory uncertainty mathematically?
Aleatory uncertainty is handled using probability theory. You assign a probability density function $f_X(x)$ to the input variable $X$ and then find the statistical quantities of the output $Y = g(X)$.
The expected value and variance of the output $Y$ can be expressed as:
So, if the probability distribution of the inputs is completely known, then (in principle) the distribution of the output can also be calculated, right?
Exactly. However, in practice, $g(\cdot)$ is a huge CAE model, so analytical integration is impossible. That's why we use Monte Carlo sampling or Polynomial Chaos Expansion (PCE).
Mathematical Representation of Epistemic Uncertainty
Can't epistemic uncertainty be represented by a probability distribution?
That's a sharp question. Actually, there are two schools of thought. The Bayesian subjective probability school believes it's acceptable to assign probability distributions to epistemic uncertainty as well. On the other hand, the non-probabilistic approach school argues that "assuming a probability distribution when data is insufficient is dangerous" and represents it using intervals or fuzzy sets.
In the interval representation, we only know the possible range of the variable $\theta$:
For example, the constant $C_\mu$ in a turbulence model has a range like $0.07 \leq C_\mu \leq 0.11$ depending on the literature. At present, we don't know which value within this range is the "true" one. This is epistemic uncertainty represented by an interval.
The propagation method must be completely different for probability distributions versus intervals, right?
Exactly. That's why frameworks like P-boxes and Dempster-Shafer theory are needed to handle the two types together.
Probability Box (P-box)
I've heard the name P-box, but honestly, I don't really understand it...
A P-box (Probability box) is a method of representing uncertainty using the upper and lower bounds of a cumulative distribution function (CDF). Its biggest advantage is that it can handle both aleatory and epistemic uncertainty simultaneously.
Let's think of a concrete example. Suppose the yield stress of a steel material follows a normal distribution $N(\mu, \sigma^2)$ (this is the aleatory uncertainty). However, the mean value $\mu$ is not known precisely due to insufficient lot data, and we only know $\mu \in [340, 360]$ MPa (this is the epistemic uncertainty). In this case, the CDF is not a single curve but a band-shaped region. This band is the P-box.
Ah, I see! It represents a state where the "shape" of the distribution is known, but its parameters have variability.
Exactly. P-boxes can also be used in cases with even less information (where even the distribution shape is unknown). For distribution-free P-boxes, upper and lower bounds are constructed using Chebyshev's inequality based only on the range of the mean and variance.
Dempster-Shafer Evidence Theory
I've also heard of Dempster-Shafer theory. How is it different from a P-box?
Dempster-Shafer (DS) theory is a framework for representing degrees of belief based on incomplete evidence. While probability theory assigns a single probability to each event, DS theory uses two measures—the belief function Bel and the plausibility function Pl—to represent the lower and upper bounds of probability.
Here's how it works. First, define a Basic Probability Assignment (BPA) $m$. This is a function representing the "mass" of each piece of evidence, assigned to subsets (excluding the empty set):
Then the belief and plausibility functions are defined as follows:
So Bel is the lower bound for "at least this much is believable," and Pl is the upper bound for "at most this much is plausible"?
Perfect understanding. In practice, it's used, for example, when multiple experts hold different opinions. Expert A says, "The parameter is in the range $[2, 5]$," and Expert B says, "It's in the range $[4, 7]$." You assign a degree of confidence (BPA) to each opinion and combine them using Dempster's rule of combination. This allows you to quantify how much evidence supports each interval.
That sounds like it could also be used as a tool for consensus building.
Position within ASME V&V 20
How is it regulated in ASME V&V 20?
ASME V&V 20-2009 "Standard for Verification and Validation in Computational Fluid Dynamics and Heat Transfer" defines the model validation metric $E$ as follows:
Here, $S$ is the simulation result and $D$ is the experimental data. Then, the uncertainty constituting this $E$ is decomposed into the following three types:
- $u_{\text{num}}$ — Numerical uncertainty (discretization error, iteration error, etc.): Epistemic
- $u_{\text{input}}$ — Input uncertainty (variation in material constants, boundary conditions): Mixed aleatory and epistemic
- $u_D$ — Experimental uncertainty (measurement error, reproducibility): Mixed aleatory and epistemic
The key point is that if $|E| \leq u_{\text{val}}$, you cannot yet say "the model is validated"; you must state that it is "consistent within the validation uncertainty bounds." Furthermore, epistemic uncertainty is required to be explicitly identified as a part that can be reduced by future research.
So by distinguishing them, you can see "where there is room for improvement." That's very rational.
Cases where distinguishing between aleatory and epistemic is difficult
- Model form uncertainty: For example, the choice of RANS turbulence model (k-epsilon vs SST). It is classified as epistemic because it's unclear which model is "correct," but creating an ensemble of model choices can sometimes make it appear probabilistic. ASME V&V 20 positions it as epistemic.
- Manufacturing variability: Lot-to-lot variation in material yield stress is aleatory, but systematic bias within a lot is epistemic. With sufficient data, they can be separated.
- Environmental loads: Seasonal variation in wind speed is aleatory, but changes due to future climate change are epistemic. The classification can change based on the time scale.
Numerical Methods and Implementation
Aleatory Uncertainty Propagation Methods
So how do you actually propagate aleatory uncertainty?
The most standard method is Monte Carlo Sampling (MCS). You generate $N$ samples from the probability distribution of the input variables and execute the CAE model for each sample. A histogram of the outputs is naturally created.
How large does $N$ need to be? CAE calculations take a long time even for a single run, right?
The convergence rate is $O(1/\sqrt{N})$, which is independent of input dimension—that's its strength. However, to accurately capture the tails of the distribution, such as the 99th percentile, you'd want $N \geq 10{,}000$. In practice, it's common to use Latin Hypercube Sampling (LHS) to significantly reduce the required number of samples.
| Method | Convergence Rate | Required Sample Count | Characteristics |
|---|---|---|---|
| Simple Monte Carlo | $O(N^{-1/2})$ | 10,000 to 1,000,000 | Highly versatile, usable even in high dimensions |
| Latin Hypercube | $O(N^{-1/2})$ or better | 100 to 10,000 | Stratified sampling reduces variance |
| Polynomial Chaos Expansion (PCE) | Exponential | $(p+1)^n$ ~ | Strong for low-dimensional, smooth responses |
| Kriging/Gaussian Process | — | 10 to 100 | Surrogate model, suitable for few samples |
Epistemic Uncertainty Propagation Methods
Epistemic uncertainty doesn't have a probability distribution, right? Can't you use Monte Carlo?
Correct. When epistemic uncertainty is represented by an interval, you use interval analysis to find the maximum and minimum values of the output:
For simple cases, you can just evaluate the endpoints if $g$ is monotonic. But if there are multiple epistemic parameters, you need to solve an optimization problem. Global optimization algorithms (like genetic algorithms or particle swarm optimization) are often used for this.
What if the epistemic uncertainty is represented by Dempster-Shafer theory?
In that case, you evaluate the model for each focal element and construct the output-side Bel and Pl. If there are $m$ focal elements, you need at least $2m$ model evaluations (the upper and lower bounds of each focal element).
Mixed Uncertainty Propagation (Double-Loop Method)
In reality, both aleatory and epistemic uncertainties are mixed, right? How do you handle both simultaneously?
Related Topics
なった
詳しく
報告