I just ran my first FEA simulation and got a stress result of 250 MPa. My boss asked "is this result validated?" — I had no idea what to say. What exactly does V&V mean?
Great question — and honestly, most engineers don't ask it enough. V&V stands for Verification and Validation, and it's the formal process of building trust in a simulation. Think of it as two very different questions: Verification asks "Am I solving the equations correctly?" — meaning, is my software and mesh giving me the right numerical answer to the math problem I set up? Validation asks "Am I solving the right equations?" — meaning, does my math model actually reflect what happens in the real world?
So even if my solver runs without errors, that doesn't mean the result is right?
Exactly. A converged solution is not a validated solution. Classic ASME V&V 10 (for FEA) and V&V 20 (for CFD) formalize this distinction. The AIAA Guide (G-077) defines it the same way. Your 250 MPa might be a perfectly computed answer to a model that uses the wrong material law or boundary conditions — that's a validation failure, not a verification failure.
"The process of determining that a computational model accurately represents the underlying mathematical model and its solution."
— Checks the math
"The process of determining the degree to which a model is an accurate representation of the real world from the perspective of the intended uses of the model."
— Checks the physics
How do you actually verify that the FEM code is doing the math right? You can't just trust the software vendor, right?
There are three main tools for code verification. First, the patch test — this checks that a FEM element reproduces a constant strain state exactly on an arbitrary mesh. Any element that passes the patch test converges to the correct solution as the mesh is refined. If your element fails the patch test, throw it away. Second, the Method of Manufactured Solutions (MMS) — you pick an exact answer first (say, a sinusoidal stress field), derive what body forces and boundary conditions would produce it, run the solver, and compare. Third, you do a grid convergence study — keep halving the mesh size and watch the error drop at the expected rate (e.g., quadratic elements should give 4× error reduction for 2× mesh refinement, which is 2nd-order convergence).
What's Richardson extrapolation? I keep seeing that term in grid convergence papers.
Richardson extrapolation is a mathematical trick to estimate the "exact" answer from two or more mesh solutions. If you have a coarse mesh result \(f_{2h}\) and a fine mesh result \(f_h\), and your method is \(p\)-th order accurate, the extrapolated "exact" value is approximately \(f^* \approx f_h + (f_h - f_{2h})/(2^p - 1)\). The Grid Convergence Index (GCI), standardized by Roache, goes further — it gives a confidence band around that extrapolated value. In practice, aerospace and nuclear codes are required to report GCI values.
The patch test, developed by Irons and Zienkiewicz, is the gold-standard consistency check for finite elements. A patch of elements is assembled on an irregular mesh. Boundary displacements corresponding to a constant strain state are applied. The test passes if the interior nodes produce the correct (constant) strain field.
MMS is the most rigorous approach to code verification. The workflow is:
MMS is independent of physical plausibility — the manufactured solution need not satisfy any physics. This isolates coding errors from modeling errors, making it uniquely powerful for code debugging.
For solution verification, the GCI quantifies numerical uncertainty due to spatial discretization. Given three systematically refined meshes with refinement ratio \(r\):
where \(e_{21} = (f_2 - f_1)/f_1\) is the relative change between medium and fine grids, \(p\) is the observed convergence order, and \(F_s = 1.25\) is the recommended safety factor (Roache, 1994). A small GCI (typically < 2–5%) indicates grid-independent results.
OK so once the code is verified, how do I actually validate the model? I run the simulation and then compare to experiments — but what does "compare" mean quantitatively?
Just plotting simulation vs. experiment curves and saying "looks good" isn't enough for serious engineering. You need formal validation metrics. The most common are: E-value (the simple relative error at a point), the u-value (normalized error accounting for combined experimental and numerical uncertainties), and for field comparisons, the Theil inequality coefficient or L2 norm of the residual. In fatigue and stochastic problems, you also check that confidence intervals overlap — called a "feature comparison" or "Bayesian model validation."
What's uncertainty quantification? How is it different from just having measurement error in the experiment?
Great distinction. Uncertainty quantification (UQ) splits uncertainty into two fundamental types. Aleatory uncertainty is irreducible randomness — like the natural scatter in material properties between different batches of steel. You can characterize it statistically but you can't eliminate it. Epistemic uncertainty is reducible ignorance — like not knowing exactly which friction coefficient to use. Getting more test data or a better model reduces epistemic uncertainty. In aerospace certification, you must quantify both and show your simulation predictions remain conservative even at the worst-case uncertainty bounds.
For a validation comparison to be meaningful, the following must be documented:
| Metric | Formula | Use Case | Threshold (typical) |
|---|---|---|---|
| Relative error (E) | \(E = |S - E_{exp}|/E_{exp}\) | Single-point comparison | < 5–10% |
| u-value (normalized) | \(u = (S - E_{exp}) / \sqrt{u_S^2 + u_E^2}\) | Uncertainty-weighted comparison | \(|u| < 1\) |
| L2 error norm | \(\|S - E\|_2 / \|E\|_2\) | Field quantity comparison | < 5% for CFD |
| GCI | Roache formula (see above) | Numerical uncertainty band | < 2–5% |
In a real project, where in the timeline do you do all this? Do you do V&V before the analysis or after?
V&V is woven throughout the project, not bolted on at the end. Here's a typical sequence: First, during software selection and model setup, you do code verification (patch tests, order-of-accuracy checks on benchmark problems). Then, before your main production runs, you do a mesh convergence study — this is solution verification. You find the mesh size where results stop changing significantly. Then, once results are in, you compare against available experimental data for model validation. Finally, you quantify residual uncertainty and document everything. In practice, companies often maintain a "V&V matrix" — a table mapping each simulation type to the verification tests and validation experiments that support it.
Are there official standards I need to follow? My company is doing aerospace work and they mentioned something about NASA standards.
Yes, there are several, and for aerospace work you'll encounter all of them. ASME V&V 10-2006 covers the guide for verification and validation in computational solid mechanics (FEA). ASME V&V 20-2009 does the same for computational fluid dynamics and heat transfer. AIAA G-077-1998 (updated 2010) is the guide for V&V in computational fluid dynamics specifically. And for NASA-funded work or missions, NASA-STD-7009A is the standard for models and simulations — it defines maturity levels (1–6) for simulation credibility and requires documented evidence at each level.
| Standard | Domain | Issuing Body | Key Requirement |
|---|---|---|---|
| ASME V&V 10 | Solid Mechanics (FEA) | ASME | Verification plan, validation plan, UQ, documentation |
| ASME V&V 20 | CFD & Heat Transfer | ASME | GCI, experimental uncertainty, validation metric |
| AIAA G-077 | CFD | AIAA | Code verification, grid convergence, comparison to experiment |
| NASA-STD-7009A | All physics | NASA | Simulation credibility levels 1–6, evidence documentation |
| NRC NUREG-2121 | Nuclear safety | NRC (US) | Validation matrix, uncertainty treatment, code qualification |
What does a V&V documentation package actually look like for a real aerospace certification? Is there a checklist?
In aerospace, the simulation evidence package for certification typically includes: (1) a Software Qualification Report showing the solver itself has been verified against benchmark problems, (2) a Model Description Document detailing every modeling assumption and its justification, (3) a Mesh Convergence Report with GCI values for each QoI, (4) a Validation Report comparing simulation to coupon-level and component-level test data, (5) an Uncertainty Quantification Report showing the simulation uncertainty band covers the test scatter, and (6) a Traceability Matrix linking each simulation claim back to supporting evidence. For nuclear, NUREG-2121 requires a validation matrix covering all relevant phenomena, often involving thousands of test cases.
NASA-STD-7009A defines six credibility factors, each scored 0–4:
A composite score (0–4) is computed. High-consequence decisions (e.g., flight certification) require scores of 3–4.