Folio I · The Question
AFP Reference · XXVIII April MMXXVI
As published in AFP training resources

Your forecast doesn’t
have a score.
It should.

The methodology behind the AFP article — operationalized in three minutes, on your own data.

Paste twelve months of paired budget-versus-actual figures. We compute your Uncertainty Exposure Score™ across four dimensions — accuracy, bias, volatility, and persistence — return a single number on a 0–100 scale, name the three line items driving most of your forecast fragility, and render a Monte Carlo distribution on the metric that’s costing you the most.

Free · No credit card · ~3 minutes · Six months minimum, twelve months for the most robust persistence score
The Statistic at the Center

“Eighty-six percent of finance teams have no structured measurement of how reliable their forecasts actually are.”

Source: Association for Financial Professionals · 2026 FP&A Benchmarking Survey

38 UES™ Moderate
FIG. I.a — The UES™ Badge. Aura color signals band — the bell curve at the center is the methodology made visible. Lower is better. The score above — 38, Moderate — is the calibrated demo reading.
i

Paste your data

Twelve months of paired budget-versus-actual figures. CSV or tab-separated. Six minimum produces a result; twelve captures the full annual cycle.

ii

Watch the instrument

Four dimensions resolve in turn — accuracy, bias, volatility, persistence — each scored against the institutional benchmark.

iii

See your range

A Monte Carlo distribution on your highest-variance line item. P10, P50, P90. Your forecast is not a number — it is a range.

iv

Take it with you

Your full UES™ Report by email. A shareable badge for your board deck. The methodology, operationalized.