Credit risk is the risk of loss from a borrower's failure to repay. It has three dimensions: the probability of failure, the amount at risk, and the fraction that cannot be recovered. All credit models quantify these three dimensions.
Credit risk is distinct from market risk (price volatility) and liquidity risk (inability to transact). They require separate analytical frameworks and capital treatments.
Banks assess whether a borrower's cash flows can service principal and interest. PD is estimated from financial ratios, industry data, and macro factors.
Investors price credit spreads above the risk-free rate. Spreads embed both actual default probability and a risk premium above the actuarially fair rate.
Project lenders model cash flow shortfalls against debt service over 15–25 year tenors. DSCR covenants define the early warning threshold before technical default.
Satellites generate revenue streams. Those streams can default if the asset fails, demand falls, or the operator is financially distressed. SFIS applies credit risk methodology at asset level.
Expected Loss is a statistical mean. A lender with a 1,000-loan book earns its spread on 975 loans and loses capital on roughly 25. EL is the average loss per unit of exposure. Pricing, IFRS 9 provisioning, and regulatory capital are all calibrated from this identity.
EL does not require predicting which loan defaults. It requires knowing the distribution of outcomes. Credit risk modelling is a statistical discipline, not a case-by-case prediction problem.
PD is estimated from three broad methodological families. The choice of methodology determines whether the output is suitable for pricing, provisioning, stress testing, or none of the above.
| Approach | Data | Best Use | Key Limitation |
|---|---|---|---|
| Historical Frequency | Cohort default rates | Large rated portfolios | Lags cycle. Needs comparable samples. |
| Market-Implied | CDS spreads, bond yields | Liquid corporate bonds | Risk-neutral. Embeds premium above true PD. |
| Structural (Merton) | Asset value, volatility | Private firms, infra | Asset value is unobservable. |
| Bayesian / ML | Mixed signals + telemetry | Sparse data, novel assets | Requires principled prior specification. |
| Asset Class | Seniority | Recovery Rate | LGD |
|---|---|---|---|
| Corporate Loans | Senior Secured | 65–80% | 20–35% |
| Corporate Bonds | Senior Unsecured | 35–50% | 50–65% |
| Infrastructure Debt | Senior Secured | 70–90% | 10–30% |
| Sovereign Debt | Unsecured | 25–60% | 40–75% |
| Satellite Assets | Operator-dependent | 5–40% (est.) | 60–95% |
For amortising infrastructure loans, EAD declines as principal is repaid. Early-year EAD is high but PD may be lower as construction risk resolves. Late-year EAD is lower but cash flow risk can rise as maintenance costs increase and residual asset value declines.
Expected Loss is the mean of the loss distribution. Lenders price it into spreads and provision for it under IFRS 9. It is a cost of doing business. Unexpected Loss is the variance around the mean. Capital is held to absorb UL at 99.9% confidence under Basel III.
Credit loss distributions are right-skewed with fat tails. Most years produce losses near EL. Rare events produce losses far exceeding EL. The tail is driven by correlated defaults — borrowers failing simultaneously due to shared macro shocks.
Risk-Adjusted Return on Capital = (Revenue − EL − OpEx) / Economic Capital. Hurdle rate typically 10–15%. Enables comparison across credit exposures.
Stage 1: 12-month EL. Stage 2: lifetime EL on significant credit deterioration. Stage 3: lifetime EL on defaulted assets. Triggers tied to PD migration.
PD = N(−DD)
Distance to Default measures standard deviations between current asset value and the default boundary. A firm with DD = 4 requires a 4-sigma decline in asset value to default over the horizon. KMV extended this framework using equity market data to infer V and σ empirically.
PD(T) = 1 − e^(−λT)
Reduced-form models treat default as a Poisson process. The hazard rate λ(t) can be time-varying and driven by state variables: macro factors, credit spreads, industry indicators. Jarrow-Turnbull and Duffie-Singleton are the dominant frameworks, both requiring liquid market prices for calibration.
| Structural | Reduced-Form | |
|---|---|---|
| Default trigger | Asset value crossing boundary | Exogenous Poisson process |
| Data requirement | Balance sheet, equity vol | CDS / bond prices |
| Default predictability | Predictable in theory | Unpredictable by design |
Baseline. Interpretable coefficients. Linear log-odds assumption. Strong on tabular financial data. Preferred by regulators for transparency.
Handles non-linear interactions and missing data. XGBoost and LightGBM are industry standards. Less interpretable without SHAP decomposition.
Ensemble of decision trees. Reduces variance. Provides feature importance scores. Useful for variable selection before final model specification.
Produces a PD distribution, not a point estimate. Updates as new data arrives. Handles sparse data via principled priors. Output: P(PD | data).
Posterior ∝ Likelihood × Prior
| Rating | Category | 1-Year PD | 5-Year PD | Typical Spread |
|---|---|---|---|---|
| AAA | Investment Grade | 0.00% | 0.09% | 20–50 bps |
| AA | Investment Grade | 0.02% | 0.27% | 40–80 bps |
| A | Investment Grade | 0.06% | 0.58% | 70–130 bps |
| BBB | Investment Grade | 0.18% | 1.90% | 120–250 bps |
| BB | High Yield | 0.85% | 7.81% | 250–500 bps |
| B | High Yield | 3.24% | 19.12% | 450–900 bps |
| CCC | Distressed | 22.57% | 49.33% | >900 bps |
Two borrowers with identical PDs produce very different portfolio risk depending on default correlation. At zero correlation, large-number effects smooth losses. At high correlation, losses cluster: near-zero defaults in good years, catastrophic losses in bad years.
Classical maximum likelihood estimation (MLE) treats model parameters as fixed unknowns. You estimate them from data alone. With 1,000 mortgage defaults, that works. With 12 recorded satellite operator defaults over 40 years of commercial spaceflight, it does not. MLE produces point estimates with artificially narrow confidence intervals that misrepresent the true state of uncertainty.
Bayesian inference treats parameters as distributions. You start with a prior belief, update it with whatever data you have, and produce a posterior distribution over the parameter. The width of that posterior is an honest representation of how much you do and do not know.
// Posterior = Likelihood × Prior / Evidence
P(data) is a normalising constant. In practice we compute: Posterior ∝ Likelihood × Prior
| Frequentist (MLE) | Bayesian | |
|---|---|---|
| Output | Point estimate + standard error | Full posterior distribution over θ |
| Prior knowledge | Ignored | Incorporated formally as P(θ) |
| Small data behaviour | Unstable, overfit | Prior regularises the estimate |
| Uncertainty reporting | 95% confidence interval (frequentist) | 95% credible interval (direct probability statement) |
| Sequential updating | Requires full re-estimation | Prior posterior becomes next prior naturally |
| New asset types | Fails without labelled default data | Prior from analogous assets, updated as data arrives |
The Beta distribution is the natural prior for a probability parameter. It is defined on [0,1], has two shape parameters α and β, and is conjugate to the Binomial likelihood — meaning the posterior is also Beta. This makes updating analytically tractable.
E[PD] = α / (α + β)
Var[PD] = αβ / [(α+β)²(α+β+1)]
Uninformative limit: α = β = 1 (uniform prior over [0,1])
You have three information sources for the prior. Each maps to a different prior parameterisation.
For a binary default outcome (defaulted = 1, survived = 0), the natural likelihood is Binomial. Given n exposures and d defaults, the probability of observing exactly d defaults if the true PD is θ is:
log L(θ | d, n) = const + d·log(θ) + (n−d)·log(1−θ)
The shape of the likelihood function carries information. A wide, flat likelihood means the data is consistent with many PD values — the data has low information content. A sharp, narrow likelihood means the data strongly localises the PD. The posterior combines the prior's shape with the likelihood's shape.
Financial default is not the only signal. Telemetry observations also carry information about default probability. ALPHA constructs a composite likelihood across three signal types.
| Signal | Likelihood Form | Update Direction |
|---|---|---|
| Binary default event | Binomial L(θ | d, n) | Direct PD observation |
| Orbital health index (SIGMA score) | Probit link: P(default) = Φ(β₀ + β₁·SIGMA) | High SIGMA → lower PD |
| DSCR deterioration | Logistic: log-odds shift per DSCR unit | DSCR < 1.1× → PD spike |
| Insurance coverage change | LGD adjustment, not PD directly | Coverage lapse → LGD up |
Data: d defaults out of n exposures
Posterior: PD | data ~ Beta(α + d, β + n − d)
This is the cleanest possible update rule: count the events and add them to the shape parameters.
The posterior mean is a weighted average of the prior mean and the MLE:
Bayesian inference is naturally sequential. When a new telemetry observation or financial report arrives, today's posterior becomes tomorrow's prior. No re-estimation from scratch is required. The update rule is the same conjugate formula applied again.
The Beta-Binomial conjugate model is analytically tractable. But ALPHA's full model is not: it conditions PD on multiple correlated inputs (SIGMA score, DSCR, operator leverage, insurance coverage, orbital regime). The joint posterior across all parameters requires MCMC.
2. Propose θ* ~ q(θ* | θ_current) // proposal distribution
3. Compute acceptance ratio r = [P(data|θ*) × P(θ*)] / [P(data|θ) × P(θ)]
4. Accept θ* with probability min(1, r)
5. Repeat N times → samples approximate the posterior
A healthy trace plot shows a "fuzzy caterpillar" — rapid mixing, no trends, no stuck regions. A pathological trace shows slow drift or periods where the chain does not move (rejection streaks).
Acceptance rate is the fraction of proposed moves accepted. Too low (below 10%): proposals jump too far, almost all rejected, chain does not explore. Too high (above 60%): proposals are tiny, chain moves very slowly. Tune the proposal scale to hit 23–44%.
Autocorrelation at lag k measures correlation between chain values θ_t and θ_{t+k}. High autocorrelation means the chain moves slowly and consecutive samples carry redundant information. Effective sample size (ESS) = N / (1 + 2 Σ autocorr_k). With 1,000 raw samples and high autocorrelation, ESS might be only 50.
R-hat (Gelman-Rubin statistic) compares variance within chains to variance between chains. Run M independent chains with different starting points. If they have converged to the same distribution, within-chain and between-chain variance should be equal. R-hat close to 1.0 (below 1.05) indicates convergence. R-hat above 1.1 is a red flag.
From the posterior samples, you extract three numbers for each horizon: the posterior mean (point estimate for pricing), and the 5th and 95th percentiles (the 90% credible interval for stress testing and capital allocation). The width of the credible interval tells the credit committee how confident the model is.
Credible interval: [Q₀.₀₅(PD | data), Q₀.₉₅(PD | data)]
Predictive distribution: P(default_new | data) = E[PD | data]
Posterior sensitivity analysis moves one input at a time and measures the shift in posterior mean PD. It identifies which variables are driving the estimate. Variables with high sensitivity require more rigorous data validation. Variables with low sensitivity can be monitored at lower frequency.
SHAP (SHapley Additive exPlanations) decomposes the model output into additive contributions from each input variable, averaged over all possible orderings of variable inclusion. It satisfies three axioms: efficiency (contributions sum to the total output), symmetry (equal contributions for equal variables), and linearity (additivity for independent models).
φⱼ measures the average marginal contribution of feature j across all subsets.
Data ingestion layer. Space-Track TLE data (orbital elements, decay rates), CelesTrak (operational status, manoeuvre history), UCS Satellite Database (operator identity, launch date, orbital regime), operator financial statements (DSCR, leverage, revenue backlog), insurance registry (coverage type, limits, claims history). Data refreshed at 24-hour cadence for telemetry, quarterly for financials.
| Source | Signal Type | Update Freq | Feeds |
|---|---|---|---|
| Space-Track | TLE orbital elements | Daily | SIGMA decay, manoeuvre components |
| CelesTrak | Operational status flags | Daily | SIGMA anomaly detection |
| Operator financials | DSCR, leverage, backlog | Quarterly | ALPHA likelihood function |
| Insurance registry | Coverage, claims | Event-driven | LGD module |
SIGMA scoring layer. Seven-factor orbital risk framework. Each factor scored 0–100, weighted and aggregated into a composite SIGMA score. SIGMA is the primary structural proxy for unobservable asset value V in the Merton framework. High SIGMA (healthy orbit) → high implied asset value → lower PD prior.
Prior construction. ALPHA uses a regime-specific prior. Satellites are classified into three prior regimes based on orbital altitude and operator type:
| Regime | Prior | Prior Mean | Rationale |
|---|---|---|---|
| GEO Commercial | Beta(2, 130) | 1.5% | Low physical failure rate, high operator maturity |
| LEO Broadband | Beta(3, 80) | 3.6% | New operators, constellation execution risk |
| MEO / Novel | Beta(2, 50) | 3.8% | Limited analogues, higher uncertainty |
Likelihood update. ALPHA constructs a composite log-likelihood from three signal types. The SIGMA score enters via a probit link function. DSCR enters via a logistic regression coefficient estimated from infrastructure default data. Insurance status enters as a binary modifier on LGD, not PD.
Each new observation appends a term to the log-likelihood. This is the mechanism by which telemetry cycles continuously update the PD estimate without requiring full model re-estimation.
Posterior PD distribution. For the univariate Beta-Binomial model (analytical). For the multi-signal model: Hamiltonian Monte Carlo via NumPyro, 4 chains × 2,000 post-warmup samples, R-hat convergence criterion < 1.05, ESS per parameter > 400. The posterior is summarised as mean, 5th, and 95th percentile for each horizon.
Output layer. Three outputs per asset, per horizon. Formatted for institutional credit committee consumption.
| Output | Value | Use Case |
|---|---|---|
| Posterior Mean PD | Point estimate for pricing | Spread floor = EL = PD × LGD × EAD |
| 95th Pct PD (stress) | Worst-case credible PD | Economic capital sizing at 99.9% |
| CI Width (95th − 5th) | Model uncertainty flag | Triggers enhanced diligence if > 3× mean |
| SHAP Attribution | Variable contributions | Explainability for credit committee |
| Implied Rating | PD mapped to S&P scale | Peer comparison to rated instruments |
The analytical loss distribution (e.g. the normal approximation used in Basel IRB) makes strong parametric assumptions. Monte Carlo makes no distributional assumption about the portfolio loss. It simulates default outcomes directly for each asset in each scenario, then counts the losses. The result is an empirical loss distribution built from the data you gave it.
Draw systematic factor Z ~ N(0,1) // shared macro shock
For each asset i:
Draw idiosyncratic ε_i ~ N(0,1)
X_i = √ρ · Z + √(1−ρ) · ε_i // asset return
Default_i = 1 if X_i < Φ⁻¹(PD_i)
Loss_s = Σᵢ Default_i × LGD_i × EAD_i
Sort {Loss_s}. VaR(99.9%) = 999th percentile of 1,000 simulations.
This is the one-factor Vasicek model underlying Basel II/III IRB capital formula.
The systematic factor Z represents the macro state of the world. When Z is very negative, the economy is in severe stress and many assets breach their default threshold simultaneously. This is how correlated defaults emerge from individual asset-level simulations without explicitly modelling pairwise correlations.
VaR(99%) answers: "what is the loss at the 99th percentile?" It says nothing about what happens in the 1% tail beyond that point. CVaR (Conditional Value at Risk, also called Expected Shortfall) answers: "given that we are in the tail beyond VaR, what is the average loss?" CVaR is subadditive — portfolio CVaR is less than or equal to the sum of individual CVaRs. VaR is not. Basel III moved from VaR to ES (Expected Shortfall) for market risk capital exactly for this reason.
= (1/(1−α)) × ∫_{VaR_α}^{∞} L · f(L) dL
Sklar's theorem: any joint distribution F(x₁,…,xₙ) can be written as C(F₁(x₁),…,Fₙ(xₙ)) where C is a copula and F₁…Fₙ are the marginal CDFs. The copula captures correlation structure independently of the marginals. You can combine any set of individual PD models with any dependence structure.
// C is the copula: the joint distribution of uniform marginals
Gaussian copula: C(u₁,…,uₙ) = Φₙ(Φ⁻¹(u₁),…,Φ⁻¹(uₙ); Σ)
t-copula: C(u₁,…,uₙ) = tₙ,ν(t_ν⁻¹(u₁),…,t_ν⁻¹(uₙ); Σ)
David Li's 2000 paper introduced the Gaussian copula to credit correlation modelling. It became the standard for CDO pricing. The Gaussian copula has zero tail dependence: in the limit of extreme stress, defaults remain approximately independent. This is mathematically elegant but empirically wrong. In the 2008 crisis, defaults became highly correlated in the tail — exactly the region the Gaussian copula said was uncorrelated.
| Copula | Tail Dependence | Parameters | Use Case | Limitation |
|---|---|---|---|---|
| Gaussian | Zero (upper and lower) | Correlation matrix Σ | Normal times, large portfolios | Underestimates crisis correlation |
| Student-t | Symmetric, non-zero | Σ, ν (degrees of freedom) | Fat-tail environments | Symmetric: same upper/lower tail dep. |
| Clayton | Lower tail only | θ (one parameter) | Downside clustering | No upper tail dependence |
| Gumbel | Upper tail only | θ (one parameter) | Boom correlation | No lower tail dependence |
| Frank | Zero both tails | θ | Symmetric mid-range | Similar failure to Gaussian in tails |
A stress test specifies a macro or sector scenario, translates it into shifts in PD, LGD, and EAD across the portfolio, and computes the resulting loss. The scenario is defined by humans, not by a model. This is its strength: it can capture scenarios that have never occurred in historical data.
Reverse stress testing is conceptually different from scenario stress testing. Instead of asking "what does scenario X do to my losses?", it asks "what scenario causes my losses to exceed my capital buffer?" This forces explicit identification of the conditions that would breach solvency. Regulators increasingly require reverse stress tests because they reveal concentration risks that forward stress tests miss.
Find: minimum PD multiplier m* such that VaR(99.9%, m*×PD, ρ) = C
When you enter a derivative with a counterparty, you are exposed to two risks: market risk (the value of the derivative moves) and credit risk (the counterparty defaults while the derivative is in-the-money to you). CVA is the expected loss from counterparty default, priced into the derivative at inception.
≈ (1 − R) × Σₜ EE(tₖ) × [PD(tₖ) − PD(tₖ₋₁)]
dPD(t): marginal default probability in time interval [t, t+dt]
CVA is paid upfront as a reduction in the derivative's fair value.
| Adjustment | Full Name | Sign | What It Prices |
|---|---|---|---|
| CVA | Credit Valuation Adj. | Debit | Counterparty defaults while contract is asset to us |
| DVA | Debit Valuation Adj. | Credit | We default while contract is liability to counterparty |
| FVA | Funding Valuation Adj. | Debit | Cost of funding uncollateralised derivative positions |
| MVA | Margin Valuation Adj. | Debit | Cost of posting initial margin under SIMM |
| KVA | Capital Valuation Adj. | Debit | Cost of holding regulatory capital against the trade |
| ColVA | Collateral Valuation Adj. | Both | Value of optionality in collateral agreement terms |
Under IAS 39 (the old standard), a bank only provisioned for a loan when a loss event had actually occurred. The 2008 crisis revealed this was procyclical: banks built provisions only after the crisis hit, amplifying the shock. IFRS 9 requires provisioning for expected future losses from day one. This is the ECL model.
D = discount factor = 1/(1 + EIR)^t
Lifetime ECL = Σₜ PDₜ × LGDₜ × EADₜ × Dₜ // sum over loan life
PDₜ: conditional PD in period t (given no default in prior periods)
The formula looks like EL but discounted over the remaining life.
The trigger for Stage 1 to Stage 2 migration is SICR. Banks must define SICR in their IFRS 9 methodology. The most common approaches:
| Approach | Trigger | Threshold (typical) |
|---|---|---|
| Absolute PD threshold | PD exceeds a fixed level | PD > 1% for prime portfolios |
| Relative PD change | PD has doubled since origination | PD_current / PD_origination > 2× |
| Watchlist flags | Internal risk management flags | 30 days past due (backstop) |
| Qualitative triggers | Covenant breach, rating downgrade | Below investment grade from IG |
Model validation has three components: discrimination (does the model rank-order risk correctly?), calibration (are the predicted PDs accurate on average?), and stability (does the model produce consistent outputs as inputs change?). Each requires different tests.
Gini Coefficient = 2 × AUC − 1 // ranges 0 (random) to 1 (perfect)
KS Statistic = max|F_defaulter(s) − F_non-defaulter(s)| // max separation of score distributions
Discrimination says nothing about calibration. A model can rank-order perfectly but systematically overstate or understate PDs. Calibration checks whether predicted PD of 5% corresponds to an observed default rate of approximately 5%.
Your result: BS = 0.186 vs naive benchmark BS = 0.207. Improvement exists but is modest.
| Test | What It Checks | Pass Threshold | Your Pipeline |
|---|---|---|---|
| AUC / Gini | Rank ordering | AUC > 0.65, Gini > 0.30 | AUC 0.663, Gini 0.326 ✓ |
| KS Statistic | Score distribution separation | KS > 0.20 | ~0.29 (estimated) |
| Brier Score | Calibration accuracy | Below naive benchmark | 0.186 vs 0.207 ✓ |
| Hosmer-Lemeshow | Calibration across deciles | p-value > 0.05 | Not reported — required |
| R-hat (Bayesian) | MCMC convergence | R̂ < 1.05 | Not reported — required |
| Acceptance Rate | MCMC efficiency | 23–44% | 8.3% — failed |
| Population Stability | Score distribution over time | PSI < 0.10 | Not yet implemented |
A CDS is a bilateral contract. The protection buyer pays a periodic premium (the CDS spread) to the protection seller. If the reference entity defaults, the protection seller pays the notional minus recovery. The CDS spread is the market price of credit risk for the reference entity.
s × Σₜ DF(t) × [1 − PD_cum(t)] × Δt ≈ LGD × Σₜ DF(t) × ΔPD(t)
// At market: s ≈ PD_hazard × LGD (for flat term structure)
Inverting: hazard rate λ ≈ s / LGD. This is the reduced-form PD extraction you saw in Module 3.
A CDO (Collateralised Debt Obligation) pools credit assets and tranches the cash flows. Senior tranches absorb losses last and receive lower yields. Junior/equity tranches absorb losses first and receive higher yields. The tranching redistributes credit risk — it does not eliminate it.
| Tranche | Attachment Point | Loss Absorption | Rating Target | Yield Premium |
|---|---|---|---|---|
| Super Senior | 30–100% | Last losses absorbed | AAA | 10–40 bps |
| Senior | 15–30% | After equity and mezz | AA–A | 50–150 bps |
| Mezzanine | 5–15% | After equity | BBB–BB | 200–500 bps |
| Equity / First Loss | 0–5% | First losses absorbed | Unrated | 15–25% IRR target |
Moody's publishes the definitive empirical study of infrastructure project finance defaults. Key findings from the 2023 update (1983–2022 cohort, 8,910 projects):
| Sector | 10-Year Cum. Default | Recovery Rate | LGD |
|---|---|---|---|
| Power (contracted) | 5.2% | 80% | 20% |
| Transportation (toll) | 8.1% | 75% | 25% |
| Oil & Gas (midstream) | 6.8% | 72% | 28% |
| Telecom / Satellite | ~7–12% (estimated) | 40–60% (thin data) | 40–60% |
| PF overall average | 6.4% | 78% | 22% |
LLCR (Loan Life Coverage Ratio) = NPV(Cash Flows over loan life) / Outstanding Debt
PLCR (Project Life Coverage Ratio) = NPV(Cash Flows over project life) / Outstanding Debt
Below 1.0×: cash flow insufficient to service debt. Technical default trigger.
LLCR and PLCR give forward-looking coverage across the full term.
Sovereign ratings integrate quantitative and qualitative factors. The four major pillars in Moody's sovereign methodology:
Basel III requires banks to hold capital against unexpected losses. The Internal Ratings-Based (IRB) approach uses the bank's own PD, LGD, and EAD estimates. The Standardised Approach (SA) uses fixed risk weights by asset class. Basel IV introduces an output floor: IRB capital cannot fall below 72.5% of the SA figure.
RWA = K × 12.5 × EAD
// Capital Ratio = Eligible Capital / RWA ≥ 8% (10.5% with buffers)
The formula computes conditional expected loss at the 99.9th percentile minus EL, isolating unexpected loss.
| Approach | PD Source | LGD Source | Risk Weights | Output Floor |
|---|---|---|---|---|
| SA | Not used | Fixed by regulation | 20%–150% by rating | N/A (is the floor) |
| F-IRB | Bank estimate | Fixed by regulation | Formula-based | 72.5% of SA |
| A-IRB | Bank estimate | Bank estimate | Formula-based | 72.5% of SA |
A transition matrix records the empirical probability of moving from one rating grade to another over a fixed time horizon (typically 1 year). S&P and Moody's publish annual transition matrices from cohort studies.
// Generator matrix approach: M(t) = exp(t × Q) where Q = log(M(1))
Cohort method: track rating at start and end of period. Duration method: continuous observation.
Concentration risk arises when exposure is not sufficiently diversified across obligors, sectors, or geographies. The Herfindahl-Hirschman Index (HHI) measures single-name concentration. The effective number of exposures is 1/HHI.
Effective Number = 1 / HHI
// Granularity adjustment: GA = (1/N_eff) × σ2(LGD) × correction factor
Regulatory large exposure limit: single-name exposure ≤ 25% of Tier 1 capital.
| Limit | Threshold | Source |
|---|---|---|
| Single-name large exposure | ≤ 25% of Tier 1 | Basel / CRR Art. 395 |
| G-SIB connected limit | ≤ 15% of Tier 1 | BCBS 283 |
| Sector concentration | No hard cap — Pillar 2 | Supervisory review |
Wrong-way risk (WWR) occurs when exposure and counterparty default probability are positively correlated. Standard CVA assumes independence: E[Exposure × 1_default] = E[Exposure] × PD. Under WWR, the actual expected loss exceeds this product.
E[Loss] = E[Exposure | Default] × PD × LGD // conditional exposure > unconditional
// General WWR: macro-driven (e.g., rates move with credit spreads)
// Specific WWR: structural link (e.g., put on own stock)
Oil producer enters pay-fixed swap. Oil price drops → producer revenue falls (credit worsens) AND swap MTM rises (exposure increases). Classic specific WWR.
Satellite operator sells capacity forward. If operator faces distress, capacity delivery risk rises AND the replacement cost of the contract increases. SFIS models this linkage explicitly.
Economic capital (EC) is the internal capital a bank holds to absorb unexpected losses at a chosen confidence level (typically 99.9%). Portfolio EC benefits from diversification — the sum of standalone ECs exceeds diversified EC. The question: how to allocate the diversification benefit fairly?
Σi ECi = ECportfolio // Euler allocations sum exactly to portfolio EC
RAROCi = (Spreadi − ELi) / ECi
RAROC: Risk-Adjusted Return on Capital. Hurdle rate typically 12–15%.
Credit Risk Transfer (CRT) allows banks to reduce regulatory capital by transferring credit risk to third parties. The Significant Risk Transfer (SRT) test determines whether the bank receives capital relief. Basel requires that the originator transfers a meaningful portion of the risk — not just the first loss or just the senior tranche.
// A = attachment point, D = detachment point, pool_EL = pool-level expected loss
Tranche thickness = D − A
| CRT Mechanism | Funded? | Capital Relief | Key Condition |
|---|---|---|---|
| CDS (single-name) | No | Substitution approach | Eligible protection seller |
| Guarantee | No | Substitution approach | Irrevocable, unconditional |
| Synthetic securitisation | No | SRT test required | Meaningful risk transfer |
| True sale securitisation | Yes | Full derecognition possible | True sale opinion + SRT |
| Insurance wrap | No | Partial (Pillar 2) | Insurer rating, claims history |
Climate risk transmits to credit risk through two channels. Physical risk: extreme weather, sea-level rise, and chronic changes damage assets and reduce cash flows. Transition risk: carbon pricing, regulation, and technology shifts strand assets and alter competitive dynamics.
// PhysRisk: physical risk score (0-1), TransRisk: transition risk score (0-1)
Hot house: high physical, low transition. Disorderly: high physical, high transition.
| Transmission Channel | Physical Risk | Transition Risk |
|---|---|---|
| Asset damage / impairment | Direct | Stranded assets |
| Revenue disruption | Supply chain breaks | Demand shift |
| Operating cost increase | Adaptation costs | Carbon pricing |
| Funding cost increase | Insurance repricing | ESG screening |
Financial networks exhibit contagion: the default of one institution imposes losses on its counterparties, potentially triggering further defaults. The Eisenberg-Noe clearing model finds the unique clearing vector — the set of payments each node can make given incoming payments from others. DebtRank extends this to partial defaults.
// h_i: stress level of node i (0=healthy, 1=defaulted)
// W_ji: relative exposure of i to j (exposure / equity)
Systemic importance: the total DebtRank impact if a single node defaults.
Deep learning models capture non-linear interactions and temporal dependencies that logistic regression cannot. LSTM networks process time-series data (financial ratios, market signals, telemetry). Attention mechanisms identify which time steps and features matter most. But credit regulators require explainability (SR 11-7, SS1/23).
| Model | Accuracy (AUC) | Explainability | Data Requirement | Regulatory Status |
|---|---|---|---|---|
| Logistic Regression | 0.70–0.75 | Full | Low (100s) | Approved |
| XGBoost / GBM | 0.78–0.83 | Partial (SHAP) | Medium (1000s) | Case-by-case |
| LSTM | 0.80–0.86 | Low | High (10000s+) | Challenger only |
| Transformer | 0.82–0.88 | Very low | Very high | Research only |
// Simple models: high bias, low variance (underfitting)
// Complex models: low bias, high variance (overfitting)
The Kessler syndrome describes a cascade where orbital debris from one collision increases the probability of subsequent collisions, creating a self-reinforcing chain reaction. For credit risk, this means the assets in a congested orbital shell are correlated not just through market factors but through shared physical destruction risk.
D(t+1) = D(t) + ΔD(t) × Pcollision(t) // debris grows after each collision
// λ: spatial density factor, D(t): debris count, σ: cross-section, v_rel: relative velocity