Trading the Breaking

Trading the Breaking

Quant Lectures

[QUANT LECTURES] Strategy evaluation & statistical validation (PART I&II)

Statistics for algorithmic traders

Quant Beckman's avatar
Quant Beckman
Dec 09, 2025
∙ Paid

Update for the part II:

I’m currently working on these parts, so it will be ready in the next few weeks:

IMPORTANT

*Again, this chapter is dense and practical, so it’s presented in multiple parts.

**The notebook with the code will be updated once the whole chapter is finnished.

Strategy evaluation & statistical validation

This chapter turns performance measurement into a falsification engine. Instead of staring at a single Sharpe Ratio, you’ll dissect every layer of a strategy: trade-by-trade behaviour, path dependency, benchmark-relative risk, tail losses, and the probability that an apparent edge survives both non-normal returns and selection bias. The workflow runs from simple trade-level ratios to benchmark metrics, tail- and drawdown-risk measures, block-bootstrap resampling, purged / embargoed cross-validation and walk-forward analysis—ending in probabilistic tests like PSR, DSR and CPCV that charge you for every backtest you’ve ever run.

What’s inside:

  1. Event-based diagnostics. Open the black box of trades: move from smoothed portfolio returns to the discrete sequence of wins and losses that actually generates P&L, and use this to check whether the data matches your narrative about the strategy.

  2. Core profitability ratios. Use Profit Factor, Awal Ratio (Average Win / Average Loss), and Win Rate to characterise the payoff profile—high-probability nickel-picking vs low-hit-rate trend following—and to see whether total gains truly dominate total losses.

  3. Expectancy engine. Build the expectancy identity linking Win%, Avg Win and Avg Loss; compute per-trade expectancy and the implied break-even win rate so you can tell whether a broken system needs better signals, better payoffs, or both.

  4. Economic vs statistical expectancy. Translate statistical edge into economic edge by subtracting commissions, fees and slippage—showing when a “significant” backtest is still worthless once real-world frictions are included.

  5. Path, risk & duration profile. Connect the trade engine to the equity curve via Rina Index, trade-level volatility, holding periods for winners vs losers, streak lengths and time-under-water—metrics that capture both variance drag and psychological pain.

  6. Benchmark comparison metrics. Evaluate performance in context using the Information Ratio (alpha per unit of tracking error) and Treynor Ratio (excess return per unit of beta) to judge whether a strategy deserves capital next to simple benchmark exposure.

  7. Distribution- and tail-risk ratios. Go beyond volatility with the Omega Ratio, Value-at-Risk, Conditional VaR and Calmar/MAR ratios, including coherent CVaR-based alternatives and drawdown-focused metrics that separate genuine tail risk from one-path anecdotes.

  8. Probabilistic Sharpe Ratio (PSR). Replace “Sharpe = 1.4” with “There’s an X% probability the true Sharpe exceeds SR*,” using skewness- and kurtosis-adjusted standard errors to penalise negative skew and fat tails.

  9. Deflated Sharpe Ratio (DSR). Pay a statistical price for data mining: estimate the Sharpe you’d get from the luckiest of N random trials and use it as the hurdle inside PSR, yielding the probability that your best backtest survives multiple-testing and selection bias.

  10. Resampling for dependent returns. Fix IID bootstrap myths by using moving, circular and stationary block bootstraps that respect autocorrelation and clustered P&L; build empirical distributions and asymmetric confidence intervals for Sharpe, max drawdown and path-shape metrics, plus randomisation and block sign-flip tests for streaks and equity curves.

  11. Purged & embargoed cross-validation (CPCV). See why standard K-fold fails on time series, then use label purging and temporal embargo to break leakage; scale up to Combinatorial Purged CV to approximate the full distribution of out-of-sample error and turn “one equity curve” into a family of realistic futures.

  12. Walk-Forward Analysis (WFA). Design rolling train/validate/test schedules with sliding or expanding windows, track parameter drift and regime-dependent configurations, and stitch OOS slices into a single equity curve—then compare WFA vs CPCV in terms of bias, variance and computational cost to decide which gatekeeper guards your production capital.

Sample
735KB ∙ PDF file
Download
Download

This post is for paid subscribers

Already a paid subscriber? Sign in
© 2025 Quant Beckman · Publisher Privacy ∙ Publisher Terms
Substack · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture