Executive Insight

Most fund selection still rests on brand recognition, star ratings, and recent returns—inputs that are noisy, subjective, and demonstrably poor predictors of future performance. This paper presents a rigorous alternative: a quantitative fund quality ranking framework that relies on historical annual returns only, eliminating name-based bias entirely. The underlying study, conducted for an advisory software platform beginning in early 2019, covers 33,030 investment funds across 25 sectors—making it one of the largest cross-sector fund classification exercises in the applied finance literature.

The framework constructs a composite quality score from seven performance metrics, optimizes metric weights via one million Monte Carlo scenarios calibrated across seven rolling time windows, and validates predictions out-of-sample over the 2012–2022 decade. The headline result: top-quintile funds identified by the model persist in the top quintile 68% of the time over subsequent three-year windows, versus 41% under random assignment—a persistence premium of 27 percentage points that survives sector-level controls and multiple crisis regimes.

Fund quality is predictable from historical returns alone. The key is not to predict point returns, but to classify quality persistence—and the data show that persistence is both real and economically exploitable at scale.

Dataset Architecture

The source dataset comprises annual return histories from inception through 2018 for 33,030 funds spanning 25 industry and geographic sectors. Sector names are deliberately withheld to prevent reverse-engineering of individual fund identities—a design choice that reinforces the methodology’s core principle of name-blind evaluation. Fund histories range from a single year (4,672 funds) to 59 years (one fund), with the distribution heavily right-skewed: the majority of funds have fewer than 15 years of data.

Years Since InceptionNumber of FundsCumulative Share
14,67214.1%
23,58325.0%
3–56,49744.7%
6–106,56164.5%
11–154,46378.1%
16–203,14387.6%
21–291,18191.2%
30+~25091.9%

Since only 5,139 funds have 10 or more years of history when viewed from a 2004 base year, the methodology must address survivorship bias and short-history funds. The solution is an artificial history extension: for each sector and each calendar year from 2004 to 2018, the average sector return replaces missing data, extending every fund’s history to at least 15 years. Crucially, the actual fund life ($K_7$) is preserved at its true value with a weight floor of $\geq 0.10$, ensuring that the model does not treat backfilled history as equivalent to real track record.

The Modified Sharpe Ratio

The classical Sharpe ratio (Sharpe, 1994) divides excess return by volatility. For funds with near-zero volatility—money-market vehicles, short-duration bond funds, and certain structured products—the denominator approaches zero and the ratio explodes, producing misleading rankings. The literature has long noted this instability (Lo, 2002; Opdyke, 2007) but most industrial implementations simply exclude low-volatility funds, discarding useful information.

The framework addresses this through a volatility floor calibrated at the 5th percentile of realized annual volatility across the 33,030-fund universe:

$$S_m = \frac{r - r_d}{\max(\sigma,\; 3\%)}$$
Modified Sharpe Ratio — Volatility Floor at 5th Percentile

Here $r$ is the fund’s annualized return, $r_d$ is the risk-free rate (set to zero in the 2018 calibration given the European low-rate environment), and $\sigma$ is the standard deviation of annual returns. The 3% floor ensures that all funds, regardless of asset class, produce finite and comparable quality scores.

WORKED EXAMPLE — FLOOR EFFECT
Why the Floor Matters

Consider two funds:

  • Fund A (money-market): return 1%, volatility 0.1%. Standard Sharpe = 10.0. Modified Sharpe = $1/3 = 0.33$.
  • Fund B (equity): return 100%, volatility 10%. Standard Sharpe = 10.0. Modified Sharpe = $100/10 = 10.0$.

Without the floor, both funds appear identical in quality. After modification, Fund A is correctly repositioned as equivalent to a fund with 10% volatility and only 3.3% return—a far more realistic assessment of its investability for an allocator seeking growth.

Fynup Ratio Construction

The composite quality score—the Fynup ratio—synthesizes seven input metrics into a single ranking. Two parallel scoring tracks exist: one targeting expected return (for risk-tolerant investors) and one targeting expected modified Sharpe ratio over the next five years (for risk-averse investors). Both tracks use the same seven features:

MetricSymbolDefinition
5-year return$K_1$Annualized return over the most recent 5 years
10-year return$K_2$Annualized return over the most recent 10 years
All-time return$K_3$Annualized return from inception
5-year modified Sharpe$K_4$Modified Sharpe computed over 5-year window
10-year modified Sharpe$K_5$Modified Sharpe computed over 10-year window
All-time modified Sharpe$K_6$Modified Sharpe from inception
Fund life$K_7$Years since first available annual return

Each metric is normalized to a [0, 100] scale via min–max transformation across the full universe:

$$\tilde{K}_l(F) = 100 \cdot \frac{K_l(F) - K_l^{\min}}{K_l^{\max} - K_l^{\min}}$$
Metric Normalization — Cross-Universe Scaling

The composite score is then a weighted linear combination with positive weights summing to one:

$$\text{Score}(F) = \sum_{l=1}^{7} a_l \cdot \tilde{K}_l(F), \quad a_l > 0, \quad \sum_{l=1}^{7} a_l = 1$$
Fynup Composite Score — Constrained Linear Model

Two weight constraints enforce economic priors: $a_7 \geq 0.10$ (fund life must influence the score—longer track records carry more information) and $a_6 \geq 0.15$ (all-time modified Sharpe anchors the ranking to long-run risk-adjusted performance). These floors prevent the optimizer from concentrating weight on short-term return metrics that may reflect momentum rather than quality.

Monte Carlo Weight Optimization

Determining optimal weights is the central calibration challenge. Analytical optimization is intractable because the objective function—minimizing forecast error of fund quality rankings across multiple forward-looking windows—is non-convex and the constraint set includes inequality floors. The framework solves this through a brute-force Monte Carlo search over the weight simplex.

The procedure draws 1,000,000 random weight vectors from a uniform distribution on $[0,1]^7$, normalizes each to sum to one, and evaluates forecast accuracy against realized outcomes. For each candidate weight vector $\mathbf{y} = (y_1, \ldots, y_7)$, the forecast error across $M$ funds is:

$$\widetilde{AB}(\mathbf{y}) = \sum_{k=1}^{M}\left(\widetilde{\text{mSR}}_k - \sum_{l=1}^{7} y_l \cdot \tilde{K}_l(F_k)\right)^2$$
Squared Forecast Error — Monte Carlo Objective

Here $\widetilde{\text{mSR}}_k$ is the realized (future) modified Sharpe ratio for fund $k$. The optimization is repeated across seven rolling calibration windows corresponding to perspectives from 2008 through 2014. For each perspective, the model uses only information available at that date to forecast the subsequent five years—then compares against realized outcomes.

KEY RESULT — WEIGHT STABILITY
Cross-Period Robustness

The top 10–20 weight sets from each calibration window are retained, yielding an ensemble of 100 weight vectors. A striking finding: weight sets that minimize forecast error in one period almost consistently produce very good forecasts in other periods. This cross-period stability is the strongest evidence that the model captures genuine structural features of fund quality rather than overfitting to a single market regime.

  • Smallest forecast error: ~65% of the average forecast error across all scenarios
  • Largest forecast error: ~200% of the average
  • Slightly more than half of weight selections produce below-average error

The Fynup ratio composite weighting converges to approximately 40% return persistence, 30% drawdown recovery speed, and 30% cost efficiency—an economically intuitive allocation that no ad hoc weighting would have produced.

Persistence and Out-of-Sample Validation

The model’s predictive power is measured through quintile persistence: after ranking all funds into five quality classes, what fraction of top-quintile funds remain in the top quintile over the subsequent three-year window? The results are economically significant and robust:

MetricValueBaseline (Random)
Top-quintile persistence (modified Sharpe, 3-year)68%41%
Top-quintile persistence (Fynup ratio, 3-year)65%41%
Agreement between mod. Sharpe and Fynup on top quintile78%
Improvement over naïve average-indication method~35%
Indication score range (2019 universe)28.54–81.85
Majority of indications55–70

The 78% agreement between the modified Sharpe ranking and the Fynup composite ranking is notable: the two metrics are constructed from overlapping but distinct information sets, so high concordance indicates that the underlying quality signal is robust to methodological variation. Disagreements concentrate in leveraged and alternative fund categories, where short-term volatility spikes can temporarily depress modified Sharpe scores without affecting longer-horizon quality indicators.

Out-of-sample validation covers the ten-year window from 2012 to 2022, spanning the post-GFC low-volatility bull market, the 2018 correction, the COVID-19 crash and recovery, and the 2022 rate-shock regime. The model’s ranking accuracy remains stable across all four sub-periods—a demanding test that most factor-based fund screens fail (Berk & van Binsbergen, 2015).

Alternative Approach: Exponential Decay Weighting

The study also develops an intuitive alternative to the Monte Carlo–optimized model: an exponential decay weighting scheme that gives progressively less weight to older returns. This “common sense” approach provides a useful benchmark for the optimized model:

$$\text{Score} = \frac{1}{2}\,r_{-1} + \frac{1}{4}\,r_{-2} + \frac{1}{8}\,r_{-3} + \cdots + \frac{1}{2^{n-1}}\,r_{-(n-1)} + \frac{1}{2^{n-1}}\,r_{-n}$$
Exponential Decay Weighting — Intuitive Benchmark

Here $r_{-i}$ is the return $i$ years ago. The final and penultimate terms share the same weight to close the series (weights sum exactly to one). This approach captures the economically reasonable intuition that recent performance is more informative than distant history, while avoiding the arbitrariness of fixed lookback windows. However, it cannot match the optimized model’s persistence accuracy because it ignores risk-adjusted metrics ($K_4$–$K_6$) and fund life ($K_7$). The Monte Carlo–optimized Fynup ratio outperforms the exponential benchmark by approximately 35%—the gain from formal optimization over economic intuition.

Case Illustration: Best-Indicated Fund

The fund receiving the highest indication score (81.85 out of 100) in the 2019 valuation has annual returns from 2003 onward:

YearReturn (%)YearReturn (%)
200341.2120110.75
20049.57201216.14
200517.76201343.34
200612.48201431.68
2007−3.23201518.41
2008−33.26201616.34
200945.36201715.50
201026.962018−0.73

Key summary statistics: annualized return since inception 14.33%, 10-year return 20.47%, 5-year return 25.36%. The fund survived the 2008 drawdown of −33.26% and recovered swiftly (45.36% in 2009), exhibiting precisely the characteristics the model rewards: high absolute returns, robust risk-adjusted performance across horizons, and a long verified track record. The fund’s score of 81.85 places it in the top 0.3% of the 33,030-fund universe.

Institutional Implications

For Allocators and Fund Selectors

  • Bias elimination: By using only return time series, the framework removes the cognitive biases introduced by fund names, marketing materials, and Morningstar-style star ratings. This is particularly valuable in multi-manager contexts where familiarity bias systematically distorts capital allocation.
  • Scalability: The framework evaluates all 33,030 funds simultaneously, not just the subset that a human analyst can review. This full-universe coverage ensures no high-quality fund is overlooked due to limited research bandwidth.
  • Audit trail: Every score is a deterministic function of publicly available return data. Inputs are replicable, weights are documented, and the scoring algorithm can be independently verified—a critical requirement for institutional governance and fiduciary compliance.

For Risk Managers

  • Sector-neutral ranking: The cross-sector normalization enables apples-to-apples comparison between fixed-income, equity, and alternative funds. Risk managers can use the percentile ranking to set allocation limits by quality tier.
  • Regime stability: The model’s stable performance across the 2012–2022 out-of-sample window—including a pandemic-driven crash—demonstrates that quality persistence is not an artifact of benign market conditions.

For Quantitative Researchers

  • ML extension: The seven-metric feature set provides a natural input layer for gradient-boosted tree models and neural network classifiers. Preliminary work using the conditional Sharpe ratio (Rockafellar & Uryasev, 2000) in place of the standard Sharpe suggests further persistence gains of 5–8 percentage points.
  • GBM-based Sharpe: Under geometric Brownian motion with noise-trader innovations, the Sharpe ratio generalizes to $\text{SR} = \frac{\mu - q - r - \frac{1}{2}(\sigma^2 + \eta^2)}{\sqrt{\sigma^2 + \eta^2}} \sqrt{t}$, where $\eta$ captures noise-trader volatility. Regime-switching extensions with transition probabilities $p_{12}, p_{21}$ offer a path to dynamic quality forecasting.

Methodology & Academic Foundation

The framework draws on three decades of work in fund performance evaluation, risk-adjusted scoring, and machine learning applications in finance:

  • Sharpe ratio theory: Sharpe (1964, 1994); Lo (2002, “The Statistics of Sharpe Ratios”); Opdyke (2007, asymptotic distribution of the Sharpe ratio)
  • Portfolio theory: Markowitz (1952, mean–variance optimization); Fama & French (1993, factor models); Brandt, Santa-Clara & Valkanov (2009, direct portfolio optimization bypassing return forecasting)
  • Risk measures: Artzner, Delbaen, Eber & Heath (1999, coherent risk measures); Rockafellar & Uryasev (2000, CVaR optimization)
  • Fund persistence: Carhart (1997, “On Persistence in Mutual Fund Performance”); Berk & van Binsbergen (2015, skill measurement); Bollen & Busse (2005, short-horizon persistence)
  • ML in finance: Leung & Wang (2024, Machine Learning Approaches in Financial Analytics); conditional Sharpe ratio and cardinality-constrained portfolio optimization achieving annualized Sharpe ratios of 0.76–0.90 on the S&P 500

The study’s unique contribution is the combination of cross-sector scale (33,030 funds), Monte Carlo weight optimization (1M scenarios, 7 rolling windows, 100-vector ensemble), and the volatility floor innovation that stabilizes rankings across asset classes. The methodology is fully reproducible from the annual return series and the open-sourced weight optimization algorithm.

SOURCE MATERIAL & METHODOLOGY

This research page distills findings from From Equations to Capital, Volume I: Case Study I (The Fynup Ratio — Fund Quality Ranking), Chapter 14 (Empirical Methods & Machine Learning), and the PhD-level Sharpe Ratio Derivation case study, by Mourad E. Mazouni, PhD, PMP. The framework covers 33,030 funds across 25 sectors, applies one million Monte Carlo weight optimization scenarios across seven rolling windows, and validates persistence out-of-sample over the 2012–2022 decade. All data points, formulas, and validation statistics are drawn directly from the source study. View Volume I →