Objective research to aid investing decisions

Value Investing Strategy (Strategy Overview)

Allocations for September 2024 (Final)
Cash TLT LQD SPY

Momentum Investing Strategy (Strategy Overview)

Allocations for September 2024 (Final)
1st ETF 2nd ETF 3rd ETF

Big Ideas

These blog entries offer some big ideas of lasting value relevant for investing and trading.

Effects of In-sample Bias and Market Adaptation on Stock Anomalies

Do stock return anomalies weaken after discovery? If so, why? In the February 2016 update of their paper entitled “Does Academic Research Destroy Stock Return Predictability?”, David McLean and Jeffrey Pontiff examine out-of-sample and post-publication performance of 97 predictors of the cross section of stock returns published in peer-reviewed finance, accounting and economics journals. For each predictor, published confidence in predictive power is at least 95%, and replication is feasible with publicly available data. The publication date is year and month on the cover of the journal. Their goal is to determine the degrees to which any future degradation in predictive power derives from: (1) statistical biases (exposed out-of-sample but pre-publication); and, (2) market adaptation to strategies used by investors to exploit anomalies (exposed post-publication). For each predictor and interval, they employ a consistent test methodology based on average monthly return for a hedge portfolio that is long (short) the fifth of stocks with the highest (lowest) expected returns based on the original study. Portfolios are equally weighted unless the original study uses value weighting. Using extensions to 2013 of the exact or best available approximations of original data for the 97 predictive variables with samples starting as early as 1926 and ending as late as 2011, they find that: Keep Reading

Sharpe Ratio, Alpha or Geometric Mean?

What is the single best performance metric an investor can use to rank performances of competing portfolios (such as mutual funds)? In his September 2016 paper entitled “Measuring Portfolio Performance: Sharpe, Alpha, or the Geometric Mean?”, Moshe Levy compares Sharpe ratio, 5-factor (market, size, book-to-market, profitability, investment) alpha and geometric mean return as portfolio performance metric. The widely used Sharpe ratio is optimal when return distributions are normal and the investor can borrow at the lending (risk-free) rate without limit for leverage. However, asset return distributions may not be normal, investors generally borrow at an interest rate above the risk-free rate and Federal Reserve Regulation T restricts borrowing to 100% of an investor’s initial capital. Moreover, investors typically restrict themselves to much lower borrowing levels. His methodology is to compare the ranking of a set of actual equity mutual funds under realistic assumptions based on each of the three metrics with the ranking produced by utility maximizing allocations for each fund paired with the risk-free asset. The better the ranking produced by the metric aligns with the utility maximization ranking, the better the metric. His baseline assumption is that actual annual borrowing rate is 3.5% above the lending rate. For robustness, he considers several levels of investor risk aversion in determining utility maximization and other gaps between borrowing and lending rates. Using theory, monthly returns for 10,145 U.S. domestic equity mutual funds, the risk-free (lending) rate and returns for the five Fama-French factors during July 2005 through June 2015, he finds that: Keep Reading

The Right Math for Analysis of Financial Markets?

Where should investors look for methodological edges in 21st century financial markets? In his brief August 2016 paper entitled “Mathematics and Economics: A Reality Check”, Marcos Lopez de Prado advises finance students (and practitioners) what mathematical/analytical expertise to acquire for successful 21st century investing and trading. Based on his experience with what kinds of analysts and mathematics are most successful in financial markets, he concludes that: Keep Reading

Best Way to Guard Against Investment Strategy Flame-outs?

Can investors avoid strategy flame-outs associated with overly enthusiastic backtesting (overfitting)? In his July 2016 paper entitled “Limitations of Quantitative Claims About Trading Strategy Evaluation”, Michael Harris presents two examples that demonstrate a key limitation of trading strategy backtesting:

  1. U.S. stock market trend following.
  2. U.S. stock market mean reversion.

Specifically, he compares performances of such strategies before and after 1997 to illustrate the interaction of backtesting and change in market conditions. Using daily S&P 500 Index returns (excluding dividends) during January 1950 through December 2015, he finds that: Keep Reading

A Few Notes on Odds On: The Making of an Evidence-based Investor

Matt Hall, cofounder and president of Hill Investment Group, introduces his 2016 book, Odds On: The Making of an Evidence-Based Investor, by stating that: “…the evidence-based movement has been studying market data and academic research to identify the groups of stocks and other investments that provide better odds of long-term success. …I’m inviting you to learn how evidence-based investing could change your life…” Based on his experience, he concludes that: Keep Reading

In-sample vs. Out-of-sample Performance of 888 Trading Strategies

Are any trading strategy backtest performance statistics predictive of out-of-sample results? In their March 2016 paper entitled “All that Glitters Is Not Gold: Comparing Backtest and Out-of-Sample Performance on a Large Cohort of Trading Algorithms”, Thomas Wiecki, Andrew Campbell, Justin Lent and Jessica Stauth compare backtest and out-of-sample performance statistics for 888 algorithmic trading strategies. They first screen a larger set of strategies to remove duplicates, outliers and algorithms unlikely to represent real strategies. They next test the selected strategies in-sample (IS) with data that was available to the developers (from 2010 through deployment dates between January and June in 2015). They then test the strategies out-of-sample (OOS) during June 2015 through February 2016. All tests employ minute-by-minute prices for trade entry/exit and include robustly estimated trading frictions. Performance metrics derive from end-of-day positions/prices. Most tests are linear regressions relating individual IS-OOS performance metrics (such as Sharpe ratio). They also examine abilities of several multivariate machine learning techniques to predict performance, ultimately via an an equal-weighted portfolio of the 10 strategies predicted to have the highest OOS Sharpe ratios. Using position and price data for the 888 strategies during the specified IS and OOS periods, plus the total number of backtest days actually employed by each strategy developer, they find that: Keep Reading

Big Picture on Prevalence of Asset Price Series Trends and Reversals

Do asset price series in general reliably exhibit trends and reversals? In his May 2016 paper entitled “Trend, Mean-Reversion or Random Walk? A Statistical Analysis of Price Behavior in Major Markets”, Theo Athanasiadis tests a wide variety of financial market price series for existence of significant trends and reversals. He considers both spot and futures price series in U.S. dollars for 56 major markets: 16 developed equity market indexes; the S&P 500 implied volatility index (VIX); 25 liquid commodities covering all basic sectors; 5 liquid currency exchange rates versus the U.S. dollar; and, 9 liquid government bonds of varying durations. For futures contract returns, he uses the most liquid contracts (typically nearest or next-nearest) and rolls accordingly. He employs three statistical tests of time-series behavior: autocorrelation, variance ratio and positive/negative runs relative to median. He considers weekly, monthly, quarterly and semiannual returns in both univariate and multivariate tests. Using spot and futures price returns at the specified frequencies for all 56 markets as available during January 1999 through March 2016, he finds that: Keep Reading

Live Performance of Alternative Beta Products

Are the backtests provided for alternative beta investment products representative of their future live performance? In their March 2016 paper entitled “Quantifying Backtest Overfitting in Alternative Beta Strategies”, Antti Suhonen, Matthias Lennkh and Fabrice Perez compare the backtested and live performances of alternative beta products offered by investment banks. The strategies underlying these products are formulaic and non-discretionary, designed to extract risk factor/style premiums (such as value, momentum, carry or term) or exploit some financial market anomaly (such as turn-of-the-month or mean reversion). Specifically, they:

  1. Present an overview of alternative beta products offered by investment banks.
  2. Compare backtested and live performance data for these products.
  3. Compare  backtested and live factor exposures four four strategy families (equity value, equity volatility, fixed income term and currency exchange carry).

Using daily returns in U.S. dollars of 215 alternative beta strategies across five asset classes and 11 strategy types offered by 15 investment banks as available during January 1990 through early March 2015, they find that: Keep Reading

Why Smart Beta Funds Will Disappoint?

What happens out-of-sample to stock portfolios with weights derived from extreme in-sample fitting? In their February 2016 paper entitled “Stock Portfolio Design and Backtest Overfitting”, David Bailey, Jonathan Borwein and Marcos Lopez de Prado examine backtest overfitting in the context of designing a stock portfolio/fund. Their test approach is:

  1. Construct split-adjusted, dividend-reinvested price series for all S&P 500 components as of January 22, 2016 with continuous monthly prices during 1991 through 2015 (277 stocks).
  2. Select a target performance profile, including annualized return (6%, 8%, 10%, 12% 15% or 18%) and “shape” of return (principally, steady increase every month).
  3. Apply an optimization program to determine the fixed stock price weights (0.1% increments) that achieve target performance profile in-sample during 1991-2005 (requiring monthly rebalancing of the portfolio to those weights).
  4. Apply these stock price weights during 2006-2015 (again, with monthly portfolio rebalancing) to measure out-of-sample performance.

In initial tests, they allow negative weights (shorting). Because of the risks of shorting, they repeat analyses with a long-only constraint. They note that their in-sample fitting process considers “an inconceivably large set” of possible weights. They use the S&P 500 Total Return Index as a benchmark. Using adjusted monthly prices for the specified stocks from the end of December 1990 through the beginning of January 2016, they find that: Keep Reading

Removing the Upward Bias of In-sample Optimized Sharpe Ratios

How can investors easily estimate the degradation from optimized in-sample Sharpe ratio to out-of-sample expected Sharpe ratio? In their February 2016 paper entitled “Noise Fit, Estimation Error and a Sharpe Information Criterion”, Dirk Paulsen and Jakob Sohl derive a simple correction for the upward bias in an optimized in-sample Sharpe ratio. The upward bias derives from fitting: (1) random noise within the backtest sample; and, (2) peculiarities in the backtest sample that make it less than perfectly representative of the entire (unknowable) series. In other words, even if no predictability exists, fitting noise “discovers” some. And, even if predictability exists, predictability within a backtest sample will likely be different from predictability in the entire series. Based on derivations addressing quantification of these two sources of bias, they conclude that: Keep Reading

Login
Daily Email Updates
Filter Research
  • Research Categories (select one or more)