r/quant

▲ 2 r/quant

I tested the classic S&P 500 reconstitution trade. The mechanism is real, but the easy trade seems dead.

I've been looking at one of the most famous forced-flow anomalies in finance: the S&P 500 inclusion/deletion effect.

The mechanism is well known. When a stock enters the index, passive funds have to buy it. When it gets deleted, passive funds have to sell it. That creates mechanical flow unrelated to fundamentals.

What I wanted to test was not whether the mechanism exists in theory, but whether the simple trade still works in practice.

So I looked at two basic implementations using daily data:

  • buy deletions after the effective date and wait for a rebound
  • test whether additions fade after the effective date

For deletions, I identified 317 historical events, but I could only reconstruct post-event price data for 121 of them using free data. The rest were often delisted, acquired, merged, renamed, bankrupt, or otherwise unavailable.

That missingness is not random, and it's a serious limitation. Many of the names that drop out of the database are exactly the ones most likely to have had extreme post-event behavior : bankruptcies going to zero, acquisitions gapping up. So any statistics I compute are a statement about the surviving 38% of the sample, not the full universe of deletions. They cannot prove the anomaly is dead in general. They can only describe what happened in the subset I was able to reconstruct.

With that caveat, even in this surviving subset there was nothing:

  • average post-effective return: -24.19 bps
  • Sharpe: -0.07
  • win rate: 41.3%
  • timing permutation p-value: 0.208
  • validation layers passed: 0/8

No rebound, no statistical support, no robustness : in the testable sample. Whether the full 317 events tell a different story is an open question I can't answer with free data.

On the addition side, the issue is different. My dataset has effective dates, but not reliable announcement dates. That matters because the canonical inclusion effect mostly happens between announcement and implementation. So a post-effective-date fade test is not really a clean test of the original anomaly.

That test also looked like noise, but I would not treat that as a mechanism kill. It is more a data limitation than a strong conclusion.

My takeaway is this:

The mechanism is still real. Forced index flow still exists. But the naive implementation of the trade : public event, daily data, enter after the effective date, wait for mean reversion: appears to be gone, or at minimum is not detectable in the data I have access to.

That makes sense. Once a flow becomes public, easy to model, and visible in advance, faster participants can arbitrage much of the obvious price impact away before passive money actually executes. Greenwood and Sammon (2023) document this weakening in detail: more anticipation, better liquidity supply, structural changes in the composition of additions and deletions.

So to me this looks like a good example of something important:

A market mechanism can remain real long after the easy trade built on top of it has died. "Structural edge exists" and "you can still monetize it" are not the same statement.

That distinction seems to matter a lot in anomaly research. A lot of "edges" are really just true stories with no remaining implementation value.

I'm curious how others here think about this. Do you see the index effect as mostly competed away now, or just pushed into narrower implementations and data regimes?

reddit.com
u/DuePhotograph6877 — 2 hours ago
▲ 9 r/quant

Perpetual Futures

The crypto market liquidity is drying up, but perpetual futures for commodities, equities etc.. are exploding in popularity and volume.

Have any of you tradfi quants ventured into the world of perps and funding rates? And if so what’s your take on this new type of futures contract?

IMO perps are the most palatable way for retail to get access to high leverage, so as a general product it is here to stay beyond crypto.

reddit.com
u/Prada-me — 8 hours ago
▲ 2 r/quant

Weekly Megathread: Education, Early Career and Hiring/Interview Advice

Attention new and aspiring quants! We get a lot of threads about the simple education stuff (which college? which masters?), early career advice (is this a good first job? who should I apply to?), the hiring process, interviews (what are they like? How should I prepare?), online assignments, and timelines for these things, To try to centralize this info a bit better and cut down on this repetitive content we have these weekly megathreads, posted each Monday.

Previous megathreads can be found here.

Please use this thread for all questions about the above topics. Individual posts outside this thread will likely be removed by mods.

reddit.com
u/AutoModerator — 6 hours ago
▲ 4 r/quant

Seeking advice on fitness functions for Genetic Algorithms

Hi everyone,

Throwing a bottle in the sea here. I’ve been struggling for days trying to find a way to optimize my algo using an evolutionary/genetic approach.

The Problem: My optimization process is prematurely converging. It hits a fitness plateau extremely fast, and the strategy stops optimizing generation after generation. It feels like the engine is getting stuck in a local optimum very early in the training loop.

What I've tried so far:

  • Evaluating and scoring the generations using the Van Tharp method (System Quality Number / SQN).
  • Building my own custom calculus and penalty functions to balance win rate, drawdown, and total profit.
  • Tuning basic hyper-parameters like mutation and crossover rates.

Everything I try seems to lack robustness needed to actually push the algorithm past that initial plateau and find a solid strategy.

My Questions for the community:

  1. What fitness functions or mathematical metrics do you guys rely on to properly evaluate a strategy generation over generation?
  2. Are you using multi-objective optimization (like NSGA-II) to balance returns and drawdowns, or do you stick to a single scalar fitness metric?
  3. What methods do you use to prevent your optimization from hitting a plateau so fast?

Any pointers, papers, or advice would be massively appreciated. Thank you!

reddit.com
u/Signal_Control_9366 — 22 hours ago
▲ 0 r/quant

Interesting experiment arbitraging favorite-longshot bias on polymarket/kalshi

I developed an application for automating and managing trades on polymarket and kalshi. About 6 months ago I started looking into whether the favorite-longshot bias from horse racing and sports betting shows up in prediction markets. It does. I pulled all 59,000 resolved binary markets off their API and ran a calibration study. Markets priced in the 40-50% range resolve Yes only about 22% of the time. Sports and games are the strongest categories.

Why I don't think it's noise

I ran Benjamini-Hochberg FDR correction at q=0.05 across 537 calibration cells. 78 survived. You'd expect about 27 by chance, so that's roughly 2.9x the false discovery rate.

Before writing any strategy code I set hard kill gates. If I couldn't find at least 8pp of miscalibration in a tradeable category, I'd stop. It passed. I also caught a Simpson's paradox artifact where the bias looked like it was growing over time but it was really just Sports becoming a bigger share of the market mix.

The part I think matters most: I tried expanding to Kalshi using 7.68 million markets and the kill gate failed. The apparent signal was a bucket-assignment artifact at the 50-cent line. I killed that track. The gates aren't decoration.

Backtest (in-sample, take it for what it's worth)

About 4,850 signals, ~150 trades after filtering. 64.6% win rate, Sharpe 1.21. Simulation from $3K to ~$8K, max drawdown 25.1%. Average hold around 20 days.

Where I'm at now

Paper trading went live this week. 12 positions open, $4K of $10K deployed. First resolutions in a week or two. I'll post an update with out-of-sample numbers when I have them, pass or fail.

I'm not sharing the specific cell map, classification system, or pipeline logic. That's the IP. But if you pulled the data yourself you'd confirm the FLB exists. Knowing it exists and knowing exactly where to trade it are different things.

What I'm curious about

  • Does the methodology hold up or am I missing something obvious?
  • Anyone ever commercialized capacity-constrained trading IP? The ceiling here is ~$50-100K before you move the markets you're trading. I'm deploying my own capital but I've been wondering if the methodology has standalone value as education or research.
  • If you trade prediction markets, does the FLB claim match your experience?

Happy to talk process and stats. Not sharing the cell map or pipeline specifics.

reddit.com
u/AlSikandar — 5 hours ago
▲ 0 r/quant

Built a systematic FLB strategy on Polymarket - 59K markets analyzed, BH FDR correction, now in paper trading. Anyone else trading structural bias here?

Introduction

Hey all. Long-time lurker, first real post. I want to share a project I have been working on and get some honest feedback — both on the methodology and on whether the IP has commercial legs.

The short version: I built a systematic trading system that exploits the favorite-longshot bias on Polymarket (CFTC-regulated prediction market). The core finding is that binary markets in the 30-60% price range are overpriced by 12-24 percentage points, and this holds up after Benjamini-Hochberg FDR correction across 59K resolved markets.

Background

Polymarket binary contracts pay $1 if an event happens, $0 if it doesn't. A contract at $0.45 implies 45% probability. If I can show the true resolution rate for that class of markets is much lower than 45%, there is a structural edge.

I collected all resolved binary markets from Polymarket's API — about 59,000 markets total. Ran a calibration study: for markets priced at X% at various time horizons before resolution, what fraction actually resolved Yes?

The favorite-longshot bias showed up clearly. Markets in the 40-50% range resolve Yes only about 22% of the time. Sports and games categories are the strongest. The bias is driven by retail traders overpaying for exciting "Yes" on longshot outcomes — the same psychological pattern that has been documented in horse racing and sports betting for decades.

Why I think this is not just data mining

This is where I expect the most pushback, so let me get ahead of it:

1. Statistical correction. I used Benjamini-Hochberg FDR correction at q=0.05 across 537 calibration cells (category x horizon x price bucket). 78 cells survived. If this were noise, you would expect roughly 27 cells to survive — getting 78 is a 2.9x multiple over the false discovery rate.

2. Pre-registered kill gates. Before writing any strategy code, I set explicit pass/fail criteria. The Phase 0 kill gate required >8pp miscalibration in at least one tradeable category. If it had failed, I would have stopped the project entirely and published the calibration study as a portfolio piece. It passed with STRONG_PASS.

3. Simpson's paradox testing. The apparent intensification of bias over time (13pp at 7 days, 24pp at 30 days) turned out to be a composition artifact — Sports grew from 7% to 26% of the market mix over the dataset period, and Sports has the strongest signal. Within categories, the bias is stable across time. I caught this with volume and category controls.

4. A kill gate that actually fired. I expanded the analysis to Kalshi (another CFTC-regulated prediction exchange) using an independent dataset of 7.68M markets. The kill gate failed — only 2 of 10 required BH cells survived, and a boundary sensitivity check revealed the apparent signal was a bucket-assignment artifact at the 50-cent line. I paused the Kalshi track based on this result. I am mentioning this specifically because it demonstrates the gates are not decoration — they fire when the signal is not there.

Backtest results (in-sample, all the usual caveats apply)

  • 4,851 signals generated, ~150 trades executed through a multi-gate filtering pipeline
  • 64.6% win rate, 23% ROI, Sharpe 1.21
  • Post-capacity-expansion simulation: $3K starting capital to ~$8K, CAGR 63.7%, Sharpe 1.07, max drawdown 25.1%
  • Average hold period: ~20 days

I am not going to pretend these are out-of-sample numbers. They are not. That is what the forward validation phase is for.

Where things stand right now

Forward validation (paper trading with live market data) went live this week. 12 open positions, about $4K of $10K budget deployed. First resolutions expected within a week or two. The system runs on 15-minute cycles with 227 automated tests and a full CI pipeline.

I do not have out-of-sample results yet. I will share an update on how forward validation went — whether it passed or failed.

What I am deliberately not sharing

I am not publishing the exact cell map (which category/horizon/bucket combinations are tradeable), the structural classification system I built for market taxonomy, or the signal pipeline gating logic. These are the core IP.

I am sharing enough of the methodology for you to evaluate whether it is rigorous, but not enough to replicate the strategy without doing the work yourself. If you ran the same calibration study on the public Gamma API data, you would confirm the FLB exists — but knowing it exists and knowing which specific cells to trade are very different things.

The commercialization question

This is the part I genuinely want community input on.

The capacity ceiling for this strategy is roughly $50-100K deployed capital before you start moving markets. That is a fundamental constraint — it means selling execution (fund, copy-trading) actively degrades the edge. But selling intelligence (methodology, data, education) does not.

The paths I am considering:

  • Education: A course teaching calibration methodology and structural bias analysis for prediction markets. The techniques generalize to any prediction market, not just Polymarket.
  • Research/data licensing: The 59K-market dataset with calibration results, licensed to platforms or research teams.
  • Signals-as-a-service: Heavily capped (5-10 seats max) and only after 100+ forward-validated trades with confirmed edge. This is the most obvious path but also the one that erodes the moat fastest.

I have a slide deck and a detailed proposal document ready if anyone wants to discuss specifics — happy to share in DMs with anyone who has relevant experience.

My questions for this community

  1. Does the methodology sound rigorous, or am I fooling myself? What holes do you see? I have been deep in this for months and could be missing something obvious.
  2. Has anyone here commercialized quantitative trading IP? What worked and what did not? I am especially interested in hearing from people who navigated the "edge is real but capacity-constrained" problem.
  3. If you were shopping a slide deck for this kind of project, who would you approach? Prediction market platforms? Quant funds doing alt-data? Fintech accelerators? Educational platforms?
  4. Any prediction market traders here who can gut-check the FLB claim from their own experience? Curious if this matches what you have seen in practice.

Happy to answer methodology questions. I will not share the specific cell map or signal pipeline details, but anything about the process, statistical approach, or commercialization thinking is fair game.

reddit.com
u/AlSikandar — 8 hours ago
▲ 0 r/quant

How do hedge funds monitor web data?

Web data is importantly as an alternative source as told by multiple quants. So this needs to be scraped.

But there should be a cadence for scraping and also the usual data scrapers break after sometime.

How do funds monitor/track the websites continuously and get the alternative data?

I am thinking of creating a prompt based monitoring system. Will that be useful? Will that be considered as alt data?

reddit.com
u/scienceofdeduction09 — 8 hours ago
ML lookahead bias profitable in real?
▲ 0 r/quant

ML lookahead bias profitable in real?

Well, basically i developed a ML using this bias, i didn't knowed about that until i reviewed again a month later. In short, i passed the strategy to a paper account in alpaca using a VPS at the begining of the year. 3 months later here are the results.

The strategy basically make the decition to enter or not at the market in 3x ETFs. What you think about that? I know that the best is that i make the backtest with the final model to all the data, but i want to know if you are converted a LookAhead bias ML in profitable even when it may be considered a sin in ML.

u/Anonimo1sdfg — 16 hours ago
Week