u/Dry-Jello194

▲ 1 r/arbitragebetting+1 crossposts

Betzy Pricing Update + VIP Channel Launch

We’ve officially updated the Betzy pricing structure and added a dedicated VIP channel for elite AI signals.

New Pricing

🔥 Weekly (7 Days)
150 Stars

💎 Monthly (30 Days)
400 Stars

Annual (365 Days)
3500 Stars

🔔 DM Alerts (30 Days)
150 Stars

🏆 VIP Channel (30 Days)
500 Stars

What’s inside the VIP channel?

The VIP channel is no longer just “more predictions.”

It now focuses on Elite Signals only — predictions that pass:

  • Main ML model threshold
  • Secondary V2 validation model
  • Pressure + momentum gates
  • Confidence filtering
  • Hybrid confluence checks

Current internal testing shows elite signals significantly outperform standard predictions in hit rate and ROI, while keeping volume intentionally low.

The idea is simple:

Less noise. Higher quality.

Why the change?

Over the last weeks we upgraded the backend heavily:

  • Hybrid ML architecture
  • Dual-model validation
  • Real-time calibration monitoring
  • Drift detection
  • Premium confidence gating
  • Reliability tracking (ECE / KS / decile monitoring)

Instead of flooding channels with quantity, the system now prioritizes precision.

Important

The public/free experience remains active.
VIP is designed for users who specifically want the highest-confidence subset of predictions.

More backend updates and transparency reports coming soon.

reddit.com
u/Dry-Jello194 — 3 days ago

Betzy ML Update — Hybrid AI System, Elite Signals & Real Calibration Monitoring

Betzy ML Update — Hybrid AI System, Elite Signals & Real Calibration Monitoring

After weeks of replay testing on 1,200+ historical predictions, Betzy’s new hybrid ML architecture is now live.

What changed?

  • V1 remains the main prediction engine
  • V2 now works as a shadow validator + confidence filter
  • “💎 Elite Signals” only trigger when BOTH models strongly agree
  • Added real-time calibration monitoring, drift detection and premium safety guards

Current projected performance

FREE SIGNALS

  • Higher volume
  • ~70–84% hit range depending on market conditions

💎 ELITE SIGNALS

  • Lower volume
  • Much higher confidence
  • Requires V1 + V2 confluence
  • Historical replay projections reached up to ~96% hit rate in selected samples

One of the biggest discoveries during testing

The model itself wasn’t actually failing.

The real issue was calibration drift after a scaler fix changed the raw score distribution.
The AI rankings were still good, but the probability layer started overstating confidence in some score ranges.

Instead of panic-retraining the entire model, Betzy now monitors:

  • ECE drift
  • KS distribution drift
  • confidence inflation
  • premium hit-rate degradation

This means the system can detect when probabilities become unreliable before the prediction quality fully collapses.


Important finding

Even during calibration instability:

> Higher-confidence predictions still outperform lower-confidence predictions consistently.

That means the predictive edge still exists.
The issue was mostly about probability mapping — not ranking quality.


Last 11 days (estimated ROI)

Date ROI
2026-05-08 +124.6%
2026-05-07 +35.7%
2026-05-06 +10.6%
2026-05-05 -8.1%
2026-05-04 +22.5%
2026-05-03 +12.5%
2026-05-02 +22.9%
2026-05-01 -3.4%
2026-04-30 +12.1%
2026-04-29 +30.1%
2026-04-28 +26.7%

Yes, there were losing days too.
This is betting — not magic.


Next steps

  • collect more post-fix data
  • rolling-window calibration refit
  • possible context-specific calibration by game state
  • survivor-bias reduction using snapshot archives

Betzy is slowly evolving from a simple prediction bot into a fully monitored probabilistic decision engine.

reddit.com
u/Dry-Jello194 — 3 days ago
▲ 4 r/PredictionMarkets+2 crossposts

11-Day Betting Model Results — Early Observations from a Live ML Football Betting System

I’ve been running a live in-play football betting model for the last 11 days and wanted to share the raw performance numbers plus some thoughts on what’s actually happening under the hood.

Important context before people scream “fake ROI”:

  • These are estimated odds simulations, not bookmaker closing odds
  • Market type is mainly live goal / next-goal style signals
  • System is ML-driven with:
    • historical similarity search
    • contextual calibration
    • pressure/momentum features
    • gating/filter logic
  • Results include variance-heavy high-volume days
  • This is still effectively a live beta / shadow deployment phase

Daily Results

Date Staked Profit ROI W/L/P
2026-05-08 $1231 +$303.51 24.6% 18/6/0
2026-05-07 $641 +$228.79 35.7% 10/3/0
2026-05-06 $1120 +$118.63 10.6% 15/8/0
2026-05-05 $678 -$55.09 -8.1% 8/7/0
2026-05-04 $1253 +$281.60 22.5% 14/6/0
2026-05-03 $5223 +$654.51 12.5% 59/30/1
2026-05-02 $4635 +$1060.84 22.9% 60/22/0
2026-05-01 $1210 -$41.60 -3.4% 17/13/0
2026-04-30 $520 +$62.78 12.1% 7/5/0
2026-04-29 $1722 +$518.69 30.1% 25/9/0
2026-04-28 $500 +$133.41 26.7% 8/1/0

Aggregate Snapshot

Total Staked

$18,733

Total Profit

+$3,266

Overall ROI

+17.4%

Win/Loss

241 Wins / 110 Losses / 1 Push

Raw win rate:

  • ~68.7%

The Interesting Part

What surprised me most wasn’t the ROI itself.

It was how the model behaves across different confidence regimes.

I recently added a secondary validation layer (“V2”) on top of the original production model (“V1”).

Initially the new model looked worse in backtests:

  • lower recall
  • lower volume
  • fewer bets
  • worse aggregate ROI

But after replaying historical predictions with proper temporal filtering (no future leakage), something interesting appeared:

The V2 model was not actually “worse.”
It was:

  • far more selective
  • heavily confidence-skewed
  • much stronger in the extreme high-confidence tail

So instead of replacing the original model, I turned it into a premium confluence filter:

  • V1 = broad signal generator
  • V2 = “are we REALLY sure?” validator

That dramatically reduced volume but massively increased projected precision.

Biggest Lessons So Far

1. Calibration matters more than raw model accuracy

A badly calibrated model with decent ranking can destroy ROI.

Fixing probability calibration improved live behavior more than retraining the RF itself.

2. Thresholds are everything

A model that looks mediocre at 0.60 can become elite at 0.80.

Most people optimize AUC and forget:

>

3. Ensemble disagreement is insanely informative

Some of the best signals came from:

  • V1 high confidence
  • V2 ALSO high confidence
  • strong historical neighborhood density
  • no contradictory gate flags

That overlap region appears much stronger than either model alone.

4. Volume and edge fight each other

The system can easily generate more bets.

But every time I loosen thresholds:

  • ROI compresses
  • variance explodes
  • false positives multiply

Right now the architecture is drifting toward:

>

Which honestly may be the only survivable long-term approach.

Caveats

Before anyone asks:

  • No, I do NOT think 17% long-term ROI is sustainable.
  • Yes, survivorship bias is possible.
  • Yes, estimated odds inflate reality.
  • Yes, bookmaker limits/slippage matter.
  • Yes, variance can nuke 11-day samples instantly.

I’m treating this as:

>

Current Focus

Next steps are:

  • larger replay datasets
  • removing selection bias from training
  • live calibration monitoring
  • confidence drift analysis
  • premium-tier signal validation

If people are interested, I can also post:

  • calibration charts
  • confidence bucket performance
  • replay methodology
  • false positive analysis
  • live vs replay divergence
  • feature engineering details
  • temporal leakage prevention setup
reddit.com
u/Dry-Jello194 — 5 days ago
▲ 2 r/arbitragebetting+1 crossposts

https://preview.redd.it/k7zzgzl2lqyg1.png?width=3814&format=png&auto=webp&s=91f42534c08ae8034c679a5d8146231b493554fb

[RESULTS] AI-powered goal predictions went 18/19 today — 95% accuracy across 10 leagues

Been testing BETZY (an AI live-match prediction engine) for a while now. Sharing today's full log — screenshot attached so you can verify every single call.

TODAY'S NUMBERS (May 2, 2026)

  • Signals fired: 21
  • Completed: 19
  • Success: 18
  • Expired (no goal): 1
  • Still pending: 2
  • Accuracy (completed): 94.7%

LEAGUES COVERED TODAY

EPL, La Liga, Bundesliga, Serie A, Ligue 1, Championship, Eredivisie, Thai League 1, Czech First League, 2. Bundesliga, Trendyol 1. Lig

NOTABLE CALLS

  • Sheffield United (66', score 1-1, COUNTER) → GOAL — 66% confidence
  • Levante UD (54', score 1-1, COUNTER) → GOAL — 62% confidence
  • VfL Bochum (67', score 1-0, COUNTER) → GOAL — 62% confidence
  • Ayutthaya United (68', score 2-0, COUNTER) → GOAL — 58% confidence
  • Nantes (57', score 2-0, NORMAL) → GOAL — 49% confidence
  • Dukla Praha (9', score 0-0, NORMAL) → GOAL — 46% confidence
  • SK Sigma Olomouc (56', score 0-2, NORMAL) → GOAL — 41% confidence

THE ONE THAT DIDN'T HIT

  • Millwall (51', score 2-0, NORMAL) → no goal, match ended — EXPIRED

HOW IT WORKS

The system watches live match data in real time and fires a signal when it detects a high-probability goal window — based on momentum, minute, scoreline, and historical league patterns. Predictions are flagged as NORMAL (with the game's flow) or COUNTER (against the current trend, usually higher value).

Full history log is in the screenshot above. Happy to answer questions.

Not financial advice. Track records can change. Always do your own research.

#soccerbetting #football #ai #predictions #betzy

reddit.com
u/Dry-Jello194 — 12 days ago
▲ 2 r/arbitragebetting+1 crossposts

I've been building a football analytics bot for the past few months and wanted to share some of the genuinely surprising things I learned from the data. This isn't a tips post — it's more about what the numbers revealed when I stopped trusting intuition.


What the system does

It polls live match data every ~90 seconds — shots, xG, momentum, pressure, big chances, possession — and decides whether to fire a "goal incoming" prediction within a configurable time window. Predictions are auto-resolved when the window expires.

961 resolved predictions across 70+ leagues so far.


What I expected vs what the data actually showed

🔴 Shots on target is a trap signal

Higher SOT in the prediction window was inversely correlated with goals. Teams piling up shots without converting are more likely to run out of steam than score. I removed SOT as a positive gate entirely after confirming this across 900+ predictions.

🔴 Big chances mislead you the same way

Same story. High big chance count = overheating attack, not imminent goal. > Success avg: 1.43 big chances | Failure avg: 1.53 big chances

Counterintuitive but statistically consistent.

🟢 Pressure flips sign depending on the half

This was the biggest finding. Total pressure is a ceiling signal in the first half (above 110 = overheating, more likely to fail) and a floor signal in the second half (below 90 = cold match, less likely to score).

Using a single global pressure threshold was actively hurting accuracy. Splitting by half gave a meaningful improvement.

🟢 Minute is the single most important feature

Not xG. Not pressure. Not momentum. When you feed all signals into a Random Forest, minute dominates at 14.2% importance. The system now uses time-weighted gates rather than flat thresholds.

🔴 Score state matters more than most models assume

Score State Prediction Accuracy Notes
high (5+ goals) 42.1% Nearly random — block it
2-1 55.6% Risky
0-0 70.7% Solid baseline
2-0 74.5% Open play, both teams active
1-1 76.9% Best state for goals

Blocking predictions in high and 2-1 states alone added ~1.6pp to overall accuracy.


Current performance

Overall: 70.8% across 961 resolved predictions

Minute Bucket Accuracy Notes
0–15 69.1% Stable
15–30 60.7% No reliable discriminator found
30–45 35.8% ⚠️ Research bucket — all signals invert here
45–60 75.7% ✅ Best window
60–75 71.1% Solid
75–90 Blocked at min 80 Was 21.4% — essentially noise

The 30–45 bucket is still unsolved. Every signal inverts: higher pressure, higher xG, higher SOT all correlate with failure in this window. Open to ideas if anyone has seen similar patterns.


How predictions are delivered

Via a Telegram bot to subscribed channels. The system runs 24/7 with a self-learning calibration engine that adjusts thresholds based on recent outcomes — so it gradually adapts to each league's scoring patterns.

Happy to answer questions about the methodology, especially the xG modelling or the pressure asymmetry finding — drop them below.


Not financial advice. Predictions are probabilistic — 70.8% means roughly 3 in 10 will be wrong.

reddit.com
u/Dry-Jello194 — 12 days ago