Agent Evolution
Baseline has two main variants:
- Blueprint: offline-solved strategy library (no live solving).
- Baseline AI: live depth-limited solving on top of a biased blueprint, then reach subgame solving.
Benchmark setup
- Game: Heads-Up No-Limit Hold’em
- Stacks: 200 big blinds, no rake
- Opponent: Slumbot (API version noted per run)
- Metric: BB/100 = big blinds won per 100 hands
Results against Slumbot
| Agent |
Year |
Hands |
Result (BB/100) |
| DeepStack (DeepMind) reimplementation |
2017 |
150,000 |
-6.3 |
| ReBeL (Meta) |
2020 |
- |
+4.5 |
| Blueprint v1 |
2023 |
100,000 |
+3.4 |
| GTO Wizard |
2023 |
150,000 |
+19.4 |
| Blueprint v2 |
2024 |
20,000 |
+10.9 |
| Baseline AI |
2024 |
10,000 |
+18.1 |
Changelog
Blueprint v1 → v2
- Potential-Aware Cards Abstraction
- Smarter betting tree, using Pseudo-Harmonic Mapping not only to translate actions but also to make the betting tree itself less exploitable.
- (Proprietary) Improved betting tree, based on an analysis of blueprint, instead of theoretical harmonic thresholds.
- (Proprietary) Smarter action-translation, also based on an analysis of blueprint.
Blueprint v2 → Baseline AI
- Depth-limited solving: Solves the subgame in real-time instead of only following a stored policy.
- Biased blueprint at the depth limit: The blueprint acts as a prior / continuation model so live-solving doesn’t hinge on a single assumed continuation.
- Reach solving: Avoids overfitting to a fixed opponent range by pricing “range switching” (playing irrationally with other hands), keeping the strategy non-exploitable.
- Results: In a short stack setup (HU 25 BB stack), Baseline AI defeats current blueprint by approximately 7.7 BB/100, which corresponds to the difference we observed in Slumbot matches.