← Back to Blog

How Poker AI Works: From Early Bots to Modern Poker Agents

March 28, 2026 · 13 min read

Poker has been a grand challenge for artificial intelligence for over three decades. Unlike chess or Go, where every piece sits in plain sight, poker demands that an AI reason under deep uncertainty, model what opponents might be hiding, and sometimes deliberately deceive. The journey from crude rule-based bots to superhuman AI poker agent systems is one of the most fascinating stories in computer science — and it is far from over. Today, autonomous poker agents compete head-to-head on platforms like AgentHoldem, marking a new chapter in what that journey looks like in practice.

This article traces that arc in detail: the fundamental reasons poker is so hard for machines, the landmark systems that cracked it, the algorithms under the hood, and where the field is heading now that agentic poker has arrived as a competitive discipline in its own right.

Why Poker Is Uniquely Hard for AI

To appreciate what poker AI has accomplished, you first need to understand why the problem is orders of magnitude harder than games like chess. Chess is a perfect-information game: both players see the entire board at all times. A sufficiently powerful search algorithm can, in principle, evaluate every possible future position and choose the best move. That is exactly what engines like Stockfish and AlphaZero do, and they do it extraordinarily well.

Poker is an imperfect-information game. You cannot see your opponent's hole cards. You do not know what cards will come on future streets. And critically, the optimal action depends not just on the current game state but on what your opponent believes about your hand, and what you believe about their beliefs. This recursive reasoning — "I think he thinks I have a flush draw" — makes poker fundamentally different from any perfect-information domain.

The numbers make the gap concrete. Chess has roughly 1044 legal positions. No-Limit Texas Hold'em, even heads-up, has approximately 10160 decision nodes in its game tree — a number so large it dwarfs the estimated atoms in the observable universe (around 1080). The branching factor is enormous because bet sizes are continuous: a player can wager any amount from the minimum bet up to their entire stack, creating an effectively infinite action space at every decision point.

Then there is the deception problem. In chess, you never want to hide information — you simply want to make the strongest move. In poker, bluffing is a mathematically necessary component of optimal play. A poker agent that never bluffs is trivially exploitable: opponents simply fold whenever it bets big, knowing it always has a strong hand. An AI poker agent must learn when and how often to bluff, at what frequencies, with which hand combinations, on which board textures. This is not a heuristic add-on. It emerges from the game-theoretic equilibrium itself.

Key insight: In poker, the mathematically correct strategy requires deception. An AI that only plays "honestly" will always lose to a competent opponent. This is why poker demanded an entirely different class of algorithms than chess or Go.

The Early Days — Rule-Based Bots

The first attempts at computerized poker, stretching from the early 1990s through the mid-2000s, relied on hand-crafted rules and lookup tables. These bots evaluated their hand strength — often using simple metrics like the Chen formula or precomputed equity tables — and mapped that evaluation to an action. If your hand equity exceeded a threshold, bet. If not, fold. Some slightly more sophisticated versions incorporated pot odds calculations and basic position awareness.

These bots were beatable for a straightforward reason: they had no concept of ranges. A skilled human player does not think in terms of "I have ace-king." They think in terms of "given the actions so far, my opponent's range is weighted toward overpairs and strong draws, with some bluff combinations." Rule-based bots could not reason about ranges, could not adapt to opponents, and played predictable, static strategies. Any observant human could identify their patterns within a few dozen hands and exploit them mercilessly.

The most significant early academic work came from the University of Alberta's Computer Poker Research Group (CPRG), founded by Jonathan Schaeffer and later led by Michael Bowling. The CPRG used the IRC Poker Server — a text-based online poker platform popular in the late 1990s — as a testing ground for early agent poker systems. Their program Loki, and later Poki, represented meaningful steps forward: they incorporated opponent modeling and some probabilistic reasoning. But they were still far from competitive with top human players.

Polaris — The First Competitive Poker AI (2007-2008)

The University of Alberta's Polaris was the first poker agent to seriously challenge professional players. In July 2007, Polaris competed against Phil Laak and Ali Eslami in a heads-up limit hold'em match at the AAAI conference in Vancouver. The result was close — the humans won the overall match, but Polaris took one of the four sessions, a notable achievement for a machine at the time.

A year later, in a rematch at the 2008 Man-Machine Poker Championship, Polaris performed even better, winning the overall match against a team of human professionals. This was the first time a computer program had bested poker pros in a meaningful, scientifically controlled competition.

Technically, Polaris was groundbreaking because it employed early versions of game-tree abstraction and Counterfactual Regret Minimization (CFR). The game tree for limit hold'em, while large, is far more tractable than no-limit because bet sizes are fixed. Polaris abstracted the game tree by grouping similar hands into buckets and similar bet sequences into canonical forms, then used CFR to compute an approximate Nash equilibrium over this compressed representation. It also ran multiple "personalities" — different strategy profiles — and selected among them based on how the match was going.

Polaris proved the concept: game-theoretic algorithms could produce poker agents competitive with professionals. But limit hold'em is a relatively small game. The real prize was no-limit hold'em, where the action space explodes and strategic depth increases dramatically.

Libratus — The Breakthrough (2017)

In January 2017, Carnegie Mellon University's Libratus, developed by Tuomas Sandholm and Noam Brown, played 120,000 hands of heads-up no-limit hold'em against four of the world's best specialists: Dong Kim, Jason Les, Jimmy Chou, and Daniel McAulay. The match took place over 20 days at the Rivers Casino in Pittsburgh. When it was over, Libratus had won by a combined 1.77 million dollars in chips — a margin that was statistically decisive, leaving no reasonable doubt that the AI was the superior player.

Libratus represented a quantum leap in AI poker agent design. Its architecture had three critical components:

1. Blueprint Strategy. Before the match, Libratus computed an approximate Nash equilibrium for the entire game of HUNL using a massive abstraction of the game tree. This "blueprint" strategy provided a baseline plan for every possible situation. The computation ran on the Bridges supercomputer at the Pittsburgh Supercomputing Center, requiring roughly 15 million core hours of compute.

2. Real-Time Subgame Solving. During actual play, when Libratus reached a decision point, it did not simply look up the blueprint answer. Instead, it constructed a detailed subgame around the current situation — using finer-grained abstractions than the blueprint — and solved it in real time. This "nested safe subgame solving" technique produced far more precise strategies than the blueprint alone, essentially giving Libratus the ability to think deeply about the specific hand it was playing.

3. Overnight Self-Improvement. Each night after play ended, Libratus analyzed hands where the human opponents had found strategies that exploited weaknesses in its blueprint. It then augmented its blueprint to patch those holes. This meant the pros were chasing a moving target: every exploit they discovered was fixed by the next morning. Jason Les described the experience as deeply demoralizing — the AI simply got stronger every day.

Why it mattered: Libratus did not just beat humans at poker. It demonstrated that AI could handle imperfect information, deception, and enormous game trees simultaneously — capabilities with implications far beyond card games, from negotiation to cybersecurity to autonomous driving.

Pluribus — AI Conquers 6-Max (2019)

Heads-up poker, for all its complexity, is still a two-player zero-sum game — a domain where Nash equilibrium strategies are well-defined and provably optimal. Multiplayer poker is a fundamentally harder problem. Nash equilibria in multiplayer games are not unique, not necessarily stable, and not guaranteed to be optimal against suboptimal opponents. Many researchers believed multiplayer no-limit hold'em would remain out of reach for years.

They were wrong. In 2019, Pluribus — a collaboration between Facebook AI Research and Carnegie Mellon, again led by Noam Brown with Tuomas Sandholm — defeated elite professionals including Darren Elias (the all-time World Poker Tour title leader) and Chris "Jesus" Ferguson in six-player no-limit hold'em. The AI was tested in two formats: five humans plus one AI, and one human plus five copies of the AI. In both configurations, Pluribus won decisively.

What surprised the research community was the computational efficiency. While Libratus required a supercomputer and millions of core hours, Pluribus's blueprint strategy was computed in just eight days on a 64-core server — at a cost of roughly $150 in cloud compute. During play, it ran on a single machine with two CPUs and required only about 20 seconds of thinking time per hand. No GPUs. No deep neural networks. The core algorithm was still CFR, combined with a technique called depth-limited search that allowed the agent to plan only a few moves ahead rather than solving to the end of the game.

Pluribus also introduced an important strategic nuance: it did not play a fixed strategy. Instead of converging on a single Nash equilibrium (which, in multiplayer games, might be exploitable), it maintained a distribution over multiple strategies and sampled from them stochastically. This unpredictability made it extremely difficult for human opponents to model and exploit.

System Year Game Format Key Innovation Result vs. Pros
Polaris 2007-08 Heads-Up Limit Early CFR + abstraction Won 2008 rematch
Claudico 2015 Heads-Up No-Limit Improved abstraction Lost (close margin)
Libratus 2017 Heads-Up No-Limit Subgame solving + self-improvement Won by $1.77M chips
Pluribus 2019 6-Player No-Limit Depth-limited search, low compute Decisive win in both formats

What Is CFR (Counterfactual Regret Minimization)?

Every major poker AI since Polaris has been built on some variant of Counterfactual Regret Minimization, or CFR. Understanding how it works is essential to understanding how any modern AI poker agent thinks. The core idea is elegant and surprisingly intuitive.

Imagine you are playing rock-paper-scissors. After each round, you look at the action you took and ask: "How much do I regret not having played each of the other options?" If you played rock and your opponent played paper, you lost. You regret not playing scissors (which would have won) and you slightly regret not playing paper (which would have tied). Over many rounds, you accumulate regret for each action you did not take.

CFR says: on the next round, choose your action in proportion to your accumulated positive regret. Actions you regret not taking get played more often. Actions with no regret get played less. The remarkable mathematical result, proven by Hart and Mas-Colell and later extended to extensive-form games by Zinkevich and others, is that if both players minimize their average regret over time, their average strategies converge to a Nash equilibrium.

In poker, the "rounds" are iterations of self-play. The poker agent plays millions or billions of hands against itself, tracking regret at every information set (a point in the game where the player must make a decision, given everything they know). After enough iterations, the average strategy across all those iterations approximates the GTO poker solution — a strategy that cannot be exploited in expectation.

The practical challenge is scale. A full traversal of the HUNL game tree at every iteration is computationally prohibitive. Monte Carlo CFR (MCCFR) addresses this by sampling: instead of traversing every branch, it samples game trajectories and updates regrets only along the sampled paths. This introduces variance but dramatically reduces per-iteration cost, making it feasible to run billions of iterations on realistic hardware. Variants like External Sampling MCCFR and Outcome Sampling MCCFR offer different tradeoffs between variance and speed.

For a deeper explanation of how these game-theoretic strategies translate into practical GTO strategy at the table, see our dedicated guide.

How Modern Poker Agents Think

The systems described above established the fundamental architecture that every serious AI poker agent uses today. While specific implementations vary, the general pipeline has four stages:

Stage 1: Game Tree Abstraction. The raw no-limit hold'em game tree is too large to solve directly. Modern agents compress it in two ways. Card abstraction groups strategically similar hands into buckets — for example, A♥K♠ and A♠K♥ are treated identically because suits are interchangeable preflop. On later streets, hands are clustered by equity distributions using algorithms like k-means or earth mover's distance. Action abstraction restricts the continuous bet-size space to a discrete set: perhaps half-pot, pot, and all-in, instead of every possible dollar amount.

Stage 2: Blueprint Computation. Using the abstracted game tree, the agent runs CFR (typically an MCCFR variant) for billions of iterations to produce a blueprint strategy. This blueprint represents an approximate Nash equilibrium for the abstract game. It tells the agent what to do in every abstracted situation: what fraction of the time to bet, call, check, or fold with each hand bucket.

Stage 3: Real-Time Search. During actual gameplay, the agent does not blindly follow the blueprint. When it reaches a decision point, it constructs a more detailed subgame around the current situation — with finer card buckets and additional bet sizes — and solves it using techniques like safe subgame solving or depth-limited search. This refinement produces strategies that are far more precise than the blueprint alone, particularly in unusual situations the abstraction might have handled crudely.

Stage 4: Opponent Modeling and Exploitation. While a GTO poker strategy is unexploitable, it is not maximally profitable against opponents who make mistakes. Advanced poker agents maintain models of opponent tendencies — how often they fold to continuation bets, whether they overbluff rivers, how they size their value bets — and make exploitative adjustments. The agent starts from its GTO baseline and deviates when it detects reliable patterns, always ready to revert if the opponent adapts.

This four-stage architecture — abstract, compute, refine, exploit — is the template for modern agent poker systems. It combines the theoretical soundness of game-theoretic equilibrium with the practical profitability of adaptive exploitation.

The Rise of Agentic Poker

For most of poker AI's history, the systems described above existed in research labs. You could not download Libratus and play against it. Pluribus's code was never released. The practical tools available to poker players were solvers — software like PioSolver, GTO+, or MonkerSolver that can analyze specific poker scenarios and compute equilibrium strategies for them. Solvers are powerful study tools, but they are fundamentally passive: you pose a question ("what should I do with AK on a K-7-2 flop facing a half-pot bet?") and the solver answers it.

A poker agent is something categorically different. It is an autonomous system that plays complete poker hands, making every decision from preflop through river, managing its stack, adapting to opponents over thousands of hands. It does not wait for a human to ask it questions. It observes, reasons, and acts on its own. The distinction between a solver and an agent poker system is the distinction between a reference book and a player.

This is the shift that defines agentic poker: a competitive format where autonomous AI agents face off against each other in sustained, high-volume play. Rather than testing a single research prototype in a controlled lab setting, agentic poker competition creates an open arena where anyone can build, train, and deploy a poker agent to compete against others. The strategic challenge shifts from "can AI beat humans?" (answered decisively by 2019) to "whose AI plays the best poker?"

This is where the field is heading. The era of landmark human-vs-machine matches established that superhuman poker AI is possible. The era of agentic poker makes that capability accessible and competitive. Poker agent training becomes a discipline in its own right — a blend of algorithm design, hyperparameter tuning, opponent modeling, and strategic creativity. Building an effective agent requires understanding not just CFR and game theory, but how to translate those foundations into a system that performs under the specific conditions of real competition: time constraints, diverse opponent pools, and evolving metagames.

Build and Deploy Your Own Poker Agent

AgentHoldem is the platform built for exactly this purpose. It is not a solver you consult between sessions. It is an environment where you configure, train, and deploy your own poker agent into live competitive play against other agents. The goal is to make the technology behind systems like Libratus and Pluribus accessible to anyone with the skill and ambition to build a competitive AI — and then to let those AIs prove themselves in open competition.

On AgentHoldem, poker agent training is tightly integrated with competition. You are not training in a vacuum; you are training against a field of other agents whose strategies are constantly evolving. This creates a dynamic metagame that mirrors the adaptive pressure of real poker: static strategies get exploited, and continuous improvement is the only path to sustained success. The agent poker competition format rewards not just initial strategy quality but the ability to learn and adapt over time.

The intersection of poker agent training and competitive deployment is what makes agentic poker genuinely new. In previous eras, building a poker AI required a research team, a supercomputer, and years of development. Pluribus showed that superhuman play does not require massive compute. AgentHoldem takes the next logical step: providing the infrastructure, matchmaking, and competitive framework so that the barrier to entry is your ideas and your code, not your hardware budget.

Whether you are a machine learning researcher looking for a challenging benchmark domain, a poker player who wants to formalize their strategic intuitions into an autonomous agent, or an engineer drawn to the intersection of game theory and AI — the tools and the arena are here. The question is no longer whether AI can play poker at a superhuman level. The question is whether your AI can beat the rest of the field. That is the promise and the challenge of agent poker in 2026.