DeepStack: expert-level artificial intelligence in heads-up no-limit poker
From MaRDI portal
Publication:4645965
Abstract: Artificial intelligence has seen several breakthroughs in recent years, with games often serving as milestones. A common feature of these games is that players have perfect information. Poker is the quintessential game of imperfect information, and a longstanding challenge problem in artificial intelligence. We introduce DeepStack, an algorithm for imperfect information settings. It combines recursive reasoning to handle information asymmetry, decomposition to focus computation on the relevant decision, and a form of intuition that is automatically learned from self-play using deep learning. In a study involving 44,000 hands of poker, DeepStack defeated with statistical significance professional poker players in heads-up no-limit Texas hold'em. The approach is theoretically sound and is shown to produce more difficult to exploit strategies than prior approaches.
Recommendations
Cited in
(34)- Limited lookahead in imperfect-information games
- Computing large market equilibria using abstractions
- scientific article; zbMATH DE number 1784984 (Why is no real title available?)
- Approximating maxmin strategies in imperfect recall games using A-loss recall property
- Identifying behaviorally robust strategies for normal form games under varying forms of uncertainty
- World-class interpretable poker
- Robust and resource-efficient identification of two hidden layer neural networks
- The Hanabi challenge: a new frontier for AI research
- Counterfactuals as modal conditionals, and their probability
- A multivariate Riesz basis of ReLU neural networks
- Mathematical consistency and long-term behaviour of a dynamical system with a self-organising vector field
- CECMLP: new cipher-based evaluating collaborative multi-layer perceptron scheme in federated learning
- Committing to correlated strategies with multiple leaders
- Generosity, selfishness and exploitation as optimal greedy strategies for resource sharing
- Successful Nash equilibrium agent for a three-player imperfect-information game
- Computing human-understandable strategies: deducing fundamental rules of poker strategy
- HSVI can solve zero-sum partially observable stochastic games
- Superhuman AI for heads-up no-limit poker: Libratus beats top professionals
- A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play
- Analysis of Hannan consistent selection for Monte Carlo tree search in simultaneous move games
- Faster algorithms for extensive-form game solving via improved smoothing functions
- The challenge of poker
- Superhuman AI for multiplayer poker
- Value functions for depth-limited solving in zero-sum imperfect-information games
- Solving zero-sum one-sided partially observable stochastic games
- Rethinking formal models of partially observable multiagent decision making
- DCENet: a dynamic correlation evolve network for short-term traffic prediction
- Multi-agent reinforcement learning: a selective overview of theories and algorithms
- Automatically designing counterfactual regret minimization algorithms for solving imperfect-information games
- Automated construction of bounded-loss imperfect-recall abstractions in extensive-form games
- Deep reinforcement learning with emergent communication for coalitional negotiation games
- Evaluating strategic structures in multi-agent inverse reinforcement learning
- Distinguishing luck from skill through statistical simulation: a case study
- Simple uncoupled no-regret learning dynamics for extensive-form correlated equilibrium
This page was built for publication: DeepStack: expert-level artificial intelligence in heads-up no-limit poker
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q4645965)