next up previous contents
Next: Bayesian poker Up: Koller and Pfeffer's work Previous: Koller and Pfeffer's work

Optimal versus maximal player

Koller and Pfeffer's goal is to create an optimal poker player. By definition, the optimal player does the best that can be done against a rational (perfect) opponent, and it does not do worse even if its strategy is revealed to its opponent. However, the optimal strategy does not take advantages of mistakes when they become apparent, and human players invariably make mistakes. A maximal player will deviate from the optimal strategy to exploit the observed weak points in the opponent's play. In theory, the maximal player must take the risk of being sub-optimal to exploit a sub-optimal opponent. In practice, the risk is small and well rewarded.

In contrast to Koller and Pfeffer's aim, Loki is not an optimal player. Our goal is to create a maximal player, which uses opponent modeling to exploit patterns in its opponents' play, with the intention of winning the most money it can in every situation. Furthermore, since it does not seem feasible to determine an optimal player for real multi-player poker, a program to play real-world poker in the near future most likely will not be a game-theoretic optimal player.

Nevertheless, Koller and Pfeffer have suggested that an alternative approach to deal with less-than-perfect players is to learn the type of mistake that a player is prone to make. This approach can be used when there is a long-term interaction with the same player. The authors point out that the ability of the Gala language to capture regularities in the game may be particularly useful in this context, since the high-level description of a game state can provide features for the learning algorithm. One can see this learning algorithm as a potential opponent modeling component for a program based on the Gala system.


next up previous contents
Next: Bayesian poker Up: Koller and Pfeffer's work Previous: Koller and Pfeffer's work
Lourdes Pena
1999-09-10