Difference between revisions of "Master Level Theses"

From AIRWiki
Jump to: navigation, search
Line 7: Line 7:
 
{{Project template
 
{{Project template
 
|title=Reinforcement Learning in Poker
 
|title=Reinforcement Learning in Poker
|tutor=Marcello Restelli
+
|tutor=Marcello Restelli (restelli-AT-elet-DOT-polimi-DOT-it
 
|description=In this years, Artificial Intelligence research has shifted its attention from fully observable environments such as Chess to more challenging partially observable ones such as Poker.
 
|description=In this years, Artificial Intelligence research has shifted its attention from fully observable environments such as Chess to more challenging partially observable ones such as Poker.
  

Revision as of 15:04, 29 September 2008

Machine Learning


Title: Reinforcement Learning in Poker
PokerPRLT.png
Description: In this years, Artificial Intelligence research has shifted its attention from fully observable environments such as Chess to more challenging partially observable ones such as Poker.

Up to this moment research in this kind of environments, which can be formalized as Partially Observable Stochastic Games, has been more from a game theoretic point of view, thus focusing on the pursue of optimality and equilibrium, with no attention to payoff maximization, which may be more interesting in many real-world contexts.

On the other hand Reinforcement Learning techniques demonstrated to be successful in solving both fully observable problems, single and multi-agent, and single-agent partially observable ones, while lacking application to the partially observable multi-agent framework.

This research aims at studying the solution of Partially Observable Stochastic Games, analyzing the possibility to combine the Opponent Modeling concept with the well proven Reinforcement Learning solution techniques to solve problems in this framework, adopting Poker as testbed.

Tutor: Marcello Restelli (restelli-AT-elet-DOT-polimi-DOT-it
Start: Anytime
Number of students: 1-2
CFU: 20-40