Home > Research > Publications & Outputs > Automated planning in repeated adversarial games
View graph of relations

Automated planning in repeated adversarial games

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
Close
Publication date2010
Host publicationProceedings of the 26th Conference on Uncertainty in Artificial Intelligence, UAI 2010
Pages376-383
Number of pages8
<mark>Original language</mark>English
Event26th Conference on Uncertainty in Artificial Intelligence, UAI 2010 - Catalina Island, CA, United States
Duration: 8/07/201011/07/2010

Conference

Conference26th Conference on Uncertainty in Artificial Intelligence, UAI 2010
Country/TerritoryUnited States
CityCatalina Island, CA
Period8/07/1011/07/10

Conference

Conference26th Conference on Uncertainty in Artificial Intelligence, UAI 2010
Country/TerritoryUnited States
CityCatalina Island, CA
Period8/07/1011/07/10

Abstract

Game theory's prescriptive power typically relies on full rationality and/or self-play interactions. In contrast, this work sets aside these fundamental premises and focuses instead on heterogeneous autonomous interactions between two or more agents. Specifically, we introduce a new and concise representation for repeated adversarial (constant-sum) games that highlight the necessary features that enable an automated planing agent to reason about how to score above the game's Nash equilibrium, when facing heterogeneous adversaries. To this end, we present TeamUP, a model-based RL algorithm designed for learning and planning such an abstraction. In essence, it is somewhat similar to R-max with a cleverly engineered reward shaping that treats exploration as an adversarial optimization problem. In practice, it attempts to find an ally with which to tacitly collude (in more than two-player games) and then collaborates on a joint plan of actions that can consistently score a high utility in adversarial repeated games. We use the inaugural Lemonade Stand Game Tournament1 to demonstrate the effectiveness of our approach, and find that TeamUP is the best performing agent, demoting the Tournament's actual winning strategy into second place. In our experimental analysis, we show hat our strategy successfully and consistently builds collaborations with many different heterogeneous (and sometimes very sophisticated) adversaries.