## Kagan Tumer's Publications

Display Publications by [Year] [Type] [Topic]

#### Abstract

In many multi agent learning problems, it is difficult to determine, a priori, the agent reward structure that will lead to good performance. This problem is particularly pronounced in continuous, noisy domains ill-suited to simple table backup schemes commonly used in TD(\lambda)/Q-learning. In this paper, we present a new reward evaluation method that allows the tradeoff between coordination among the agents and the difficulty of the learning problem each agent faces to be visualized. This method is independent of the learning algorithm and is only a function of the problem domain and the agents' reward structure. We then use this reward efficiency visualization method to determine an effective reward without performing extensive simulations. We test this method in both a static and a dynamic multi-rover learning domain where the the agents have continuous state spaces and where their actions are noisy (e.g., the agents' movement decisions are not always carried out properly). Our results show that in the more difficult dynamic domain, the reward efficiency visualization method provides a two order of magnitude speedup in selecting a good reward. Most importantly it allows one to quicklycreate and verify rewards tailored to the observational limitations of the domain.

#### BibTeX Entry

@inproceedings{tumer-agogino_aamas05,
author = {A. Agogino and K. Tumer},
title = {Multiagent Reward Analysis for Learning in Noisy Domains},
booktitle = {Proceedings of the Fourth International Joint Conference on Autonomous Agents and Multiagent Systems},
month = {July},