Kagan Tumer's Publications

Display Publications by [Year] [Type] [Topic]


Decentralized Coordination via Task Decomposition and Reward Shaping (extended abstract). A. Iscen and K. Tumer. In Proceedings of the Twelveth International Joint Conference on Autonomous Agents and Multiagent Systems, Minneapolis, MN, May 2013.

Abstract

In this work, we introduce a method for decentralized coordination in cooperative multiagent multi-task problems where the subtasks and agents are homogeneous. Using the method proposed, the agents cooperate at the high level task selection using the knowledge they gather by learning subtasks. We introduce a subtask selection method for single agent multi-task MDPs and we extend the work to multiagent multi-task MDPs by using reward shaping at the subtask level to coordinate the agents. Our results on a multi-rover problem show that agents which use the combination of task decomposition and subtask based difference rewards result in significant improvement both in terms of learning speed, and converged policies.

Download

[PDF]210.6kB  

BibTeX Entry

@inproceedings{tumer-iscen-helm_aamas13,
        author = {A. Iscen and  K. Tumer},
        title = {Decentralized Coordination via Task Decomposition and Reward Shaping (extended abstract)},
        booktitle = {Proceedings of the Twelveth International Joint Conference on Autonomous Agents and Multiagent Systems},
	month = {May},
	address = {Minneapolis, MN},
	abstract={In this work, we introduce a method for decentralized coordination in cooperative multiagent multi-task problems where the subtasks and agents are homogeneous. Using the method proposed, the agents cooperate at the high level task selection using the knowledge they gather by learning subtasks. We introduce a subtask selection method for single agent multi-task MDPs and we extend the work to multiagent multi-task MDPs by using reward shaping at the subtask level to coordinate the agents. Our results on a multi-rover problem show that agents which use the combination of task decomposition and subtask based difference rewards result in significant improvement both in terms of learning speed, and converged policies.},
	bib2html_pubtype = {Refereed Conference Papers},
	bib2html_rescat = {Multiagent Systems, Reinforcement Learning},
        year = {2013}
}

Generated by bib2html.pl (written by Patrick Riley ) on Wed Apr 01, 2020 17:39:43