Kagan Tumer's Publications

Display Publications by [Year] [Type] [Topic]


Curriculum Learning for Tightly Coupled Multiagent Systems (Extended Abstract). G. Rockefeller, P. Mannion, and K. Tumer. In Proceedings of the International Joint Conference on Autonomous Agents and Multiagent Systems, pp. , Montreal, Canada, May 2019.

Abstract

In this paper, we leverage curriculum learning (CL) to improve the performance of multiagent systems (MAS) that are trained with the cooperative coevolution of artificial neural networks. We design curricula to progressively change two dimensions: scale (i.e. domain size) and coupling (i.e. the number of agents required to complete a subtask). We demonstrate that CL can successfully mitigate the challenge of learning on a sparse reward signal resulting from a high degree of coupling in complex MAS. We also show that, in most cases, the combination of difference reward shaping with CL can improve performance by up to 56%. We evaluate our CL methods on the tightly coupled multi-rover domain. CL increased converged system performance on all tasks presented. Furthermore, agents were only able to learn when trained with CL for most tasks.

Download

[PDF]483.6kB  

BibTeX Entry

@InProceedings{tumer-rockefeller_aamas19,
author = {G. Rockefeller and P. Mannion and K. Tumer},
title = {Curriculum Learning for Tightly Coupled Multiagent Systems (Extended Abstract)},
booktitle = {Proceedings of the International Joint Conference on Autonomous Agents and Multiagent Systems},
address = {Montreal, Canada},
month = {May},
 pages={},
 abstract={In this paper, we leverage curriculum learning (CL) to improve the performance of multiagent systems (MAS) that are trained with the cooperative coevolution of artificial neural networks. We design curricula to progressively change two dimensions: scale (i.e. domain size) and coupling (i.e. the number of agents required to complete a subtask). We demonstrate that CL can successfully mitigate the challenge of learning on a sparse reward signal resulting from a high degree of coupling in complex MAS. We also show that, in most cases, the combination of difference reward shaping with CL can improve performance by up to 56%. We evaluate our CL methods on the tightly coupled multi-rover domain. CL increased converged system performance on all tasks presented. Furthermore, agents were only able to learn when trained with CL for most tasks.},
	bib2html_pubtype = {Refereed Conference Papers},
	bib2html_rescat = {Multiagent Systems},
year = {2019}
}

Generated by bib2html.pl (written by Patrick Riley ) on Wed Apr 01, 2020 17:39:43