Display Publications by [Year] [Type] [Topic]

Difference evaluation functions have resulted in excellent multiagent behavior in many domains, including air traffic and mobile robot control. However, calculating difference evaluation functions requires determining the value of a counterfactual system objective function, which is often difficult when the system objective function is unknown or global state and action information is unavailable. In this work, we demonstrate that a local estimate of the system evaluation function may be used to estimate difference evaluations using readily available information, allowing for difference evaluations to be computed in multiagent systems where the mathematical form of the objective function is not known. This approximation technique is tested in two domains, and we demonstrate that approximating difference evaluation functions results in better performance and faster learning than when using global evaluation functions. Finally, we demonstrate the effectiveness of the learned policies on a set of Pioneer P3-DX robots.

(unavailable)

@InProceedings{tumer-colby_aamas16, author = {M. Colby and T. Duchow-Pressley and J. J. Chung and K. Tumer}, title = {Local Approximation of Difference Evaluation Functions}, booktitle = {Proceedings of the Fifteenth International Joint Conference on Autonomous Agents and Multiagent Systems}, address = {Singapore}, month = {May}, abstract ={Difference evaluation functions have resulted in excellent multiagent behavior in many domains, including air traffic and mobile robot control. However, calculating difference evaluation functions requires determining the value of a counterfactual system objective function, which is often difficult when the system objective function is unknown or global state and action information is unavailable. In this work, we demonstrate that a local estimate of the system evaluation function may be used to estimate difference evaluations using readily available information, allowing for difference evaluations to be computed in multiagent systems where the mathematical form of the objective function is not known. This approximation technique is tested in two domains, and we demonstrate that approximating difference evaluation functions results in better performance and faster learning than when using global evaluation functions. Finally, we demonstrate the effectiveness of the learned policies on a set of Pioneer P3-DX robots.}, pages={521-529}, bib2html_pubtype = {Refereed Conference Papers}, bib2html_rescat = {Multiagent Systems, Robotics}, year = {2016} }

Generated by bib2html.pl (written by Patrick Riley ) on Tue Jun 26, 2018 19:10:42