Experience-based Reinforcement Learning to Acquire Effective Behavior in a Multiagent Domain
Abstract
In this paper, we discuss Profit-sharing, an experience-based reinforcement learning approach (which is similar to a Monte-Carlo based reinforcement learning method) that can be used to learn robust and effective actions within uncertain, dynamic, multi-agent systems. We introduce the cut-loop routine that discards looping behavior, and demonstrate its effectiveness empirically within a simplified NEO (non-combatant evacuation operation) domain. This domain consists of several agents which ferry groups of evacuees to one of several shelters. We demonstrate that the cut-loop routine makes the Profit-sharing approach adaptive and robust within a dynamic and uncertain domain, without the need for predefined knowledge or subgoals. We also compare it empirically with the popular Q-learning approach.
BibTeX
@conference{Arai-2000-16761,author = {Sachiyo Arai and Katia Sycara and Terence Payne},
title = {Experience-based Reinforcement Learning to Acquire Effective Behavior in a Multiagent Domain},
booktitle = {Proceedings of 6th Pacific Rim International Conference on Artificial Intelligence (PRICAI '00)},
year = {2000},
month = {August},
pages = {125 - 135},
}