Experience-based Reinforcement Learning to Acquire Effective Behavior in a Multiagent Domain - Robotics Institute Carnegie Mellon University

Experience-based Reinforcement Learning to Acquire Effective Behavior in a Multiagent Domain

Sachiyo Arai, Katia Sycara, and Terence Payne
Conference Paper, Proceedings of 6th Pacific Rim International Conference on Artificial Intelligence (PRICAI '00), pp. 125 - 135, August, 2000

Abstract

In this paper, we discuss Profit-sharing, an experience-based reinforcement learning approach (which is similar to a Monte-Carlo based reinforcement learning method) that can be used to learn robust and effective actions within uncertain, dynamic, multi-agent systems. We introduce the cut-loop routine that discards looping behavior, and demonstrate its effectiveness empirically within a simplified NEO (non-combatant evacuation operation) domain. This domain consists of several agents which ferry groups of evacuees to one of several shelters. We demonstrate that the cut-loop routine makes the Profit-sharing approach adaptive and robust within a dynamic and uncertain domain, without the need for predefined knowledge or subgoals. We also compare it empirically with the popular Q-learning approach.

BibTeX

@conference{Arai-2000-16761,
author = {Sachiyo Arai and Katia Sycara and Terence Payne},
title = {Experience-based Reinforcement Learning to Acquire Effective Behavior in a Multiagent Domain},
booktitle = {Proceedings of 6th Pacific Rim International Conference on Artificial Intelligence (PRICAI '00)},
year = {2000},
month = {August},
pages = {125 - 135},
}