The Principle of Maximum Causal Entropy for Estimating Interacting Processes - Robotics Institute Carnegie Mellon University

The Principle of Maximum Causal Entropy for Estimating Interacting Processes

Brian D. Ziebart, J. Andrew (Drew) Bagnell, and Anind Dey
Journal Article, IEEE Transactions on Information Theory, Vol. 59, No. 4, pp. 1966 - 1980, April, 2013

Abstract

The principle of maximum entropy provides a powerful framework for estimating joint, conditional, and marginal probability distributions. However, there are many important distributions with elements of interaction and feedback where its applicability has not been established. This work presents the principle of maximum causal entropy—an approach based on directed information theory for estimating an unknown process based on its interactions with a known process. We demonstrate the breadth of the approach using two applications: a predictive solution for inverse optimal control in decision processes and computing equilibrium strategies in sequential games.

BibTeX

@article{Ziebart-2013-7667,
author = {Brian D. Ziebart and J. Andrew (Drew) Bagnell and Anind Dey},
title = {The Principle of Maximum Causal Entropy for Estimating Interacting Processes},
journal = {IEEE Transactions on Information Theory},
year = {2013},
month = {April},
volume = {59},
number = {4},
pages = {1966 - 1980},
keywords = {Maximum entropy, statistical estimation, causal entropy, directed information, inverse optimal control, inverse reinforcement learning, correlated equilibrium},
}