The Principle of Maximum Causal Entropy for Estimating Interacting Processes
Journal Article, IEEE Transactions on Information Theory, Vol. 59, No. 4, pp. 1966 - 1980, April, 2013
Abstract
The principle of maximum entropy provides a powerful framework for estimating joint, conditional, and marginal probability distributions. However, there are many important distributions with elements of interaction and feedback where its applicability has not been established. This work presents the principle of maximum causal entropy—an approach based on directed information theory for estimating an unknown process based on its interactions with a known process. We demonstrate the breadth of the approach using two applications: a predictive solution for inverse optimal control in decision processes and computing equilibrium strategies in sequential games.
BibTeX
@article{Ziebart-2013-7667,author = {Brian D. Ziebart and J. Andrew (Drew) Bagnell and Anind Dey},
title = {The Principle of Maximum Causal Entropy for Estimating Interacting Processes},
journal = {IEEE Transactions on Information Theory},
year = {2013},
month = {April},
volume = {59},
number = {4},
pages = {1966 - 1980},
keywords = {Maximum entropy, statistical estimation, causal entropy, directed information, inverse optimal control, inverse reinforcement learning, correlated equilibrium},
}
Copyright notice: This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. These works may not be reposted without the explicit permission of the copyright holder.