Inverse Reinforcement Learning with Explicit Policy Estimates - Robotics Institute Carnegie Mellon University
Loading Events

PhD Speaking Qualifier

April

9
Fri
Navyata Nilesh Sanghvi PhD Student Robotics Institute,
Carnegie Mellon University
Friday, April 9
10:00 am to 11:00 am
Inverse Reinforcement Learning with Explicit Policy Estimates

Abstract:
Various methods for solving the inverse reinforcement learning (IRL) problem have been developed independently in machine learning and economics. In particular, the method of Maximum Causal Entropy IRL is based on the perspective of entropy maximization, while related advances in the field of economics instead assume the existence of unobserved action shocks to explain expert behavior (Nested Fixed Point Algorithm, Conditional Choice Probability method, Nested Pseudo-Likelihood Algorithm). In this work, we make previously unknown connections between these related methods from both fields. We achieve this by showing that they all belong to a class of optimization problems, characterized by a common form of the objective, the associated policy and the objective gradient. We demonstrate key computational and algorithmic differences which arise between the methods due to an approximation of the optimal soft value function, and describe how this leads to more efficient algorithms. The development of a common theoretical framework results in a unified perspective of related methods from both fields. Our insights enable us to investigate the suitability of methods for various problem scenarios, based on their underlying assumptions and the resultant quality of solutions.

Committee:
Kris Kitani (Advisor)
Maxim Likhachev
Deepak Pathak
Xiaofang Wang