Skip to content. | Skip to navigation

Personal tools

Navigation

You are here: Home / Papers / Nonparametric Bayesian Inverse Reinforcement Learning for Multiple Reward Functions

Jaedeug Choi and Kee-Eung Kim (2012)

Nonparametric Bayesian Inverse Reinforcement Learning for Multiple Reward Functions

In: Proceedings of Neural Information Processing Systems (NIPS).

We present a nonparametric Bayesian approach to inverse reinforcement learning (IRL) for multiple reward functions. Most previous IRL algorithms assume that the behaviour data is obtained from an agent who is optimizing a single reward function, but this assumption is hard to guarantee in practice. Our approach is based on integrating the Dirichlet process mixture model into Bayesian IRL. We provide an efficient Metropolis-Hastings sampling algorithm utilizing the gradient of the posterior to estimate the underlying reward functions, and demonstrate that our approach outperforms previous ones via experiments on a number of problem domains.