Reasoning about Counterfactuals to Improve Human Inverse Reinforcement Learning - Robotics Institute Carnegie Mellon University

Reasoning about Counterfactuals to Improve Human Inverse Reinforcement Learning

Conference Paper, Proceedings of (IROS) IEEE/RSJ International Conference on Intelligent Robots and Systems, October, 2022

Abstract

To collaborate well with robots, we must be able to understand their decision making. Humans naturally infer other agents' beliefs and desires by reasoning about their observable behavior in a way that resembles inverse reinforcement learning (IRL). Thus, robots can convey their beliefs and desires by providing demonstrations that are informative for a human learner's IRL. An informative demonstration is one that differs strongly from the learner's expectations of what the robot will do given their current understanding of the robot's decision making. However, standard IRL does not model the learner's existing expectations, and thus cannot do this counterfactual reasoning. We propose to incorporate the learner's current understanding of the robot's decision making into our model of human IRL, so that a robot can select demonstrations that maximize the human's understanding. We also propose a novel measure for estimating the difficulty for a human to predict instances of a robot's behavior in unseen environments. A user study finds that our test difficulty measure correlates well with human performance and confidence. Interestingly, considering human beliefs and counterfactuals when selecting demonstrations decreases human performance on easy tests, but increases performance on difficult tests, providing insight on how to best utilize such models.

BibTeX

@conference{Lee-2022-133240,
author = {Michael S. Lee and Henny Admoni and Reid Simmons},
title = {Reasoning about Counterfactuals to Improve Human Inverse Reinforcement Learning},
booktitle = {Proceedings of (IROS) IEEE/RSJ International Conference on Intelligent Robots and Systems},
year = {2022},
month = {October},
keywords = {inverse reinforcement learning, learning from demonstration, scaffolding, counterfactual reasoning, policy summarization, machine teaching},
}