You2Me: Inferring Body Pose in Egocentric Video via First and Second Person Interactions - Robotics Institute Carnegie Mellon University

You2Me: Inferring Body Pose in Egocentric Video via First and Second Person Interactions

Evonne Ng, Donglai Xiang, Hanbyul Joo, and Kristen Grauman
Conference Paper, Proceedings of (CVPR) Computer Vision and Pattern Recognition, pp. 9887 - 9897, June, 2020

Abstract

The body pose of a person wearing a camera is of great interest for applications in augmented reality, healthcare, and robotics, yet much of the person's body is out of view for a typical wearable camera. We propose a learning-based approach to estimate the camera wearer's 3D body pose from egocentric video sequences. Our key insight is to leverage interactions with another person---whose body pose we can directly observe---as a signal inherently linked to the body pose of the first-person subject. We show that since interactions between individuals often induce a well-ordered series of back-and-forth responses, it is possible to learn a temporal model of the interlinked poses even though one party is largely out of view. We demonstrate our idea on a variety of domains with dyadic interaction and show the substantial impact on egocentric body pose estimation, which improves the state of the art.

BibTeX

@conference{Ng-2020-126802,
author = {Evonne Ng and Donglai Xiang and Hanbyul Joo and Kristen Grauman},
title = {You2Me: Inferring Body Pose in Egocentric Video via First and Second Person Interactions},
booktitle = {Proceedings of (CVPR) Computer Vision and Pattern Recognition},
year = {2020},
month = {June},
pages = {9887 - 9897},
}