Transfer Learning via Temporal Contrastive Learning - Robotics Institute Carnegie Mellon University

Transfer Learning via Temporal Contrastive Learning

Master's Thesis, Tech. Report, CMU-RI-TR-24-10, April, 2024

Abstract

This thesis introduces a novel transfer learning framework for deep reinforcement learning. The approach automatically combines goalconditioned policies with temporal contrastive learning to discover meaningful sub-goals. The approach involves pre-training a goal-conditioned agent, finetuning it on the target domain, and using contrastive learning to construct a planning graph that guides the agent via sub-goals. Experiments on PointMaze and multi-agent coordination Overcooked tasks demonstrate improved sample efficiency, the ability to solve sparse-reward and long-horizon problems, and enhanced interpretability compared to baselines. The results highlight the effectiveness of integrating goalconditioned policies with unsupervised temporal abstraction learning for complex multi-agent transfer learning. Compared to state-of-the-art baselines, our method achieves the same or better performances while requiring only 23.4% of the training samples.

BibTeX

@mastersthesis{Zeng-2024-140512,
author = {Weihao Zeng},
title = {Transfer Learning via Temporal Contrastive Learning},
year = {2024},
month = {April},
school = {Carnegie Mellon University},
address = {Pittsburgh, PA},
number = {CMU-RI-TR-24-10},
}