Reinforcement Learning with Prototypical Representations

Denis Yarats, Rob Fergus, Alessandro Lazaric, Lerrel Pinto
Proceedings of the 38th International Conference on Machine Learning, PMLR 139:11920-11931, 2021.

Abstract

Learning effective representations in image-based environments is crucial for sample efficient Reinforcement Learning (RL). Unfortunately, in RL, representation learning is confounded with the exploratory experience of the agent – learning a useful representation requires diverse data, while effective exploration is only possible with coherent representations. Furthermore, we would like to learn representations that not only generalize across tasks but also accelerate downstream exploration for efficient task-specific training. To address these challenges we propose Proto-RL, a self-supervised framework that ties representation learning with exploration through prototypical representations. These prototypes simultaneously serve as a summarization of the exploratory experience of an agent as well as a basis for representing observations. We pre-train these task-agnostic representations and prototypes on environments without downstream task information. This enables state-of-the-art downstream policy learning on a set of difficult continuous control tasks.

Cite this Paper


BibTeX
@InProceedings{pmlr-v139-yarats21a, title = {Reinforcement Learning with Prototypical Representations}, author = {Yarats, Denis and Fergus, Rob and Lazaric, Alessandro and Pinto, Lerrel}, booktitle = {Proceedings of the 38th International Conference on Machine Learning}, pages = {11920--11931}, year = {2021}, editor = {Meila, Marina and Zhang, Tong}, volume = {139}, series = {Proceedings of Machine Learning Research}, month = {18--24 Jul}, publisher = {PMLR}, pdf = {http://proceedings.mlr.press/v139/yarats21a/yarats21a.pdf}, url = {https://proceedings.mlr.press/v139/yarats21a.html}, abstract = {Learning effective representations in image-based environments is crucial for sample efficient Reinforcement Learning (RL). Unfortunately, in RL, representation learning is confounded with the exploratory experience of the agent – learning a useful representation requires diverse data, while effective exploration is only possible with coherent representations. Furthermore, we would like to learn representations that not only generalize across tasks but also accelerate downstream exploration for efficient task-specific training. To address these challenges we propose Proto-RL, a self-supervised framework that ties representation learning with exploration through prototypical representations. These prototypes simultaneously serve as a summarization of the exploratory experience of an agent as well as a basis for representing observations. We pre-train these task-agnostic representations and prototypes on environments without downstream task information. This enables state-of-the-art downstream policy learning on a set of difficult continuous control tasks.} }
Endnote
%0 Conference Paper %T Reinforcement Learning with Prototypical Representations %A Denis Yarats %A Rob Fergus %A Alessandro Lazaric %A Lerrel Pinto %B Proceedings of the 38th International Conference on Machine Learning %C Proceedings of Machine Learning Research %D 2021 %E Marina Meila %E Tong Zhang %F pmlr-v139-yarats21a %I PMLR %P 11920--11931 %U https://proceedings.mlr.press/v139/yarats21a.html %V 139 %X Learning effective representations in image-based environments is crucial for sample efficient Reinforcement Learning (RL). Unfortunately, in RL, representation learning is confounded with the exploratory experience of the agent – learning a useful representation requires diverse data, while effective exploration is only possible with coherent representations. Furthermore, we would like to learn representations that not only generalize across tasks but also accelerate downstream exploration for efficient task-specific training. To address these challenges we propose Proto-RL, a self-supervised framework that ties representation learning with exploration through prototypical representations. These prototypes simultaneously serve as a summarization of the exploratory experience of an agent as well as a basis for representing observations. We pre-train these task-agnostic representations and prototypes on environments without downstream task information. This enables state-of-the-art downstream policy learning on a set of difficult continuous control tasks.
APA
Yarats, D., Fergus, R., Lazaric, A. & Pinto, L.. (2021). Reinforcement Learning with Prototypical Representations. Proceedings of the 38th International Conference on Machine Learning, in Proceedings of Machine Learning Research 139:11920-11931 Available from https://proceedings.mlr.press/v139/yarats21a.html.

Related Material