Google Research

Self-Imitation Learning

ICML (2018)

Abstract

This paper proposes Self-Imitation Learning (SIL), a method that learns to imitate the agent’s past good trajectories and when combined with an actor-critic architecture can achieve better exploration and better performance. Specifically, our empirical results show that SIL improves the advantage actor-critic (A2C) on several hard exploration Atari games and is competitive to the state-of-the-art count-based exploration methods. We also show that SIL improves proximal policy optimization (PPO) on continuous control tasks.

Research Areas

Learn more about how we do research

We maintain a portfolio of research projects, providing individuals and teams the freedom to emphasize specific types of work