Towards Control-Centric Representations in Reinforcement Learning from Images
Published in arxiv
Recommended citation:Liu C, Zang H, Li X, et al. Towards Control-Centric Representations in Reinforcement Learning from Images[J]. arXiv preprint arXiv:2310.16655, 2023
Paper link:Image-based Reinforcement Learning is a practical yet challenging task. A major hurdle lies in extracting control-centric representations while disregarding irrelevant information. While approaches that follow the bisimulation principle exhibit the potential in learning state representations to address this issue, they still grapple with the limited expressive capacity of latent dynamics and the inadaptability to sparse reward environments. To address these limitations, we introduce ReBis, which aims to capture control-centric information by integrating reward-free control information alongside reward-specific knowledge. ReBis utilizes a transformer architecture to implicitly model the dynamics and incorporates block-wise masking to eliminate spatiotemporal redundancy. Moreover, ReBis combines bisimulation-based loss with asymmetric reconstruction loss to prevent feature collapse in environments with sparse rewards. Empirical studies on two large benchmarks, including Atari games and DeepMind Control Suit, demonstrate that ReBis has superior performance compared to existing methods, proving its effectiveness.
@article{liu2023towards,
title={Towards Control-Centric Representations in Reinforcement Learning from Images},
author={Liu, Chen and Zang, Hongyu and Li, Xin and Heng, Yong and Wang, Yifei and Fang, Zhen and Wang, Yisen and Wang, Mingzhong},
journal={arXiv preprint arXiv:2310.16655},
year={2023}
}