The Information Bottleneck in RL

State representations hold the promise of simplifying control, allowing a reinforcement learning (RL) agent to solve a task more quickly, and to generalize better to new tasks. While this representation can be learned in a multi-task setting, doing so requires manually constructing a suitable task distribution, an onerous requirement. Instead, we propose to learn a representation that encodes as few bits of the input as possible, subject to the constraint that the agent is still able to solve this task. This essentially amounts to placing “blinkers” on our agent, with the aim of ignoring spurious attributes of the state. Formally, we adopt the information bottleneck (IB) as a measure of representational complexity, and augment the standard RL objective with a lower bound.

References

  • Emergence of Invariance and Disentanglement in Deep Representations, Alessandro Achille, Stefano Soatto. arXiv:1706.01350 [cs, stat] (2018)
  • Deep Variational Information Bottleneck, Alex Alemi, Ian Fischer, Josh Dillon, Kevin Murphy. Proceedings of ICLR 2017 (2017)
  • Generalization in reinforcement learning with selective noise injection and information bottleneck, Maximilian Igl, Kamil Ciosek, Yingzhen Li, Sebastian Tschiatschek, Cheng Zhang, Sam Devlin, Katja Hofmann. Advances in neural information processing systems (2019)