Skip to main content

Learning Markov State Abstractions for Deep Reinforcement Learning

Publication ,  Conference
Allen, C; Parikh, N; Gottesman, O; Konidaris, G
Published in: Advances in Neural Information Processing Systems
January 1, 2021

A fundamental assumption of reinforcement learning in Markov decision processes (MDPs) is that the relevant decision process is, in fact, Markov. However, when MDPs have rich observations, agents typically learn by way of an abstract state representation, and such representations are not guaranteed to preserve the Markov property. We introduce a novel set of conditions and prove that they are sufficient for learning a Markov abstract state representation. We then describe a practical training procedure that combines inverse model estimation and temporal contrastive learning to learn an abstraction that approximately satisfies these conditions. Our novel training objective is compatible with both online and offline training: it does not require a reward signal, but agents can capitalize on reward information when available. We empirically evaluate our approach on a visual gridworld domain and a set of continuous control benchmarks. Our approach learns representations that capture the underlying structure of the domain and lead to improved sample efficiency over state-of-the-art deep reinforcement learning with visual features—often matching or exceeding the performance achieved with hand-designed compact state information.

Duke Scholars

Published In

Advances in Neural Information Processing Systems

ISSN

1049-5258

ISBN

9781713845393

Publication Date

January 1, 2021

Volume

10

Start / End Page

8229 / 8241

Related Subject Headings

  • 4611 Machine learning
  • 1702 Cognitive Sciences
  • 1701 Psychology
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Allen, C., Parikh, N., Gottesman, O., & Konidaris, G. (2021). Learning Markov State Abstractions for Deep Reinforcement Learning. In Advances in Neural Information Processing Systems (Vol. 10, pp. 8229–8241).
Allen, C., N. Parikh, O. Gottesman, and G. Konidaris. “Learning Markov State Abstractions for Deep Reinforcement Learning.” In Advances in Neural Information Processing Systems, 10:8229–41, 2021.
Allen C, Parikh N, Gottesman O, Konidaris G. Learning Markov State Abstractions for Deep Reinforcement Learning. In: Advances in Neural Information Processing Systems. 2021. p. 8229–41.
Allen, C., et al. “Learning Markov State Abstractions for Deep Reinforcement Learning.” Advances in Neural Information Processing Systems, vol. 10, 2021, pp. 8229–41.
Allen C, Parikh N, Gottesman O, Konidaris G. Learning Markov State Abstractions for Deep Reinforcement Learning. Advances in Neural Information Processing Systems. 2021. p. 8229–8241.

Published In

Advances in Neural Information Processing Systems

ISSN

1049-5258

ISBN

9781713845393

Publication Date

January 1, 2021

Volume

10

Start / End Page

8229 / 8241

Related Subject Headings

  • 4611 Machine learning
  • 1702 Cognitive Sciences
  • 1701 Psychology