Skip to main content

Minimax Optimal and Computationally Efficient Algorithms for Distributionally Robust Offline Reinforcement Learning

Publication ,  Conference
Liu, Z; Xu, P
Published in: Advances in Neural Information Processing Systems
January 1, 2024

Distributionally robust offline reinforcement learning (RL), which seeks robust policy training against environment perturbation by modeling dynamics uncertainty, calls for function approximations when facing large state-action spaces. However, the consideration of dynamics uncertainty introduces essential nonlinearity and computational burden, posing unique challenges for analyzing and practically employing function approximation. Focusing on a basic setting where the nominal model and perturbed models are linearly parameterized, we propose minimax optimal and computationally efficient algorithms realizing function approximation and initiate the study on instance-dependent suboptimality analysis in the context of robust offline RL. Our results uncover that function approximation in robust offline RL is essentially distinct from and probably harder than that in standard offline RL. Our algorithms and theoretical results crucially depend on a novel function approximation mechanism incorporating variance information, a new procedure of suboptimality and estimation uncertainty decomposition, a quantification of the robust value function shrinkage, and a meticulously designed family of hard instances, which might be of independent interest.

Duke Scholars

Published In

Advances in Neural Information Processing Systems

ISSN

1049-5258

Publication Date

January 1, 2024

Volume

37

Related Subject Headings

  • 4611 Machine learning
  • 1702 Cognitive Sciences
  • 1701 Psychology
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Liu, Z., & Xu, P. (2024). Minimax Optimal and Computationally Efficient Algorithms for Distributionally Robust Offline Reinforcement Learning. In Advances in Neural Information Processing Systems (Vol. 37).
Liu, Z., and P. Xu. “Minimax Optimal and Computationally Efficient Algorithms for Distributionally Robust Offline Reinforcement Learning.” In Advances in Neural Information Processing Systems, Vol. 37, 2024.
Liu, Z., and P. Xu. “Minimax Optimal and Computationally Efficient Algorithms for Distributionally Robust Offline Reinforcement Learning.” Advances in Neural Information Processing Systems, vol. 37, 2024.

Published In

Advances in Neural Information Processing Systems

ISSN

1049-5258

Publication Date

January 1, 2024

Volume

37

Related Subject Headings

  • 4611 Machine learning
  • 1702 Cognitive Sciences
  • 1701 Psychology