Information-Theoretic Generalization Bounds for Batch Reinforcement Learning.

Entropy (Basel)

School of Computing Science, Simon Fraser University, 8888 University Dr W, Burnaby, BC V5A 1S6, Canada.

Published: November 2024

We analyze the generalization properties of batch reinforcement learning (batch RL) with value function approximation from an information-theoretic perspective. We derive generalization bounds for batch RL using (conditional) mutual information. In addition, we demonstrate how to establish a connection between certain structural assumptions on the value function space and conditional mutual information. As a by-product, we derive a generalization bound via conditional mutual information, which was left open and may be of independent interest.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11593174PMC
http://dx.doi.org/10.3390/e26110995DOI Listing

Publication Analysis

Top Keywords

conditional mutual
12
generalization bounds
8
bounds batch
8
batch reinforcement
8
reinforcement learning
8
derive generalization
8
information-theoretic generalization
4
batch
4
learning analyze
4
analyze generalization
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!