Information-Theoretic Confidence Bounds for Reinforcement Learning.
Xiuyuan LuBenjamin Van RoyPublished in: NeurIPS (2019)
Keyphrases
- information theoretic
- confidence bounds
- reinforcement learning
- mutual information
- information theory
- theoretic framework
- log likelihood
- information bottleneck
- jensen shannon divergence
- kullback leibler divergence
- information theoretic measures
- entropy measure
- learning algorithm
- minimum description length
- relative entropy
- multi modal
- kl divergence
- multi modality