Login / Signup
Strategies for Safe Multi-Armed Bandits with Logarithmic Regret and Risk.
Tianrui Chen
Aditya Gangrade
Venkatesh Saligrama
Published in:
ICML (2022)
Keyphrases
</>
multi armed bandits
bandit problems
multi armed bandit problems
multi armed bandit
decision problems
regret bounds
worst case
online learning
decision making
lower bound
optical flow
maximum likelihood
optimal strategy