Login / Signup
Explore no more: improved high-probability regret bounds for non-stochastic bandits.
Gergely Neu
Published in:
CoRR (2015)
Keyphrases
</>
multi armed bandit
regret bounds
online learning
lower bound
reinforcement learning
linear regression
upper bound
image sequences
support vector
probability distribution
information theoretic
conditional probabilities