Keyphrases
- bandit problems
- decision problems
- multi armed bandits
- exploration exploitation
- maximum likelihood
- gaussian distribution
- multi armed bandit problems
- expected utility
- gaussian mixture
- learning algorithm
- linear programming problems
- gaussian noise
- optimal policy
- computational complexity
- search algorithm
- reinforcement learning
- multiscale