Login / Signup

A Field Test of Bandit Algorithms for Recommendations: Understanding the Validity of Assumptions on Human Preferences in Multi-armed Bandits.

Liu LeqiGiulio ZhouFatma Kilinç-KarzanZachary C. LiptonAlan L. Montgomery
Published in: CoRR (2023)
Keyphrases
  • bandit problems
  • multi armed bandits
  • multi armed bandit
  • learning algorithm
  • decision making
  • user preferences
  • e learning
  • bayesian networks
  • decision makers