/Multi-armed-bandits

Multi-armed bandits experiment with epsilon-greedy, UCB, Thompson sampling, Bayesian-greedy and HA-UCB

Primary LanguageJupyter Notebook

Watchers