Skip to content

Latest commit

 

History

History
13 lines (8 loc) · 641 Bytes

README.md

File metadata and controls

13 lines (8 loc) · 641 Bytes

multi-armed-bandit

This repo is set up for a blog post I wrote on "The Multi-Armed Bandit Problem and Its Solutions".


The result of a small experiment on solving a Bernoulli bandit with K = 10 slot machines, each with a randomly initialized reward probability.

Alt text

  • (Left) The plot of time step vs the cumulative regrets.
  • (Middle) The plot of true reward probability vs estimated probability.
  • (Right) The fraction of each action is picked during the 5000-step run.