-
Notifications
You must be signed in to change notification settings - Fork 0
MartinRuz/BA_Multi-armed-bandit
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
the accompanying python-script simulates a multi-armed bandit with two different policies as documented in my thesis. the index 100 in lines 5-9 marks the number of repetitions of the same experiment. N is the number of rounds, the policy choosese N times. d is the number of possible choices in each round. in the for-loop 17-23 we fill the array m with random normal distributed values, with mean 100+r and variance v, where r is a random number between -1 and 1 and v is a random number between 0.05 and 0.25. Then we perform the two algorithms and print the results.
About
No description, website, or topics provided.
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published