babaniyi / bandits

Algorithms for multiarmed bandits such as e-greedy, thompson sampling, etc.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Multiarmed bandits

I introduced various multiarmed bandits algorithms such as e-greedy, annealing epsilon greedy, thompson sampling, UCB etc. I also compared the performance of these algorithms and how they can quickly find the best arm.

About

Algorithms for multiarmed bandits such as e-greedy, thompson sampling, etc.


Languages

Language:Jupyter Notebook 98.6%Language:Rich Text Format 0.9%Language:Python 0.4%