bandits
bandits copied to clipboard
Python library for Multi-Armed Bandits
Bandits
Python library for Multi-Armed Bandits
Implements the following algorithms:
- Epsilon-Greedy
- UCB1
- Softmax
- Thompson Sampling (Bayesian)
- Bernoulli, Binomial <=> Beta Distributions
Installation
You can install bandits
with:
git clone https://github.com/bgalbraith/bandits.git
cd bandits
pip install .
Examples
References
Wikipedia
Blog Posts
- When to Run Bandit Tests Instead of A/B/n Tests
- Bandit theory, part I
- Bandit theory, part II
- Bandits for Recommendation Systems
- Recommendations with Thompson Sampling
- Personalization with Contextual Bandits
- Bayesian Bandits - optimizing click throughs with statistics
- Mulit-Armed Bandits
- Bayesian Bandits
- Python Multi-armed Bandits (and Beer!)
Presentations
- Boston Bayesians Meetup 2016 - Bayesian Bandits From Scratch
- ODSC East 2016 - Bayesian Bandits
- NYC ML Meetup 2010 - Learning for Contextual Bandits
Books and Book Chapters
- Reinforcement Learning: An Introduction
- Multi-armed Bandit Allocation Indices
- Bandit Algorithms for Website Optimization
- Multi-Armed Bandit Problems (in Foundations and Applications of Sensor Management)