multi-armed-bandit
Here are 116 public repositories matching this topic...
Demo for NAIST Spring Seminar 2017
-
Updated
May 14, 2017 - JavaScript
A simple implementation of the multi_arm_bandit problem which can be used in Open AI gym as well.
-
Updated
Aug 28, 2017 - Python
Reinforcement Learning
-
Updated
Nov 30, 2017 - Python
Implementations of the bandit algorithms with unordered and ordered slates that are described in the paper "Non-Stochastic Bandit Slate Problems", by Kale et al. 2010.
-
Updated
May 30, 2018 - Python
In probability theory, the multi-armed bandit problem is a problem in which a fixed limited set of resources must be allocated between competing (alternative) choices in a way that maximizes their expected gain, when each choice's properties are only partially known at the time of allocation, and may become better understood as time passes or by…
-
Updated
Jun 1, 2018 - Jupyter Notebook
A bandit server that implements the multi-armed bandit for running a single experiment
-
Updated
Jun 6, 2018 - Go
Reinforcement Learning in Scala
-
Updated
Jun 18, 2018 - Scala
Implementation of the X-armed Bandits algorithm, as detailed in the paper, "X-armed Bandits", Bubeck et al., 2011.
-
Updated
Jul 12, 2018 - Python
Python implementation of multi-armed bandit using epsilon-greedy exploration and reward-average sampling estimation
-
Updated
Sep 25, 2018 - Jupyter Notebook
Il bandito bracciuto: Introduzione leggera ma non troppo all'apprendimento con rinforzo.
-
Updated
Jan 20, 2019 - Python
A golang library for solving multi armed bandit problem which can optimize your business choice on the fly without A/B testing
-
Updated
Apr 15, 2019 - Go
An implementation of solvers for the multi-armed-bandit-problem in JavaScript.
-
Updated
Apr 25, 2019 - JavaScript
-
Updated
May 6, 2019 - JavaScript
-
Updated
May 6, 2019 - Python
More about the exploration-exploitation tradeoff with harder bandits
-
Updated
May 12, 2019 - Jupyter Notebook
Software for the experiments reported in the RecSys 2019 paper "Multi-Armed Recommender System Bandit Ensembles"
-
Updated
Aug 16, 2019 - Java
A/B Testing Framework for Python (with optional Multi-armed bandit implementation)
-
Updated
Aug 18, 2019 - Python
Using the Airspeed Velocity tool (https://asv.readthedocs.io/) to benchmark SMPyBandits (https://github.com/SMPyBandits/SMPyBandits/)
-
Updated
Nov 6, 2019 - Python
Improve this page
Add a description, image, and links to the multi-armed-bandit topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the multi-armed-bandit topic, visit your repo's landing page and select "manage topics."