You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+6-11Lines changed: 6 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -15,18 +15,13 @@ A lightweight python library for bandit algorithms
15
15
16
16
## Introduction
17
17
18
-
This library is intended for fast and robust build of bandit algorithms. Hence it has the following features:
18
+
This library is intended to enable fast and robust comparison between different bandit algorithms. It provides following features:
19
19
20
-
* object-oriented design
21
-
*multiprocesses support
22
-
* friendly runtime info
20
+
* object-oriented design: this allows unnecessary environmental information to be hidden from learners. Besides, it is easy extend the library and implement new algorithms.
21
+
*multi-process support: it is not uncommon to run a game muitiple repetitions. One can run multiple repetitions simultaneously with this feature.
22
+
* friendly runtime information: useful information is provided when necessary, which reduces the difficulty of debug.
23
23
24
-
The library consists of four components i.e., `arms`, `bandits`, `learners` and `protocols`, which are explained in the following:
25
-
26
-
*`arms`: a set of arms used to build bandit environments
27
-
*`bandits`: bandit environments
28
-
*`learners`: bandit algorithms
29
-
*`protocols`: protocols which are used to coordinate the interactions between the learner and the bandit environment
24
+
The library consists of four submodules and they are `arms`, `bandits`, `learners` and `protocols` respectively, among which `protocols` are those used to coordinate the interactions between the learner and the bandit environment.
30
25
31
26
## Implemented Policies
32
27
@@ -150,7 +145,7 @@ The following figure shows the simulation results.
150
145
151
146

152
147
153
-
Please check this [notebook](examples/ordinary_bandit.ipynb) to figure out more details.
148
+
Please check this [notebook](examples/multi_armed_bandit.ipynb) to figure out more details.
0 commit comments