Skip to content

Commit

Permalink
update examples
Browse files Browse the repository at this point in the history
  • Loading branch information
usaito committed Feb 1, 2021
1 parent dd68eef commit 36f544a
Show file tree
Hide file tree
Showing 3 changed files with 21 additions and 30 deletions.
2 changes: 0 additions & 2 deletions examples/multiclass/evaluate_off_policy_estimators.py
Original file line number Diff line number Diff line change
Expand Up @@ -48,8 +48,6 @@
SelfNormalizedInverseProbabilityWeighting(),
DoublyRobust(),
SelfNormalizedDoublyRobust(),
SwitchInverseProbabilityWeighting(tau=1, estimator_name="switch-ipw (tau=1)"),
SwitchInverseProbabilityWeighting(tau=100, estimator_name="switch-ipw (tau=100)"),
SwitchDoublyRobust(tau=1, estimator_name="switch-dr (tau=1)"),
SwitchDoublyRobust(tau=100, estimator_name="switch-dr (tau=100)"),
DoublyRobustWithShrinkage(lambda_=1, estimator_name="dr-os (lambda=1)"),
Expand Down
2 changes: 2 additions & 0 deletions examples/obd/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,8 @@
Here, we use the open bandit dataset and pipeline to implement and evaluate OPE.
Specifically, we evaluate the estimation performances of well-known off-policy estimators using the ground-truth policy value of an evaluation policy, which is calculable with our data using on-policy estimation.

Please clone [the obp repository](https://github.com/st-tech/zr-obp) and download [the small sized Open Bandit Dataset](https://github.com/st-tech/zr-obp/tree/master/obd) to run this example.

## Evaluating Off-Policy Estimators

We evaluate the estimation performances of off-policy estimators, including Direct Method (DM), Inverse Probability Weighting (IPW), and Doubly Robust (DR).
Expand Down
Loading

0 comments on commit 36f544a

Please sign in to comment.