Skip to content

Commit

Permalink
Revert "Revert "[tune-telemetry] Tag searcher and scheduler types. (r…
Browse files Browse the repository at this point in the history
…ay-project#33561)"" (ray-project#33740)

Additionally fix `test_usage_test.py`.

Signed-off-by: xwjiang2010 <xwjiang2010@gmail.com>
Signed-off-by: Jonathan Carter <jonathan.carter@magd.ox.ac.uk>
  • Loading branch information
xwjiang2010 authored and joncarter1 committed Apr 2, 2023
1 parent 5d488f4 commit d706309
Show file tree
Hide file tree
Showing 11 changed files with 129 additions and 15 deletions.
99 changes: 91 additions & 8 deletions python/ray/air/_internal/usage.py
Original file line number Diff line number Diff line change
@@ -1,9 +1,11 @@
from typing import TYPE_CHECKING
from typing import TYPE_CHECKING, Set, Union

from ray._private.usage.usage_lib import TagKey, record_extra_usage_tag

if TYPE_CHECKING:
from ray.train.trainer import BaseTrainer
from ray.tune.schedulers import TrialScheduler
from ray.tune.search import BasicVariantGenerator, Searcher

AIR_TRAINERS = {
"HorovodTrainer",
Expand All @@ -18,16 +20,97 @@
"XGBoostTrainer",
}

# searchers implemented by Ray Tune.
TUNE_SEARCHERS = {
"AxSearch",
"BayesOptSearch",
"TuneBOHB",
"DragonflySearch",
"HEBOSearch",
"HyperOptSearch",
"NevergradSearch",
"OptunaSearch",
"SkOptSearch",
"ZOOptSearch",
}

# These are just wrappers around real searchers.
# We don't want to double tag in this case, otherwise, the real tag
# will be overwritten.
TUNE_SEARCHER_WRAPPERS = {
"ConcurrencyLimiter",
"Repeater",
}

TUNE_SCHEDULERS = {
"FIFOScheduler",
"AsyncHyperBandScheduler",
"AsyncHyperBandScheduler",
"MedianStoppingRule",
"HyperBandScheduler",
"HyperBandForBOHB",
"PopulationBasedTraining",
"PopulationBasedTrainingReplay",
"PB2",
"ResourceChangingScheduler",
}


def _find_class_name(obj, allowed_module_path_prefix: str, whitelist: Set[str]):
"""Find the class name of the object. If the object is not
under `allowed_module_path_prefix` or if its class is not in the whitelist,
return "Custom".
Args:
obj: The object under inspection.
allowed_module_path_prefix: If the `obj`'s class is not under
the `allowed_module_path_prefix`, its class name will be anonymized.
whitelist: If the `obj`'s class is not in the `whitelist`,
it will be anonymized.
Returns:
The class name to be tagged with telemetry.
"""
module_path = obj.__module__
cls_name = obj.__class__.__name__
if module_path.startswith(allowed_module_path_prefix) and cls_name in whitelist:
return cls_name
else:
return "Custom"


def tag_air_trainer(trainer: "BaseTrainer"):
from ray.train.trainer import BaseTrainer

assert isinstance(trainer, BaseTrainer)
module_path = trainer.__module__
if module_path.startswith("ray.train"):
trainer_name = trainer.__class__.__name__
if trainer_name not in AIR_TRAINERS:
trainer_name = "Custom"
else:
trainer_name = "Custom"
trainer_name = _find_class_name(trainer, "ray.train", AIR_TRAINERS)
record_extra_usage_tag(TagKey.AIR_TRAINER, trainer_name)


def tag_searcher(searcher: Union["BasicVariantGenerator", "Searcher"]):
from ray.tune.search import BasicVariantGenerator, Searcher

if isinstance(searcher, BasicVariantGenerator):
# Note this could be highly inflated as all train flows are treated
# as using BasicVariantGenerator.
record_extra_usage_tag(TagKey.TUNE_SEARCHER, "BasicVariantGenerator")
elif isinstance(searcher, Searcher):
searcher_name = _find_class_name(
searcher, "ray.tune.search", TUNE_SEARCHERS.union(TUNE_SEARCHER_WRAPPERS)
)
if searcher_name in TUNE_SEARCHER_WRAPPERS:
# ignore to avoid double tagging with wrapper name.
return
record_extra_usage_tag(TagKey.TUNE_SEARCHER, searcher_name)
else:
assert False, (
"Not expecting a non-BasicVariantGenerator, "
"non-Searcher type passed in for `tag_searcher`."
)


def tag_scheduler(scheduler: "TrialScheduler"):
from ray.tune.schedulers import TrialScheduler

assert isinstance(scheduler, TrialScheduler)
scheduler_name = _find_class_name(scheduler, "ray.tune.schedulers", TUNE_SCHEDULERS)
record_extra_usage_tag(TagKey.TUNE_SCHEDULER, scheduler_name)
6 changes: 5 additions & 1 deletion python/ray/tests/test_usage_stats.py
Original file line number Diff line number Diff line change
Expand Up @@ -1197,7 +1197,7 @@ def ready(self):
payload["extra_usage_tags"]["num_actor_tasks"] = "0"
payload["extra_usage_tags"]["num_normal_tasks"] = "0"
payload["extra_usage_tags"]["num_drivers"] = "0"
assert payload["extra_usage_tags"] == {
expected_payload = {
"extra_k1": "extra_v1",
"_test1": "extra_v2",
"_test2": "extra_v3",
Expand All @@ -1215,6 +1215,10 @@ def ready(self):
"gcs_storage": gcs_storage_type,
"dashboard_used": "False",
}
if os.environ.get("RAY_MINIMAL") != "1":
expected_payload["tune_scheduler"] = "FIFOScheduler"
expected_payload["tune_searcher"] = "BasicVariantGenerator"
assert payload["extra_usage_tags"] == expected_payload
assert payload["total_num_nodes"] == 1
assert payload["total_num_running_jobs"] == 1
if os.environ.get("RAY_MINIMAL") == "1":
Expand Down
3 changes: 3 additions & 0 deletions python/ray/train/base_trainer.py
Original file line number Diff line number Diff line change
Expand Up @@ -56,6 +56,9 @@ class BaseTrainer(abc.ABC):
Note: The base ``BaseTrainer`` class cannot be instantiated directly. Only
one of its subclasses can be used.
Note to AIR developers: If a new AIR trainer is added, please update
`air/_internal/usage.py`.
**How does a trainer work?**
- First, initialize the Trainer. The initialization runs locally,
Expand Down
2 changes: 1 addition & 1 deletion python/ray/tune/schedulers/async_hyperband.py
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@ def __init__(
if mode:
assert mode in ["min", "max"], "`mode` must be 'min' or 'max'!"

FIFOScheduler.__init__(self)
super().__init__()
self._reduction_factor = reduction_factor
self._max_t = max_t

Expand Down
2 changes: 1 addition & 1 deletion python/ray/tune/schedulers/hyperband.py
Original file line number Diff line number Diff line change
Expand Up @@ -97,7 +97,7 @@ def __init__(
if mode:
assert mode in ["min", "max"], "`mode` must be 'min' or 'max'!"

FIFOScheduler.__init__(self)
super().__init__()
self._eta = reduction_factor
self._s_max_1 = int(np.round(np.log(max_t) / np.log(reduction_factor))) + 1
self._max_t_attr = max_t
Expand Down
2 changes: 1 addition & 1 deletion python/ray/tune/schedulers/median_stopping_rule.py
Original file line number Diff line number Diff line change
Expand Up @@ -54,7 +54,7 @@ def __init__(
min_time_slice: int = 0,
hard_stop: bool = True,
):
FIFOScheduler.__init__(self)
super().__init__()
self._stopped_trials = set()
self._grace_period = grace_period
self._min_samples_required = min_samples_required
Expand Down
4 changes: 2 additions & 2 deletions python/ray/tune/schedulers/pbt.py
Original file line number Diff line number Diff line change
Expand Up @@ -391,7 +391,7 @@ def __init__(
if mode:
assert mode in ["min", "max"], "`mode` must be 'min' or 'max'."

FIFOScheduler.__init__(self)
super().__init__()
self._metric = metric
self._mode = mode
self._metric_op = None
Expand Down Expand Up @@ -1024,7 +1024,7 @@ def _load_policy(self, policy_file: str) -> Tuple[Dict, List[Tuple[int, Dict]]]:
policy = []
last_new_tag = None
last_old_conf = None
for (old_tag, new_tag, old_step, new_step, old_conf, new_conf) in reversed(
for old_tag, new_tag, old_step, new_step, old_conf, new_conf in reversed(
raw_policy
):
if last_new_tag and old_tag != last_new_tag:
Expand Down
13 changes: 12 additions & 1 deletion python/ray/tune/schedulers/trial_scheduler.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,6 @@
from typing import Dict, Optional

from ray.air._internal.usage import tag_scheduler
from ray.tune.execution import trial_runner
from ray.tune.result import DEFAULT_METRIC
from ray.tune.experiment import Trial
Expand All @@ -8,7 +9,11 @@

@DeveloperAPI
class TrialScheduler:
"""Interface for implementing a Trial Scheduler class."""
"""Interface for implementing a Trial Scheduler class.
Note to Tune developers: If a new scheduler is added, please update
`air/_internal/usage.py`.
"""

CONTINUE = "CONTINUE" #: Status for continuing trial execution
PAUSE = "PAUSE" #: Status for pausing trial execution
Expand All @@ -23,6 +28,9 @@ class TrialScheduler:

_supports_buffered_results = True

def __init__(self):
tag_scheduler(self)

@property
def metric(self):
return self._metric
Expand Down Expand Up @@ -127,6 +135,9 @@ def restore(self, checkpoint_path: str):
class FIFOScheduler(TrialScheduler):
"""Simple scheduler that just runs trials in submission order."""

def __init__(self):
super().__init__()

def on_trial_add(self, trial_runner: "trial_runner.TrialRunner", trial: Trial):
pass

Expand Down
2 changes: 2 additions & 0 deletions python/ray/tune/search/basic_variant.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@
import warnings
import numpy as np

from ray.air._internal.usage import tag_searcher
from ray.tune.error import TuneError
from ray.tune.experiment.config_parser import _make_parser, _create_trial_from_spec
from ray.tune.search.sample import np_random_generator, _BackwardsCompatibleNumpyRng
Expand Down Expand Up @@ -294,6 +295,7 @@ def __init__(
Union[int, "np_random_generator", np.random.RandomState]
] = None,
):
tag_searcher(self)
self._trial_generator = []
self._iterators = []
self._trial_iter = None
Expand Down
5 changes: 5 additions & 0 deletions python/ray/tune/search/searcher.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@
import warnings
from typing import Dict, Optional, List, Union, Any, TYPE_CHECKING

from ray.air._internal.usage import tag_searcher
from ray.tune.search.util import _set_search_properties_backwards_compatible
from ray.util.annotations import DeveloperAPI, PublicAPI
from ray.util.debug import log_once
Expand Down Expand Up @@ -32,6 +33,9 @@ class Searcher:
Not all implementations support multi objectives.
Note to Tune developers: If a new searcher is added, please update
`air/_internal/usage.py`.
Args:
metric: The training result objective value attribute. If
list then list of training result objective value attributes
Expand Down Expand Up @@ -76,6 +80,7 @@ def __init__(
metric: Optional[str] = None,
mode: Optional[str] = None,
):
tag_searcher(self)
self._metric = metric
self._mode = mode

Expand Down
6 changes: 6 additions & 0 deletions src/ray/protobuf/usage.proto
Original file line number Diff line number Diff line change
Expand Up @@ -118,4 +118,10 @@ enum TagKey {
// Name of AIR trainer, or "Custom" if user-defined.
// Example: "TorchTrainer"
AIR_TRAINER = 500;
// Name of Tune search algorithm or "Custom" if user-defined.
// Example: "TuneBOHB", "BasicVariantGenerator"
TUNE_SEARCHER = 501;
// Name of Tune scheduler algorithm or "Custom" if user-defined.
// Example: "FIFOScheduler"
TUNE_SCHEDULER = 502;
}

0 comments on commit d706309

Please sign in to comment.