Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Custom Metrics] Enabled support for logging of numerical metrics #5389

Merged
merged 12 commits into from
Feb 25, 2022

Conversation

MarkYHZhang
Copy link
Contributor

@MarkYHZhang MarkYHZhang commented Feb 18, 2022

Signed-off-by: Mark Zhang mark.zhang@databricks.com

What changes are proposed in this pull request?

Added support for users to pass in custom metric functions into mlflow.evaluate. This PR is a part of more PRs to come for custom metrics support. Specifically, this PR focuses on enabling logging of numerical metrics generated by user's custom metric functions into MLflow.

How is this patch tested?

Wrote unit tests located in mlflow/tests/models/test_default_evaluator.py

Does this PR change the documentation?

Updates to the documentation will be done once the entire feature of custom metrics is complete.

  • No. You can skip the rest of this section.
  • Yes. Make sure the changed pages / sections render correctly by following the steps below.
  1. Check the status of the ci/circleci: build_doc check. If it's successful, proceed to the
    next step, otherwise fix it.
  2. Click Details on the right to open the job page of CircleCI.
  3. Click the Artifacts tab.
  4. Click docs/build/html/index.html.
  5. Find the changed pages / sections and make sure they render correctly.

Release Notes

Is this a user-facing change?

  • No. You can skip the rest of this section.
  • Yes. Give a description of this change to be included in the release notes for MLflow users.

Enables MLflow tracking of custom metrics produced by user-provided metric functions.

(Details in 1-2 sentences. You can just refer to another PR with a description if this PR is part of a larger change.)

What component(s), interfaces, languages, and integrations does this PR affect?

Components

  • area/artifacts: Artifact stores and artifact logging
  • area/build: Build and test infrastructure for MLflow
  • area/docs: MLflow documentation pages
  • area/examples: Example code
  • area/model-registry: Model Registry service, APIs, and the fluent client calls for Model Registry
  • area/models: MLmodel format, model serialization/deserialization, flavors
  • area/projects: MLproject format, project running backends
  • area/scoring: MLflow Model server, model deployment tools, Spark UDFs
  • area/server-infra: MLflow Tracking server backend
  • area/tracking: Tracking Service, tracking client APIs, autologging

Interface

  • area/uiux: Front-end, user experience, plotting, JavaScript, JavaScript dev server
  • area/docker: Docker use across MLflow's components, such as MLflow Projects and MLflow Models
  • area/sqlalchemy: Use of SQLAlchemy in the Tracking Service or Model Registry
  • area/windows: Windows support

Language

  • language/r: R APIs and clients
  • language/java: Java APIs and clients
  • language/new: Proposals for new client languages

Integrations

  • integrations/azure: Azure and Azure ML integrations
  • integrations/sagemaker: SageMaker integrations
  • integrations/databricks: Databricks integrations

How should the PR be classified in the release notes? Choose one:

  • rn/breaking-change - The PR will be mentioned in the "Breaking Changes" section
  • rn/none - No description will be included. The PR will be mentioned only by the PR number in the "Small Bugfixes and Documentation Updates" section
  • rn/feature - A new user-facing feature worth mentioning in the release notes
  • rn/bug-fix - A user-facing bug fix worth mentioning in the release notes
  • rn/documentation - A user-facing documentation change worth mentioning in the release notes

Signed-off-by: Mark Zhang <mark.zhang@databricks.com>
@github-actions github-actions bot added area/tracking Tracking service, tracking client APIs, autologging rn/feature Mention under Features in Changelogs. labels Feb 18, 2022
@MarkYHZhang
Copy link
Contributor Author

Hey @apurvakoti @dbczumar @jinzhang21 @WeichenXu123, can you all take a look? It seems like I don't have permission to add reviewers. So I'm tagging you all here. Thanks!

Copy link
Collaborator

@jinzhang21 jinzhang21 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Added some initial comment on the API. I'll review the rest later today.

mlflow/models/evaluation/base.py Show resolved Hide resolved
mlflow/models/evaluation/base.py Outdated Show resolved Hide resolved
Signed-off-by: Mark Zhang <mark.zhang@databricks.com>
Signed-off-by: Mark Zhang <mark.zhang@databricks.com>
Copy link
Collaborator

@apurva-koti apurva-koti left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Some small comments, will do a full review again

mlflow/models/evaluation/base.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Show resolved Hide resolved
Signed-off-by: Mark Zhang <mark.zhang@databricks.com>
Signed-off-by: Mark Zhang <mark.zhang@databricks.com>
…. Since DefaultEvaluator saves Dataframe as CSV. And CSV files can also be viewed in the MLflow UI

Signed-off-by: Mark Zhang <mark.zhang@databricks.com>
Copy link
Collaborator

@apurva-koti apurva-koti left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks good! Let's use [] as the default argument for custom_metrics rather than None.

mlflow/models/evaluation/base.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/base.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/base.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Outdated Show resolved Hide resolved
@MarkYHZhang
Copy link
Contributor Author

MarkYHZhang commented Feb 23, 2022

Looks good! Let's use [] as the default argument for custom_metrics rather than None.

Thanks! I think we should avoid using mutable default argument in python due to issues described here. Although, I could initialize it to a list inside the constructor, but considering other default arguments like

feature_names: list = None,

are not initialized, I think it might be good idea just to leave it as is.

@apurva-koti
Copy link
Collaborator

Ah good catch @MarkYHZhang ! Let's keep it None then.

Signed-off-by: Mark Zhang <mark.zhang@databricks.com>
Copy link
Collaborator

@apurva-koti apurva-koti left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM once all comments addressed!

mlflow/models/evaluation/default_evaluator.py Outdated Show resolved Hide resolved
tests/models/test_default_evaluator.py Outdated Show resolved Hide resolved
Signed-off-by: Mark Zhang <mark.zhang@databricks.com>
Copy link
Collaborator

@dbczumar dbczumar left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@MarkYHZhang This looks great! Left a few small comments - should be ready to merge soon!

…ate_custom_metric function. Modified docstring to only include metrics support

Signed-off-by: Mark Zhang <markzhang.inbox@gmail.com>
Copy link
Collaborator

@dbczumar dbczumar left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@MarkYHZhang Do we plan to add an example to the examples section or extend one of the existing examples? Can we do that as part of this PR?

Copy link
Collaborator

@dbczumar dbczumar left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM when small docs nits have been addressed and example has been included. Thanks @MarkYHZhang ! Awesome stuff!

mlflow/models/evaluation/base.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/base.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Outdated Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Show resolved Hide resolved
…urn validation. Plus a few minor stylistic changes

Signed-off-by: Mark Zhang <markzhang.inbox@gmail.com>
mlflow/models/evaluation/base.py Show resolved Hide resolved
mlflow/models/evaluation/base.py Show resolved Hide resolved
mlflow/models/evaluation/default_evaluator.py Outdated Show resolved Hide resolved
tests/models/test_default_evaluator.py Show resolved Hide resolved
tests/models/test_default_evaluator.py Show resolved Hide resolved
…builtin_metrics per custom metric function

Signed-off-by: Mark Zhang <markzhang.inbox@gmail.com>
@MarkYHZhang MarkYHZhang merged commit cb7b361 into mlflow:master Feb 25, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area/tracking Tracking service, tracking client APIs, autologging rn/feature Mention under Features in Changelogs.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

5 participants