Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add n_jobs as parameter for AutoML #403

Merged
merged 21 commits into from Feb 26, 2020
Merged

Add n_jobs as parameter for AutoML #403

merged 21 commits into from Feb 26, 2020

Conversation

jeremyliweishih
Copy link
Contributor

@jeremyliweishih jeremyliweishih commented Feb 20, 2020

Fixes #382.

@codecov
Copy link

codecov bot commented Feb 20, 2020

Codecov Report

Merging #403 into master will not change coverage.
The diff coverage is n/a.

Impacted file tree graph

@@          Coverage Diff           @@
##           master    #403   +/-   ##
======================================
  Coverage    97.4%   97.4%           
======================================
  Files         103     103           
  Lines        3239    3239           
======================================
  Hits         3155    3155           
  Misses         84      84
Impacted Files Coverage Δ
evalml/automl/auto_regression_search.py 100% <ø> (ø) ⬆️
evalml/automl/auto_classification_search.py 100% <ø> (ø) ⬆️

Continue to review full report at Codecov.

Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 2d45f6a...2e87415. Read the comment docs.

@jeremyliweishih jeremyliweishih changed the title [WIP] Add n_jobs as parameter for AutoML Add n_jobs as parameter for AutoML Feb 20, 2020
Copy link
Contributor

@angela97lin angela97lin left a comment

Yay, looks good to me! 🚢

@jeremyliweishih jeremyliweishih self-assigned this Feb 21, 2020
Copy link
Collaborator

@dsherry dsherry left a comment

LGTM 👍

Unit testing suggestions: Could we add some sort of unit test which passes in a bad value for n_jobs and makes sure we get an exception we like? Could we add a unit test which verifies n_jobs is actually passed into the transformers/estimators, ideally using mock transformers/estimators?

@jeremyliweishih jeremyliweishih requested a review from dsherry Feb 24, 2020
Copy link
Collaborator

@dsherry dsherry left a comment

LGTM, left one test suggestion

@jeremyliweishih jeremyliweishih requested a review from dsherry Feb 25, 2020
Copy link
Collaborator

@dsherry dsherry left a comment

Left two suggestions, good to merge after that

evalml/pipelines/pipeline_base.py Show resolved Hide resolved
evalml/tests/pipeline_tests/test_pipelines.py Show resolved Hide resolved
@jeremyliweishih jeremyliweishih requested a review from dsherry Feb 26, 2020
@jeremyliweishih jeremyliweishih merged commit d50cfde into master Feb 26, 2020
2 checks passed
@angela97lin angela97lin mentioned this pull request Mar 9, 2020
@dsherry dsherry deleted the 382_n_jobs branch May 26, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Have AutoBase pass n_jobs and random_state to pipelines
3 participants