Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[dask] Complete features in regressor and classifier. #6471

Merged
merged 1 commit into from Dec 13, 2020

Conversation

trivialfis
Copy link
Member

  • Add eval_metric.
  • Add custom objective.
  • Add callback.
  • Add feature weights.

@hcho3
Copy link
Collaborator

hcho3 commented Dec 5, 2020

Is this blocking?

@trivialfis
Copy link
Member Author

No. It's a PR for reaching feature parity with native model.

@trivialfis
Copy link
Member Author

trivialfis commented Dec 5, 2020

Close #6121 .
Close #6468 .

@trivialfis
Copy link
Member Author

There are still some features like ntree_limit, xgb_model arguments in sklearn wrapper, which I do not indent to support. Also, some additional attributes like n_features_in_ will be considered in future PRs as right now I don't have a complete list of them. We will need to run scikit-learn's model interface check to declare feature complete.

@pseudotensor
Copy link
Contributor

There are still some features like ntree_limit, xgb_model arguments in sklearn wrapper, which I do not indent to support. Also, some additional attributes like n_features_in_ will be considered in future PRs as right now I don't have a complete list of them. We will need to run scikit-learn's model interface check to declare feature complete.

FYI ntree_limit = 0 usually has to be passed for gblinear during predict or one gets an assertion hit, so if no ntree_limit support is done it would have to automaticlally pass that properly.

@trivialfis
Copy link
Member Author

@pseudotensor gblinear is not used in distributed setting.

@trivialfis
Copy link
Member Author

trivialfis commented Dec 5, 2020

@pseudotensor We added a new feature for slicing tree model:

booster[0: 3]

will get the trees in [0, 3) boosting rounds. See https://xgboost.readthedocs.io/en/latest/python/model.html .

@codecov-io
Copy link

codecov-io commented Dec 5, 2020

Codecov Report

Merging #6471 (4d57fee) into master (d6386e4) will decrease coverage by 0.04%.
The diff coverage is 79.16%.

Impacted file tree graph

@@            Coverage Diff             @@
##           master    #6471      +/-   ##
==========================================
- Coverage   80.13%   80.09%   -0.05%     
==========================================
  Files          13       13              
  Lines        3513     3531      +18     
==========================================
+ Hits         2815     2828      +13     
- Misses        698      703       +5     
Impacted Files Coverage Δ
python-package/xgboost/dask.py 80.85% <79.16%> (-0.29%) ⬇️

Continue to review full report at Codecov.

Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update d6386e4...4d57fee. Read the comment docs.

* Add eval_metric.
* Add callback.
* Add feature weights.
* Add custom objective.
@trivialfis
Copy link
Member Author

@hcho3 @RAMitchell please help reviewing.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

5 participants