-
Notifications
You must be signed in to change notification settings - Fork 111
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[RMP] Support Tree Ranking Models (like XGBoost) in Merlin Models and Systems #105
Comments
@benfred , could you map your XGboost tickets here ? |
@marcromeyn Re-assigned this to you for the time being since it sounds like you'll be working with our new hire on both the Models and Systems sides. Once he's in our GH org, might make him the lead on this though? |
An update on this issue. Merlin Models We have a first version of the XGBoost API in Merlin Models from release 22.06. Once we have an example that part can be considered complete, for the purposes of this issue. With ongoing support for feature improvements via other issues. Merlin Systems Operators have been added to Merlin Systems to enable XGBoost, LightGBM, Scikit-Learn (Random Forest), cuML (Random Forest) models being added as part of a serving ensemble in Triton. There are a couple of small issues being addressed (operator outputs) to make this usable from the next release 22.07. NVTabular The comment "NVTabular - Operators for batch prediction with these models" in the description I think may be out of scope of this issue. At least, it's unclear to me what the relationship with NVTabular is in this context. In terms of batch prediction for the XGBoost integration in Merlin Models. This is supported as part of the predict method, since we're using the Dask version of XGBoost, which can be used in a distributed setting. In terms of a common pattern for running evaluation / which will involve batch/distributed prediction. The ongoing work in this area may be covered by #407 or #405 |
@radekosmulski , could you please add the ticket tracking the examples here in the top |
Closing this ticket. Example is pending and is planned for 22.08 |
Problem
Gradient-boosted decision trees (GBDTs) are commonly used in the industry as part of the scoring phase of recommender systems. Supporting serving of these models and integrating with the Merlin ecosystem will help facilitate usage of these models in these systems.
The Triton Inference Server has a backend called FIL (Forest Inference Library) to facilitate GPU accelerated serving of these models.
Goals
Constraints
Starting Point
Merlin-models (Data Scientist)
NVTabular (Data Scientist)
Merlin-systems (Product Engineer)
Supports the following models: XGBoost, LightGBM, Scikit-Learn (Random Forest), cuML (Random Forest)
Examples and Docs (Everyone)
Aha! Link: https://nvaiinfa.aha.io/features/MERLIN-828
The text was updated successfully, but these errors were encountered: