[tabular] Parallel Model Training #4215
Labels
enhancement
New feature or request
feature: distributed
Related to Distributed AutoGluon
module: tabular
Milestone
Related: #4213
We should add support for parallel model training via ray that goes beyond the currently implemented parallel fold model fitting and parallel HPO.
Given a portfolio of 100 models, we should be able to train multiple of them at the same time if we have sufficient resources.
This should also be extended to work with a distributed cluster.
Pseudocode Example
Current logic:
Code that does this in mainline:
autogluon/core/src/autogluon/core/trainer/abstract_trainer.py
Lines 2478 to 2504 in 7b782df
Proposed logic:
The text was updated successfully, but these errors were encountered: