-
Notifications
You must be signed in to change notification settings - Fork 28
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
100 million data results #14
Comments
How many columns in your data set? Is it sparse? If so what is the percentage of missing values. |
Subsample of the airline dataset, 8 columns (though after one-hot-encoding ~700 sparse), no missing values. |
Also h2o and catboost (CPU) on r4.8xlarge (32 cores, 1 NUMA, only physical cores/no HT):
UPDATE 2020-09-08 catboost: run time 930sec, RAM data ~5GB, RAM train max ~50GB, RAM after ~5GB, AUC 0.7358616 CPU:
RAM train end/gc is when training ends and then after calling |
GPU on p3.8xlarge (4 GPUs, but only 1 GPU used!) (needed larger than p3.2xlarge with 1 GPU because data reading/prep did not fit in 60GB RAM)
GPU:
|
4x GPU on p3.8xlarge:
TODO: XGBoost must be compiled with NCCL to use more than one GPU. catboost crashes (OOM) even on 4 GPUs. Note: lightgbm and h2o xgboost don't support multiple GPUs currently. |
All my prev results in 1 place: 100M records CPU (r4.8xlarge):
GPU (Tesla V100):
|
CPU on m5 (faster than r4 above). m5 is prob the fastest CPU on EC2 for this because for larger data more cores matter most than high frequency CPU (m5>c5, see #13 (comment))
|
100M records CPU (m5.12xlarge):
GPU (Tesla V100):
|
catboost crashes for larger sizes on the GPU (mem 16GB):
|
2020-09-08 update: catboost GPU still crashes (runs out of GPU memory):
|
Using hardware from here: #12
Using dmlc/xgboost@84d992b and microsoft/LightGBM@5ece53b
100M obtained using 10x 10m data.
CPU:
1x Quadro P1000:
4x Quadro P1000:
RAM usage:
LightGBM: 2739 MB on GPU
xgboost 1 GPU: CRASH
xgboost 4 GPUs: 2077 MB on each GPU
The text was updated successfully, but these errors were encountered: