You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I tested SDV single tabular models and my training data are of the shape (10000, 7)
This actually is a pretty small dataset. As a highlight, we use all the default hyperparameter settings.
GaussianCopula is quick just in few minutes.
TVAE returned the results around 5.5 hours.
Copular GAN and CTGAN are slow. I've waited for more than 20 hours but it is still running.
(1)
I am thinking is there a way to simply pass a parameter or do some setting to allow the GAN model training with GPU?
(2)
Due to the long training process, I would highly suggest allowing us track the progress of the training. For example, you can always report how many epochs have been trained as well as the total expected training epochs.
Also, a more friendly way is to pass a parameter to the model so that we can dump the log file out.
The text was updated successfully, but these errors were encountered:
Something strange must be happening there, because by default all the Deep Learning models try to use any CUDA device that is available to them.
Can you check whether, torch.cuda.is_available() returns True, and whether nvidia-smi reports any GPU usage while CTGAN (or any of the other models) is running?
WRT the parameters that you suggest (1) and (2), both of them are already available (you can see them in the API Reference):
For (1) all the DL models have a cuda argument, which defaults to True and allows you to either enable/disable CUDA usage, or even explicitly tell the model which CUDA device to use (in case there is more than one available)
For (2) the CTGAN-based models have a verbose parameter, which defaults to False, that if set to True will make CTGAN print the progress and losses at every epoch. Notice that this one is not available on TVAE yet, though.
General problem description:
I tested SDV single tabular models and my training data are of the shape (10000, 7)
This actually is a pretty small dataset. As a highlight, we use all the default hyperparameter settings.
GaussianCopula is quick just in few minutes.
TVAE returned the results around 5.5 hours.
Copular GAN and CTGAN are slow. I've waited for more than 20 hours but it is still running.
(1)
I am thinking is there a way to simply pass a parameter or do some setting to allow the GAN model training with GPU?
(2)
Due to the long training process, I would highly suggest allowing us track the progress of the training. For example, you can always report how many epochs have been trained as well as the total expected training epochs.
Also, a more friendly way is to pass a parameter to the model so that we can dump the log file out.
The text was updated successfully, but these errors were encountered: