You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Making this issue to compile a list of minor tweaks we could/should make to the data generation process:
Log the parameters used to generate the cached data to the directory where it is saved
Instead of specifying the number of batches in the config, only require total number of images desired and the batch size, and the number of batches can be calculated at runtime
Similarly, remove checks for certain amount of data - just load whatever is given
Don't require that all three splits are present when loading CachedSimulatedDataset - we might want to generate just a separate test split and test on that without train/validation data for example. We can instead raise an exception when the user tries to use data that wasn't loaded (e.g. in train/val/test_dataloader)
Discussed with @zhixiangteoh - the reason for using num_batches instead of a number of images is 1) to keep things in multiples of batch size without ceil/flooring, and 2) consistency with SimulatedDataset, so the second and third points aren't necessary.
Making this issue to compile a list of minor tweaks we could/should make to the data generation process:
Instead of specifying the number of batches in the config, only require total number of images desired and the batch size, and the number of batches can be calculated at runtimeSimilarly, remove checks for certain amount of data - just load whatever is givenbliss/bliss/simulator/simulated_dataset.py
Lines 120 to 125 in 760e909
CachedSimulatedDataset
- we might want to generate just a separate test split and test on that without train/validation data for example. We can instead raise an exception when the user tries to use data that wasn't loaded (e.g. intrain/val/test_dataloader
)bliss/bliss/simulator/simulated_dataset.py
Lines 127 to 130 in 760e909
The text was updated successfully, but these errors were encountered: