-
Notifications
You must be signed in to change notification settings - Fork 21.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.
Already on GitHub? Sign in to your account
Shared Dataset Functionality #24915
Comments
An additional wish: unify transforms interface. One idea: make them regular autograd functions / modules. This enables writing them more shared between domains (imagine various forms of cutout / warps, synthetic noise), moving them from cpu to gpu, performing them at other parts of the model, etc |
all datasets are not moving to core pytorch. If you want, create a |
One of the plans is to keep the raw data links and utils in pytorch/data and have the preprocessing func and associated ops in each domain. |
cc @ssnl - what is your take on this? |
I'm wonder what functionalities are planned for For example,
sounds similar to
sounds doable just with the current data loading infra. |
@soumith Any thoughts on my suggestion about |
Yes, I agree, and torchvision is going to be doing this, see pytorch/vision#1375 |
Thoughts from offline chat with @taylorgordon20 How to help in debugging in a natural python way?
Wishes?
|
馃殌 Feature
We want to build a unified data pipeline interface that offers building blocks for others to build on with the following objectives:
Motivation
Additional Information
torchdataDatasets:
Dataloader:
Features:
cc @ssnl @fmassa @zhangguanheng66 @vincentqb @mrshenli
The text was updated successfully, but these errors were encountered: