Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ENH: Deprecate WindowDataset #555

Closed
Tracked by #612
NickleDave opened this issue Jul 29, 2022 · 3 comments
Closed
Tracked by #612

ENH: Deprecate WindowDataset #555

NickleDave opened this issue Jul 29, 2022 · 3 comments

Comments

@NickleDave
Copy link
Collaborator

NickleDave commented Jul 29, 2022

Currently training is tightly coupled to the WindowDataset abstraction.
A clear drawback of this is that it prevents us from training in any other way, e.g. feeding in an entire vocalization (e.g. a birdsong bout) as one sample in a batch.
It's also the case that the logic wrapped up in this function is super hard to read, there's a lot of array-oriented programming that can lead to subtle errors, see for example #169 #213 #217 #219 #220

So this issue formalizes the idea of deprecating WindowDataset.
Instead each sample in a dataset will now be one vocalization, however it is defined for that particular dataset. Typically this will map to one audio file or one spectrogram, e.g. for birdsong one bout of song. In other words, one row from a dataframe representing a dataset, as produced by vocles,

@NickleDave NickleDave changed the title ENH: Replace WindowDataset with RandomCrop transform ENH: Deprecate WindowDataset; add RandomCrop transform Jul 29, 2022
@NickleDave NickleDave changed the title ENH: Deprecate WindowDataset; add RandomCrop transform ENH: Deprecate WindowDataset Jul 29, 2022
@NickleDave
Copy link
Collaborator Author

For more detail on what we get from the WindowDataset abstraction, and how we will achieve the same thing while deprecating it, see #556 and #557.

Key things are:

@NickleDave
Copy link
Collaborator Author

Not going to do this for now, as discussed in #651 and issues linked there. Closing

@NickleDave
Copy link
Collaborator Author

A clear drawback of this is that it prevents us from training in any other way, e.g. feeding in an entire vocalization (e.g. a birdsong bout) as one sample in a batch.
It's also the case that the logic wrapped up in this function is super hard to read, there's a lot of array-oriented programming that can lead to subtle errors, see for example #169 #213 #217 #219 #220

So this issue formalizes the idea of deprecating WindowDataset.
Instead each sample in a dataset will now be one vocalization, however it is defined for that particular dataset. Typically this will map to one audio file or one spectrogram, e.g. for birdsong one bout of song. In other words, one row from a dataframe representing a dataset, as produced by vocles,

We will still add some sort of abstraction like FileDataset where file can be an audio file or spectrogram file

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant