Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

batch_size control #4

Closed
davidsebfischer opened this issue Jul 31, 2018 · 5 comments
Closed

batch_size control #4

davidsebfischer opened this issue Jul 31, 2018 · 5 comments
Assignees

Comments

@davidsebfischer
Copy link
Contributor

Allow user to set batch size. Do this as part of training strategy? You can leave defaults as they are now.

@falexwolf
Copy link

We had a long discussion about the notions chunks and batches? In the end, we settled for chunks as this is won't be confused with experimental batches and is analogous with h5py, pytables, xarray and zarr naming conventions on the file level.

So, in anndata and Scanpy, there is a parameter chunk_size, not batch_size... Maybe a consideration...

@davidsebfischer
Copy link
Contributor Author

Agreed, that makes sense.

@davidsebfischer
Copy link
Contributor Author

Opened a separate issue for batch vs chunk naming.

@Hoeze
Copy link
Contributor

Hoeze commented Jul 31, 2018

The user can change the batch size by specifying it in __init__:
Estimator(input_data, batch_size=500)

Sadly, this cannot be changed at runtime, since this would require to edit the tensorflow graph.

Nevertheless, this is an option which should be accessible in diffxpy:
theislab/diffxpy#8

@Hoeze Hoeze closed this as completed Jul 31, 2018
@davidsebfischer
Copy link
Contributor Author

Yes, if this is set in TrainingStrategy, changes in batch size at run time are not necessary anyway.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants