Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Transfer large synthetic data approach from dask-glm #67

Closed
PeterDSteinberg opened this Issue Oct 25, 2017 · 1 comment

Comments

Projects
None yet
2 participants
@PeterDSteinberg
Copy link
Collaborator

PeterDSteinberg commented Oct 25, 2017

See make_poisson and similar functions in dask-glm's datasets.py. Transfer that approach into dask-ml's datasets.py so larger synthetic data sets can be created. Currently the approach in dask-ml's datasets.py just wraps sklearn.datasets.

@mrocklin

This comment has been minimized.

Copy link
Member

mrocklin commented Oct 26, 2017

I also ran into this just last night

TomAugspurger added a commit to TomAugspurger/dask-ml that referenced this issue May 22, 2018

REF: Large dataset generators
Change the methods in dask_ml.datasets to not overwhelm a single
machine when generating large random datasets.

Closes dask#67

TomAugspurger added a commit to TomAugspurger/dask-ml that referenced this issue May 24, 2018

REF: Large dataset generators
Change the methods in dask_ml.datasets to not overwhelm a single
machine when generating large random datasets.

Closes dask#67

TomAugspurger added a commit that referenced this issue May 24, 2018

REF: Large dataset generators (#173)
Change the methods in dask_ml.datasets to not overwhelm a single
machine when generating large random datasets.

Closes #67

TomAugspurger pushed a commit to TomAugspurger/dask-ml that referenced this issue Jun 28, 2018

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
You can’t perform that action at this time.