Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Parallelism fix to reduce errors on large datasets like scRNAseq #92

Merged
merged 2 commits into from
Mar 6, 2020

Conversation

TheAustinator
Copy link
Contributor

This fix circumvents pickling errors generated by gseapy/algorithm.py line 374, which large datasets (e.g. scRNAseq) are particularly vulnerable to. The gsea tensor computation can generate blocks which are larger than 4 GB, which do not fit in the i struct formatter used by python's build in multiprocessing module.

@zqfang
Copy link
Owner

zqfang commented Mar 6, 2020

Thank you very much for your patch. I' will check these more detailly later.

@zqfang zqfang merged commit adb3742 into zqfang:master Mar 6, 2020
@zqfang
Copy link
Owner

zqfang commented Apr 19, 2020

refer to a bug #94

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants