Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Memory error for large datasets #26

Open
jf2016GH opened this issue Nov 11, 2019 · 3 comments
Open

Memory error for large datasets #26

jf2016GH opened this issue Nov 11, 2019 · 3 comments

Comments

@jf2016GH
Copy link

$time python nclist.py

Subject file: chainOrnAna1.bed (1.956864), Query file: chainRn4.bed (2.350965)
Building NCLS took 0.36449146270751953
Traceback (most recent call last):
File "nclist.py", line 36, in
time_ncls(s, q)
File "nclist.py", line 22, in time_ncls
n.all_overlaps_both(q[1].values, q[2].values, q.index.values)
File "ncls/src/ncls.pyx", line 278, in ncls.src.ncls.NCLS.all_overlaps_both
File "ncls/src/ncls.pyx", line 321, in ncls.src.ncls.NCLS.all_overlaps_both
File "/home/john/anaconda3/lib/python3.6/site-packages/numpy/core/fromnumeric.py", line 1320, in resize
a = concatenate((a,) * n_copies)
MemoryError

@endrebak
Copy link
Collaborator

Thanks. This happens when the number of enumerated intervals is very large. I’ll have a look :)

@endrebak
Copy link
Collaborator

While I cannot do anything to prevent this from happening, I can write a user-friendly error message instead. I will do that :)

If this means that the NCLS fails on large data where the AIList this is a very interesting result for me :)

@jf2016GH
Copy link
Author

jf2016GH commented Nov 14, 2019 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants