dense_retriever -- MemoryError: std::bad_alloc #33
Comments
Hi Aaron, |
How much RAM is required for the flat index? |
The flat index setup showed max 95 GB ram consumption for the entire process (i.e index+ wikipedia passages data in memory, etc.). |
Hmm, I'm confused. Why is 128GB RAM not enough for the flat index setup, if the flat index setup only requires 95GB RAM? |
I have no idea why 128 it not enough, I run it on 512 GB server and measured the highest RES consumption. Virtual max was 135. |
I've checked that, prior to running |
How many CPU cores does the server have and how long does it take to run dense_retriever with the server? |
It has 64 cores
…On Sun, Aug 9, 2020 at 6:48 AM Sohee Yang ***@***.***> wrote:
I run it on 512 GB server and measured the highest RES consumption.
Virtual max was 135.
How many CPU cores does the server have?
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
<#33 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ABDJ5QCWYZHJKLWIWUCCNXDR72SL3ANCNFSM4PGKTZZQ>
.
|
Thank you :D |
Seems like this can be closed now. |
I have the same issue of MemoryError: std::bad_alloc when I use dense-retriever.py, the size of my indexing is 50G and my server has 86G Memory. I changed the function iterate_encoded_files in faiss_indexers.py, specifically as follows, def iterate_encoded_files(vector_files: list) -> Iterator[Tuple[object, np.array]]: the function loads all the indexing at once (since 50G is less than 86G, it is fine), so that DenseIndexer index_data all indexing at once rather than using a buffer. now MemoryError: std::bad_alloc has gone. |
Hi! It seems that no matter what value I set
index_buffer
to, I get the following error when runningdense_retriever.py
:For reference, the machine I'm running this on has 128GB RAM, but it doesn't seem to be enough. Could you please help me with this issue? Thanks!
The text was updated successfully, but these errors were encountered: