You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When building indexes of varying sizes I ran into some issues with some of the larger sizes..
Here's what my index creation code looks like:
# imagine `vectors` is an ndarray with multiple vectors of dimension 1728 ...num_dimensions=vectors[0].shape[0]
index=Index(Space.Cosine, num_dimensions=num_dimensions)
index.add_items(vectors)
index.save(filename)
And my test code looks like this:
# imagine `vector` is a sample query vector of (matching) dimension 1728index=Index.load(filename)
index.query(vector, k=200)
This works fine when vectors is of cardinality 10k, 50k, 100k, 500k, and 1M ...
but when vectors has 5M or 10M vectors in it, index creation runs fine, but upon querying ...
hey @loisaidasam thank you for reporting, that result definitely shouldn't be happening! We've seen similar behavior when accidentally adding vectors with NaN values into the index, but I don't think the issue is related. It's more likely due to a race condition somewhere so we'll investigate this and get back to you!
When building indexes of varying sizes I ran into some issues with some of the larger sizes..
Here's what my index creation code looks like:
And my test code looks like this:
This works fine when
vectors
is of cardinality 10k, 50k, 100k, 500k, and 1M ...but when
vectors
has 5M or 10M vectors in it, index creation runs fine, but upon querying ...I tried creating the index with slices of the same
vectors
array of size 1M:and it seems I can query this index just fine. Maybe some sort of limitation with the
add_items()
function?The text was updated successfully, but these errors were encountered: