Would love to see this capability in pgvector: https://qdrant.tech/articles/binary-quantization/
Essentially, BQ converts any vector embedding of floating point numbers into a vector of binary or boolean values.
In exchange for reducing our 32 bit embeddings to 1 bit embeddings we can see up to a 40x retrieval speed up gain!
One of the reasons vector search still works with such a high compression rate is that these large vectors are over-parameterized for retrieval. This is because they are designed for ranking, clustering, and similar use cases, which typically need more information encoded in the vector.
Would love to see this capability in pgvector: https://qdrant.tech/articles/binary-quantization/
Essentially, BQ converts any vector embedding of floating point numbers into a vector of binary or boolean values.