-
Notifications
You must be signed in to change notification settings - Fork 2.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: Milvus standalone service cannot start #31858
Comments
I think the etcd is a bit slow and milvus components lost the keepalive with etcd service. /assign @0215Arthur |
![]() |
what about your etcd disk? |
@0215Arthur |
Yes. The etcd share the disk with milvus node. (Just standalone) |
mainly because milvus eatup all iops and ebs rate limiting your disk.
|
@xiaofan-luan Thanks for your help very much. |
we use local nvme ssd. One possible solution is to buy 10 GP3 disk and raid them together |
Ok. Besides, how to setup the config of etcd disk to separate it from the milvus node in standalone mode?
|
you need to mount at least 2 disk. etcd data dir has separate configs than local dir. If you can use zilliz cloud for diskann that would save you a lot of troubles. We hit into similar issues every day that's why we got the knowledge about how to handle those |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
I am having similar issues trying to set up Milvus Standalone with 10k vectors of dim=1024. Disk is NVME, I get 160k
|
you need to check your system time. seems that etcd and milvus has a large clock skew. this could cause serious issue |
Isn't etcd being set up by the same container that does Milvus? It's a standalone deployment -- single node, single container, everything in it. In which case, could they be operating on different times? Also, I just ran more tests — using vectors of dim=1024, if I try to index 1023 vectors it works, but if I try 1024 it hangs:
|
I doubt that it is caused by building diskann index, because Milvus only builds index when there are 1024 or more vectors in the collection. please file a new issue and offer full milvus logs for invesigation. @nelsonspbr |
Is there an existing issue for this?
Environment
Current Behavior
I have built a collection (diskann index) with 12M embeddings (1536 dim), and I move the volume to the new machine with higher hardware config. But the milvus standalone cannot start in the new machine.
And I use the
fio
to check the disk performance (EBS gp3 disk): IOPS 850full log: milvus.ec2.log
Expected Behavior
No response
Steps To Reproduce
No response
Milvus Log
No response
Anything else?
No response
The text was updated successfully, but these errors were encountered: