Capacity Planning while using GPU based Indices #37632
Unanswered
rohitreddy1698
asked this question in
Q&A and General discussion
Replies: 2 comments 5 replies
-
I am using the milvus-2.4.9-gpu version of the docker image. |
Beta Was this translation helpful? Give feedback.
5 replies
-
@rohitreddy1698 I am currently working on implementing this feature, and if everything goes smoothly, I should be able to finish it by next week. Thank you for your patience. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello,
I am trying to compare the performances of GPU vs CPU based indices in Milvus and have Milvus setup on GKE for the same.
I have deployed Milvus using the Milvus Operator.
I am trying to do the benchmarking using the Zilliz Vector DB Bench tool : https://github.com/zilliztech/VectorDBBench
I have
4 : e2-highmen-16 nodes ( for other components )
30 : n1-highmen-8 nodes ( 25 for queyrnode and 5 for indexnode ) with 2 T4 GPUs ( 16 GiB memory ) per node
The data set I am using is 100M Laion dataset with 768 vector dimension. I am trying to test pure Search performance without any fiters.
I have a total of 25 * 2 * 16 = 800 GiB of GPU memory , which should be sufficient for GPU_IVF_FLAT and IVF_FLAT , but I am getting the following failed to deserialise error :
Please help me understand the sizes taken up by the GPU indices and help me get over this issue.
Beta Was this translation helpful? Give feedback.
All reactions