-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Slow chunking the text file #9
Comments
Can you perhaps try this? |
This one took 6.88s seem to be faster.🤔 |
Just make sure that the embedding model you used to generate the vector collection / snapshot is the same as the one rag-api-server starts with. |
I'm not quite sure which line i've to check, I follow step from readme which is
and
it should be same? |
You started the rag api server with all-MiniLM-L6-v2-ggml-model-f16.gguf So, the command you used to create the embeddings should also be all-MiniLM-L6-v2-ggml-model-f16.gguf If you just ran the steps in the docs, you should be fine. |
Yes i just run 100% steps in the docs(for many times by now), but it's still slow. I think I miss something pretty obvious 🤔. |
after try step from readme
It took 590824.84 ms = nearly 1 minute for only chunking 306 lines (91KB) file on m3 max.
This is just me or I miss some flag?
The text was updated successfully, but these errors were encountered: