-
Hi , im new to this i have installed this pdf chat bot by docker compose on unraid , but its not using gpu and taking 15 mins to 20mis to generate response any help and guidance on this ? I want to ask if this supports gpu nividia |
Beta Was this translation helpful? Give feedback.
Replies: 7 comments 2 replies
-
Hello @drmetro09. I have added a If you're having a Nvidia GPU and you've installed the correct drivers, you should be able to use the GPU with the PS: If you're setting up and running the setup for the very first time, you will need to wait for the model to download and the embedding model to load. Once these steps are done, the responses should be pretty fast. Also, I've added benchmarks in the readme file for reference. Happy exploring! |
Beta Was this translation helpful? Give feedback.
-
Hi , thanks for quick response but when gpu support is enabled it gives cublas error 15 and the bot doesn’t not give any response i have sent you a request on insta could you please guide me with this am new to this |
Beta Was this translation helpful? Give feedback.
-
The error is originating from the Ollama server module. Do you have the correct Nvidia drivers installed? If you have, could you share some screenshots for the same? And maybe share the details of your execution environment to get a better understanding about the issue. You could have a look at the Ollama docker setup documentation to see if it helps you. There seems to be a similar issue reported in Ollama repo. See if you can correlate your issue to that. Also,
I didn't quite understand this. :) |
Beta Was this translation helpful? Give feedback.
-
Have tried various ollama models like orca-mini , llama2, starling-lm all are behaving same |
Beta Was this translation helpful? Give feedback.
-
Update: Its finally working now , i downgraded from to v 0.1.10 and everything working fine its using the gpu without errors, tried using LLAMA2 model haven’t tried other models.
|
Beta Was this translation helpful? Give feedback.
-
Great to hear that you were able to get it working! I personally have tried different models and found About the feature/enhancement suggestions:
Cheers! |
Beta Was this translation helpful? Give feedback.
Update: Its finally working now , i downgraded from to v 0.1.10 and everything working fine its using the gpu without errors, tried using LLAMA2 model haven’t tried other models.
Thanks for your amazing work , i have few suggestions listed below , hope you would consider them for this project.