This guide will help you set up Ollama for Devika. Ollama is a tool that allows you to run open-source large language models (LLMs) locally on your machine. It supports varity of models like Llama-2, mistral, code-llama and many more.
- go to the Ollama website.
- Download the latest version of the Ollama.
- After installing the Ollama, you have to download the model you want to use. Models
- select the model you want to download and copy the command. for example,
ollama run llama2
.it will download the model and start the server. ollama list
will show the list of models you have downloaded.- if the server isn't running then you can manually start by
ollama serve
. default address for the server ishttp://localhost:11434
- for changing port and other configurations, follow the FAQ here
- for more information,
ollama [command] --help
will show the help menu. for example,ollama run --help
will show the help menu for the run command.
- if you serve the Ollama on a different address, you can change the port in the
config.toml
file or you can change it via UI. - if you are using the default address, devika will automatically detect the server and and fetch the models list.