Ollama is an open-source tool that allows to run large language models (LLMs) locally on their own computers. To use Ollama, you can install it here and download the model you want to run with the ollama run
command.
We recommend configuring Llama3.1 8B as your chat model.
We recommend configuring Qwen2.5-Coder 1.5B as your autocomplete model.
We recommend configuring Nomic Embed Text as your embeddings model.
Ollama currently does not offer any reranking models.
Click here to see a list of reranking model providers.
To configure a remote instance of Ollama, add the "apiBase"
property to your model in config.json: