Ollama is an open-source tool that allows to run large language models (LLMs) locally on their own computers. To use Ollama, you can install it here and download the model you want to run with the ollama run
command.
Chat model
We recommend configuring Llama3.1 8B as your chat model.
models:
- name: Llama3.1 8B
provider: ollama
model: llama3.1:8b
models:
- name: Llama3.1 8B
provider: ollama
model: llama3.1:8b
{
"models": [
{
"title": "Llama3.1 8B",
"provider": "ollama",
"model": "llama3.1:8b"
}
]
}
Autocomplete model
We recommend configuring Qwen2.5-Coder 1.5B as your autocomplete model.
models:
- name: Qwen2.5-Coder 1.5B
provider: ollama
model: qwen2.5-coder:1.5b-base
roles:
- autocomplete
models:
- name: Qwen2.5-Coder 1.5B
provider: ollama
model: qwen2.5-coder:1.5b-base
roles:
- autocomplete
{
"tabAutocompleteModel": {
"title": "Qwen2.5-Coder 1.5B",
"provider": "ollama",
"model": "qwen2.5-coder:1.5b-base"
}
}
Embeddings model
We recommend configuring Nomic Embed Text as your embeddings model.
models:
- name: Nomic Embed Text
provider: ollama
model: nomic-embed-text
roles:
- embed
models:
- name: Nomic Embed Text
provider: ollama
model: nomic-embed-text
roles:
- embed
{
"embeddingsProvider": {
"provider": "ollama",
"model": "nomic-embed-text"
}
}
Reranking model
Ollama currently does not offer any reranking models.
Click here to see a list of reranking model providers.
Using a remote instance
To configure a remote instance of Ollama, add the "apiBase"
property to your model in config.json:
models:
- name: Llama3.1 8B
provider: ollama
model: llama3.1:8b
apiBase: http://<my endpoint>:11434
models:
- name: Llama3.1 8B
provider: ollama
model: llama3.1:8b
apiBase: http://<my endpoint>:11434
{
"models": [
{
"title": "Llama3.1 8B",
"provider": "ollama",
"model": "llama3.1:8b",
"apiBase": "http://<my endpoint>:11434"
}
]
}