To get started, create an API key on the OVHcloud AI Endpoints
website. For more information, including
pricing, visit the OVHcloud AI Endpoints product
page.
Chat Model
We recommend configuring Qwen2.5-Coder-32B-Instruct as your chat model. Check our catalog to see all of our models hosted on AI Endpoints.Available Models
OVHcloud AI Endpoints provides access to the following models: Llama Models:llama3.1-8b- Llama 3.1 8B Instruct (supports function calling)llama3.1-70b- Llama 3.1 70B Instructllama3.3-70b- Llama 3.3 70B Instruct (supports function calling)
qwen2.5-coder-32b- Qwen 2.5 Coder 32B Instruct (supports function calling)qwen3-32b- Qwen 3 32B (supports function calling)qwen3-coder-30b-a3b- Qwen 3 Coder 30B A3B Instruct (supports function calling)qwen2.5-vl-72b- Qwen 2.5 VL 72B Instruct (vision-language model)
mistral-7b- Mistral 7B Instruct v0.3mistral-8x7b- Mixtral 8x7B Instruct v0.1mistral-nemo- Mistral Nemo Instruct 2407 (supports function calling)mistral-small-3.2-24b- Mistral Small 3.2 24B Instruct (supports function calling)
gpt-oss-20b- GPT-OSS 20B (supports function calling)gpt-oss-120b- GPT-OSS 120B (supports function calling)
DeepSeek-R1-Distill-Llama-70B- DeepSeek R1 Distill Llama 70B (supports function calling)
codestral-mamba-latest- Codestral Mamba 7B v0.1
Function Calling Support
Many OVHcloud models support function calling (tool use), which enables the model to interact with external tools and APIs. Models with function calling support include:- Llama 3.1 8B, Llama 3.3 70B
- All Qwen 3 models and Qwen 2.5 Coder
- GPT-OSS 20B and 120B
- DeepSeek R1 Distill Llama 70B
- Mistral Small 3.2 24B, Mistral Nemo
- YAML
- JSON
config.yaml
Example: Model with Function Calling
Hereβs an example configuration for a model that supports function calling:- YAML
- JSON
config.yaml
Embeddings Model
We recommend configuring bge-multilingual-gemma2 as your embeddings model.- YAML
- JSON
config.yaml