Popular Model Providers
These are the most commonly used model providers that offer a wide range of capabilities:| Provider | Description | Capabilities |
|---|---|---|
| Anthropic | Providers of Claude models, known for long context windows and strong reasoning | Chat, Edit, Apply, Embeddings |
| OpenAI | Creators of GPT models with strong coding capabilities | Chat, Edit, Apply, Embeddings |
| Azure | Microsoft’s cloud platform offering OpenAI models | Chat, Edit, Apply, Embeddings |
| Amazon Bedrock | AWS service offering access to various foundation models | Chat, Edit, Apply, Embeddings |
| Ollama | Run open-source models locally with a simple interface | Chat, Edit, Apply, Embeddings, Autocomplete |
| Google Gemini | Google’s multimodal AI models | Chat, Edit, Apply, Embeddings |
| DeepSeek | Specialized code models with strong performance | Chat, Edit, Apply |
| Mistral | High-performance open models with commercial offerings | Chat, Edit, Apply, Embeddings |
| xAI | Grok models from xAI | Chat, Edit, Apply |
| Vertex AI | Google Cloud’s machine learning platform | Chat, Edit, Apply, Embeddings |
| Inception | On-premises open-source model runners | Chat, Edit, Apply |
Additional Model Providers
Beyond the top-level providers, Continue supports many other options:Hosted Services
| Provider | Description |
|---|---|
| Groq | Ultra-fast inference for various open models |
| Together AI | Platform for running a variety of open models |
| DeepInfra | Hosting for various open source models |
| OpenRouter | Gateway to multiple model providers |
| Tetrate Agent Router Service | Gateway with intelligent routing across multiple model providers |
| Cohere | Models specialized for semantic search and text generation |
| NVIDIA | GPU-accelerated model hosting |
| Cloudflare | Edge-based AI inference services |
| HuggingFace | Platform for open source models |
Local Model Options
| Provider | Description |
|---|---|
| LM Studio | Desktop app for running models locally |
| llama.cpp | Optimized C++ implementation for running LLMs |
| LlamaStack | Stack for running Llama models locally |
| llamafile | Self-contained executable model files |
Enterprise Solutions
| Provider | Description |
|---|---|
| SambaNova | Enterprise AI platform |
| Watson x | IBM’s enterprise AI platform |
| Sagemaker | AWS machine learning platform |
| Nebius | Cloud-based machine learning platform |
How to Choose a Model Provider
When selecting a model provider, consider:- Hosting preference: Do you need local models for offline use or privacy, or are you comfortable with cloud services?
- Performance requirements: Different providers offer varying levels of speed, quality, and context length.
- Specific capabilities: Some models excel at code generation, others at embeddings or reasoning tasks.
- Pricing: Costs vary significantly between providers, from free local options to premium cloud services.
- API key requirements: Most cloud providers require API keys that you’ll need to configure.
Configuration Format
You can add models to yourconfig.yaml file like this: