Instinct is a 7 billion parameter model. You should expect slow responses if running on a laptop. To learn how to inference Instinct on a GPU, see our HuggingFace model card.
We recently released Instinct, a state-of-the-art open Next Edit model. Robustly fine-tuned from Qwen2.5-Coder-7B, Instinct intelligently predicts your next move to keep you in flow. To learn more about the model, check out our blog post.

1. Install Ollama

If you haven’t already installed Ollama, see our guide here.

2. Download Instinct

ollama run nate/instinct

3. Update your config.yaml

Open your config.yaml and add Instinct to the models section:
# ... rest of config.yaml ...

models:
  - uses: continuedev/instinct
Alternatively, you can just click to add the block at https://hub.continue.dev/continuedev/instinct.