Continue.Dev is an AI Code Editing tool built on top of VSCode
Setup
Continue.dev can be used in a fully offline way with local model support via ollama. It supports a yaml configuration file structure. Here is an example of how it can be configured with a combination of Ollama on the current machine and API calls to a LiteLLM proxy server:
name: my-configuration
version: 0.0.1
scheme: v1
models:
- name: qwen2.5-coder 1.5b
provider: ollama
model: qwen2.5-coder:1.5b
roles:
- chat
- edit
- apply
- autocomplete
- name: Autodetect
provider: ollama
model: AUTODETECT
- name: Autodetect
provider: openai
model: AUTODETECT
apiBase: https://litellm.yoursite.example/v1/
apiKey: sk-some-api-key
Setting num_ctx
for Ollama
By default ollama uses a context window of 2048
which is way too small for useful code assistants.
You can override the size of the context window by passing num_ctx
as an extra body parameter:
- name: Qwen 3 MoE 3x10B
provider: ollama
model: qwen3:30b-a3b
roles:
- chat
- edit
requestOptions:
extraBodyProperties:
num_ctx: 32384
It is probably worth also turning on context quantization and flash attention in Ollama first.
Agent Mode and Tool Support
Within continue’s config file you can add a block overriding the detection of tool use capability which allows you to use any model as an agent:
- name: Qwen 3 MoE 3x10B
provider: ollama
model: qwen3:30b-a3b
roles:
- chat
- edit
- apply
capabilities:
- tool_use
Currently Continue only supports agent mode for certain models that support tool use. How do they detect whether a model supports tool usage? They have a hard coded pattern matching routine in their source code
There is an issue about this open here. I proposed a new ‘role’ that would allow people to override the tool detection.