Skip to main content
Last Updated: April 3, 2026

Introduction

Continue is an open-source AI coding assistant for VS Code and JetBrains. Unlike pure agentic tools, Continue focuses on the full coding workflow: autocomplete, inline chat, codebase-aware Q&A, and multi-file edits. Supports any OpenAI-compatible endpoint with custom headers. Continue works with SaladCloud in two ways:
  • Salad AI Gateway - no infrastructure to deploy or manage. Sign up for access, point Continue at a single shared endpoint, and use your Salad API key directly. Currently in closed beta, available via monthly subscription.
  • Self-hosted model - deploy your own SaladCloud container group, billed per hour, for full control over the model, hardware, and configuration. Still very easy to set up and use.

Prerequisites

Before getting started, make sure you have:

Step-by-Step Setup

Step 1: Choose Your Backend

Salad AI Gateway is the fastest way to get started - no container groups to deploy, no cold starts to wait for.
  1. Sign up for early access at salad.com/ai-gateway.
  2. Once approved, find your Salad API key in the portal.
Available models:
ModelDescription
qwen3.5-35b-a3bQwen 3.5 35B Mixture of Experts - best for agentic tasks and complex reasoning
qwen3.5-27bQwen 3.5 27B - strong balance of capability and speed
qwen3.5-9bQwen 3.5 9B - fastest response times, suited for lighter tasks

Step 2: Install Continue

VS Code:
  1. Open the Extensions view (Ctrl+Shift+X or Cmd+Shift+X)
  2. Search for Continue
  3. Click Install on the Continue extension by Continue Dev
  4. Once installed, click the Continue icon in the sidebar
JetBrains: Install the Continue plugin from the JetBrains Marketplace via Settings → Plugins → Marketplace.

Step 3: Configure Continue to Use Your SaladCloud Endpoint

Continue is configured via ~/.continue/config.yaml. Open it from the Continue panel by clicking the gear icon, or edit it directly. The configuration UI does not currently support all necessary fields, so you will need to add the model config manually in the YAML file.
Add your AI Gateway model:
models:
  - name: Salad AI Gateway
    provider: openai
    model: qwen3.5-35b-a3b
    apiBase: https://ai.salad.cloud:40404/v1
    apiKey: your-salad-api-key
No custom headers are needed - your Salad API key in apiKey is all that’s required.

Step 4: Test the Connection

Reload VS Code (Ctrl+Shift+PReload Window), then open the Continue chat panel and ask:
“Summarize what this project does in one sentence.”
If Continue responds, your setup is complete.

Tips for Best Results

Set Context Window

Add contextLength to your model config to help Continue manage prompt sizes correctly:
models:
  - name: SaladCloud Qwen 3.5
    provider: openai
    model: qwen3.5-35b-a3b
    apiBase: https://your-endpoint.salad.cloud/v1
    apiKey: dummy
    roles:
      - chat
      - autocomplete
    contextLength: 32768
    requestOptions:
      headers:
        Salad-Api-Key: your-salad-api-key
The Qwen 3.5-35B-A3B recipe supports up to 262,144 tokens.

Handle Autocomplete Timeouts

SaladCloud has a 100-second request timeout. Autocomplete requests are short and typically complete well within this limit, but for large context windows, consider setting a shorter maxTokens for autocomplete to reduce latency:
tabAutocompleteModel:
  name: SaladCloud Qwen 3.5
  provider: openai
  model: qwen3.5-35b-a3b
  apiBase: https://your-endpoint.salad.cloud/v1
  apiKey: dummy
  defaultCompletionOptions:
    maxTokens: 512
  requestOptions:
    headers:
      Salad-Api-Key: your-salad-api-key

Model Recommendations

  • Qwen 3.5-35B-A3B: Best for chat, codebase Q&A, and multi-file edits
  • Qwen 3.5-9B: Fast enough for autocomplete; good balance of latency and quality