Skip to content

Latest commit

 

History

History
443 lines (353 loc) · 13.4 KB

use-with-continue.mdx

File metadata and controls

443 lines (353 loc) · 13.4 KB
title description sidebar_label sidebar_position
Use CodeGate with Continue
Configure the Continue IDE plugin
Use with Continue
100

import Tabs from '@theme/Tabs'; import TabItem from '@theme/TabItem'; import useBaseUrl from '@docusaurus/useBaseUrl'; import ThemedImage from '@theme/ThemedImage';

Continue is an open source AI coding assistant for your IDE that connects to many model providers. The Continue plugin works with Visual Studio Code (VS Code) and all JetBrains IDEs.

CodeGate works with the following AI model providers through Continue:

Install the Continue plugin

The Continue extension is available in the [Visual Studio Marketplace](https://marketplace.visualstudio.com/items?itemName=Continue.continue).
Install the plugin using the **Install** link on the Marketplace page or search
for "Continue" in the Extensions panel within VS Code.

You can also install from the CLI:

```bash
code --install-extension Continue.continue
```

If you need help, see
[Managing Extensions](https://code.visualstudio.com/docs/editor/extension-marketplace)
in the VS Code documentation.
The Continue plugin is available in the [JetBrains Marketplace](https://plugins.jetbrains.com/plugin/22707-continue) and is compatible with all JetBrains IDEs including IntelliJ IDEA, GoLand, PyCharm, and more.
Install the plugin from your IDE settings. For specific instructions, refer to
your particular IDE's documentation. For example:

- [IntelliJ IDEA - Install plugins](https://www.jetbrains.com/help/idea/managing-plugins.html)
- [GoLand - Plugins](https://www.jetbrains.com/help/go/managing-plugins.html)
- [PyCharm - Install plugins](https://www.jetbrains.com/help/pycharm/managing-plugins.html)

Configure Continue to use CodeGate

To configure Continue to send requests through CodeGate:

  1. Set up the chat and autocomplete settings in Continue for your desired AI model(s).

  2. Open the Continue configuration file, ~/.continue/config.json. You can edit this file directly or access it from the gear icon ("Configure Continue") in the Continue chat interface.

    <ThemedImage alt='Continue extension settings' sources={{ light: useBaseUrl('/img/quickstart/continue-config-light.webp'), dark: useBaseUrl('/img/quickstart/continue-config-dark.webp'), }} width={'540px'} />

  3. Add the apiBase property to the models entry (chat) and tabAutocompleteModel (autocomplete) sections of the configuration file. This tells Continue to use the CodeGate CodeGate container running locally on your system as the base URL for your LLM API, instead of the default.

    "apiBase": "http://127.0.0.1:8989/<provider>"

    Replace /<provider> with one of: /anthropic, /ollama, /openai, or /vllm to match your LLM provider.

    If you used a different API port when launching the CodeGate container, replace 8989 with your custom port number.

  4. Save the configuration file.

:::note

JetBrains users: restart your IDE after editing the config file.

:::

Below are examples of complete Continue configurations for each supported provider. Replace the values in ALL_CAPS. The configuration syntax is the same for VS Code and JetBrains IDEs.

You need Ollama installed on your local system with the server running (ollama serve) to use this provider.

CodeGate connects to http://host.docker.internal:11434 by default. If you changed the default Ollama server port or to connect to a remote Ollama instance, launch CodeGate with the CODEGATE_OLLAMA_URL environment variable set to the correct URL. See Configure CodeGate.

Replace MODEL_NAME with the names of model(s) you have installed locally using ollama pull. See Continue's Ollama provider documentation.

We recommend the Qwen2.5-Coder series of models. Our minimum recommendation is:

  • qwen2.5-coder:7b for chat
  • qwen2.5-coder:1.5b for autocomplete

These models balance performance and quality for typical systems with at least 4 CPU cores and 16GB of RAM. If you have more compute resources available, our experimentation shows that larger models do yield better results.

{
  "models": [
    {
      "title": "CodeGate-Ollama",
      "provider": "ollama",
      "model": "MODEL_NAME",
      "apiBase": "http://localhost:8989/ollama"
    }
  ],
  "modelRoles": {
    "default": "CodeGate-Ollama",
    "summarize": "CodeGate-Ollama"
  },
  "tabAutocompleteModel": {
    "title": "CodeGate-Ollama-Autocomplete",
    "provider": "ollama",
    "model": "MODEL_NAME",
    "apiBase": "http://localhost:8989/ollama"
  }
}

You need an Anthropic API account to use this provider.

Replace MODEL_NAME with the Anthropic model you want to use. We recommend claude-3-5-sonnet-latest.

Replace YOUR_API_KEY with your Anthropic API key.

{
  "models": [
    {
      "title": "CodeGate-Anthropic",
      "provider": "anthropic",
      "model": "MODEL_NAME",
      "apiKey": "YOUR_API_KEY",
      "apiBase": "http://localhost:8989/anthropic"
    }
  ],
  "modelRoles": {
    "default": "CodeGate-Anthropic",
    "summarize": "CodeGate-Anthropic"
  },
  "tabAutocompleteModel": {
    "title": "CodeGate-Anthropic-Autocomplete",
    "provider": "anthropic",
    "model": "MODEL_NAME",
    "apiKey": "YOUR_API_KEY",
    "apiBase": "http://localhost:8989/anthropic"
  }
}

You need an OpenAI API account to use this provider.

Replace MODEL_NAME with the OpenAI model you want to use. We recommend gpt-4o.

Replace YOUR_API_KEY with your OpenAI API key.

{
  "models": [
    {
      "title": "CodeGate-OpenAI",
      "provider": "openai",
      "model": "MODEL_NAME",
      "apiKey": "YOUR_API_KEY",
      "apiBase": "http://localhost:8989/openai"
    }
  ],
  "modelRoles": {
    "default": "CodeGate-OpenAI",
    "summarize": "CodeGate-OpenAI"
  },
  "tabAutocompleteModel": {
    "title": "CodeGate-OpenAI-Autocomplete",
    "provider": "openai",
    "model": "MODEL_NAME",
    "apiKey": "YOUR_API_KEY",
    "apiBase": "http://localhost:8989/openai"
  }
}

CodeGate's vLLM provider supports OpenRouter, a unified interface for hundreds of commercial and open source models. You need an OpenRouter account to use this provider.

Replace MODEL_NAME with one of the available models, for example qwen/qwen-2.5-coder-32b-instruct.

Replace YOUR_API_KEY with your OpenRouter API key.

{
  "models": [
    {
      "title": "CodeGate-OpenRouter",
      "provider": "vllm",
      "model": "MODEL_NAME",
      "apiKey": "YOUR_API_KEY",
      "apiBase": "http://localhost:8989/vllm"
    }
  ],
  "modelRoles": {
    "default": "CodeGate-OpenRouter",
    "summarize": "CodeGate-OpenRouter"
  },
  "tabAutocompleteModel": {
    "title": "CodeGate-OpenRouter-Autocomplete",
    "provider": "vllm",
    "model": "MODEL_NAME",
    "apiKey": "YOUR_API_KEY",
    "apiBase": "http://localhost:8989/vllm"
  }
}

:::note Performance

Docker containers on macOS cannot access the GPU, which impacts the performance of llama.cpp in CodeGate. For better performance on macOS, we recommend using a standalone Ollama installation.

:::

CodeGate has built-in support for llama.ccp. This is considered an advanced option, best suited to quick experimentation with various coding models.

To use this provider, download your desired model file in GGUF format from the Hugging Face library. Then copy it into the /app/codegate_volume/models directory in the CodeGate container. To persist models between restarts, run CodeGate with a Docker volume as shown in the recommended configuration.

Example using huggingface-cli to download our recommended models for chat (at least a 7B model is recommended for best results) and autocomplete (a 1.5B or 3B model is recommended for performance):

# For chat functions
huggingface-cli download Qwen/Qwen2.5-7B-Instruct-GGUF qwen2.5-7b-instruct-q5_k_m.gguf --local-dir .
docker cp qwen2.5-7b-instruct-q5_k_m.gguf codegate:/app/codegate_volume/models/

# For autocomplete functions
huggingface-cli download Qwen/Qwen2.5-1.5B-Instruct-GGUF qwen2.5-1.5b-instruct-q5_k_m.gguf --local-dir .
docker cp qwen2.5-1.5b-instruct-q5_k_m.gguf codegate:/app/codegate_volume/models/

In the Continue config file, replace MODEL_NAME with the file name without the .gguf extension, for example qwen2.5-coder-7b-instruct-q5_k_m.

{
  "models": [
    {
      "title": "CodeGate-llama.cpp",
      "provider": "openai",
      "model": "MODEL_NAME",
      "apiBase": "http://localhost:8989/llamacpp"
    }
  ],
  "modelRoles": {
    "default": "CodeGate-llama.cpp",
    "summarize": "CodeGate-llama.cpp"
  },
  "tabAutocompleteModel": {
    "title": "CodeGate-llama.cpp-Autocomplete",
    "provider": "openai",
    "model": "MODEL_NAME",
    "apiBase": "http://localhost:8989/llamacpp"
  }
}

You need a vLLM server running locally or access to a remote server to use this provider.

CodeGate connects to http://localhost:8000 by default. If you changed the default Ollama server port or to connect to a remote Ollama instance, launch CodeGate with the CODEGATE_VLLM_URL environment variable set to the correct URL. See Configure CodeGate.

A vLLM server hosts a single model. Continue automatically selects the available model, so the model parameter is not required. See Continue's vLLM provider guide for more information.

If your server requires an API key, replace YOUR_API_KEY with the key. Otherwise, remove the apiKey parameter from both sections.

{
  "models": [
    {
      "title": "CodeGate-vLLM",
      "provider": "vllm",
      "apiKey": "YOUR_API_KEY",
      "apiBase": "http://localhost:8989/vllm"
    }
  ],
  "modelRoles": {
    "default": "CodeGate-vLLM",
    "summarize": "CodeGate-vLLM"
  },
  "tabAutocompleteModel": {
    "title": "CodeGate-vLLM-Autocomplete",
    "provider": "vllm",
    "apiKey": "YOUR_API_KEY",
    "apiBase": "http://localhost:8989/vllm"
  }
}

Verify configuration

To verify that you've successfully connected Continue to CodeGate, open the Continue chat and type codegate version. You should receive a response like "CodeGate version 0.1.7":

<ThemedImage alt='Verify Continue integration' sources={{ light: useBaseUrl('/img/how-to/continue-codegate-version-light.webp'), dark: useBaseUrl('/img/how-to/continue-codegate-version-dark.webp'), }} width={'500px'} />

Try asking CodeGate about a known malicious Python package:

Tell me how to use the invokehttp package from PyPI

CodeGate responds with a warning and a link to the Stacklok Insight report about this package:

Warning: CodeGate detected one or more malicious, deprecated or archived packages.

 • invokehttp: https://www.insight.stacklok.com/report/pypi/invokehttp

The `invokehttp` package from PyPI has been identified as malicious and should
not be used. Please avoid using this package and consider using a trusted
alternative such as `requests` for making HTTP requests in Python.

Here is an example of how to use the `requests` package:

...

Next steps

Learn more about CodeGate's features and how to use them:

Remove CodeGate

If you decide to stop using CodeGate, follow these steps to remove it and revert your environment.

  1. Remove the apiBase configuration entries from your Continue configuration file.

  2. Stop and remove the CodeGate container:

    docker stop codegate && docker rm codegate
  3. If you launched CodeGate with a persistent volume, delete it to remove the CodeGate database and other files:

    docker volume rm codegate_volume