Select your language

Live Chat Scroll naar beneden

Local AI Coding Setup: Continue, Ollama & Devstral / DeepSeek Coder

Local AI Coding Setup: Continue, Ollama & Devstral / DeepSeek Coder

Auteur: Siu-Ho Fung

May 12, 2025

Want a better coding experience with local models? Try Continue - a code assistant extension for Visual Studio Code. Combined with Ollama and advanced models like Devstral / DeepSeek Coder, it gives you code suggestions, edits, and chat - all without sending your source code online.

Personal note:
For those looking for a robust free alternative to Cursor, I personally use a combination of Continue, Ollama, and Devstral 24B / DeepSeek Coder 33B. I find that DeepSeek Coder is more capable and gives better results. This setup works very well for me in my daily development workflow.

Everything runs locally, so your code stays private and secure - no cloud dependency.

It brings Cursor-like features to your existing VS Code setup.

Whether you're coding in Python, C++, or JavaScript, PHP, this setup provides smart completions, code edits, and chat-based AI assistance, all running locally and offline.


πŸ“Š Roo Code vs. Continue – Feature Comparison

FeatureRoo CodeContinue
VS Code Extensionβœ… Yesβœ… Yes
Local Model Supportβœ… Yes (via Ollama or local APIs)βœ… Yes (via Ollama)
Cloud Model Supportβœ… Yes (e.g. OpenAI, Anthropic, DeepSeek)βœ… Yes (e.g. OpenAI, Anthropic)
Chat Interfaceβœ… Yesβœ… Yes
Autocomplete Integrationβœ… Yesβœ… Yes (tight integration)
Code Edit Suggestionsβœ… Yesβœ… Yes
Multi-model Configurationβœ… Yes (custom prompt modes)βœ… Yes (`continue.config.yaml`)
File System Accessβœ… Yesβœ… Yes
Terminal Command Executionβœ… Yesβœ… Yes
Browser Automationβœ… Yes❌ No
Offline Useβœ… Fully supportedβœ… Fully supported
Open Sourceβœ… Yesβœ… Yes
Ease of Setupβœ… Easy (VS Code Marketplace)βœ… Easy (VS Code Marketplace)
Customizationβœ… High (Modes, commands, routing)βœ… High (YAML config for models and roles)
Community Size🟑 Medium (growing)🟒 Large (mature VS Code ecosystem)

βœ… Roo Code – Pros and Cons

Pros

  • Works with both local and cloud-based LLMs.
  • Supports browser automation and custom assistant commands.
  • Highly customizable prompt modes (e.g., Debug, Code, Architect).
  • Terminal and file access for AI agents.
  • Clean UI within VS Code with multi-tab chat.

Cons

  • Slightly heavier footprint compared to Continue.
  • Some features (e.g., agent commands) may be overkill for basic dev tasks.
  • Smaller community and less documentation than Continue.

βœ… Continue – Pros and Cons

Pros

  • Lightweight and tightly integrated into VS Code interface.
  • Seamless autocomplete, chat, and code edit experience.
  • Optimized for local model usage with continue.config.yaml.
  • Larger community and robust documentation.
  • Fast setup via Marketplace + Ollama.

Cons

  • Lacks browser automation and advanced multi-modal tools.
  • No native agent system for autonomous tasks.
  • Requires manual config for advanced multi-model workflows.

πŸ’‘ Recommendation: Use Continue if you want a focused, efficient AI coding assistant tightly integrated into VS Code with excellent local model support.
Choose Roo Code if you prefer an assistant-style developer tool that can automate browser tasks, run terminal commands, and adapt to diverse workflows.


πŸ› οΈ How to Install Continue in VS Code

1. Install the Continue Extension

  1. Open VS Code.
  2. Go to the Extensions view (Ctrl+Shift+X or Cmd+Shift+X).
  3. Search for Continue and install the one by Continue Dev.

Or use the terminal:

code --install-extension Continue.continue

2. Install Ollama

Ollama is a local model runner that works seamlessly with Continue.

πŸͺŸ On Windows

  1. Download the installer from https://ollama.com
  2. Follow the setup instructions.
  3. Open a terminal and run:
    ollama run llama3

🍎 On macOS

brew install ollama
ollama run llama3

3. Use "Better" Models (e.g., Devstral)

Ollama supports many high-performance open-source models. To pull and run Devstral 24B:

ollama pull devstral:24b-small-2505-q4_K_M
ollama pull deepseek-coder:6.7b

Then, test it:

ollama run devstral:24b-small-2505-q4_K_M

⚠️ System Requirements: To run this model, a GPU with at least 24 GB of VRAM is recommended, such as an NVIDIA RTX 3090 or higher. If you use a model variant with fewer parameters or quantization, 16 GB of VRAM may be sufficient in some cases.

This model is well-suited for chat, code editing, and longer prompts due to its extended context window.


βš™οΈ Optional: Advanced Configuration with continue.config.yaml

For deeper integration and multi-model control, you can create a custom config file. This allows Continue to:

  • Assign specific models to tasks like chat, edit, or autocomplete
  • Increase token limits for models with long context windows
  • Add semantic search and embeddings with models like nomic-embed-text

πŸ“„ Example continue.config.yaml

Save this in the root of your workspace or inside a .continue/ folder:

name: Local Assistant
version: 1.0.0
schema: v1models:
- name: Devstral Small 24B (Q4_K_M)
provider: ollama
model: devstral:24b-small-2505-q4_K_M
maxTokens: 131072 # explicitly set
contextLength: 128000 # if supported by Continue
roles:
- chat
- edit
- apply- name: DeepSeek-Coder 6.7B
provider: ollama
model: deepseek-coder:6.7b
maxTokens: 16384
contextLength: 16384
roles:
- autocomplete- name: Nomic Embed
provider: ollama
model: nomic-embed-text:latest
roles:
- embedcontext:
- provider: code
- provider: docs
- provider: diff
- provider: terminal
- provider: problems
- provider: folder
- provider: codebase

πŸ“Œ Tip: After editing the config, restart the Continue extension or reload VS Code.


πŸš€ Benefits of Continue + Ollama

βœ… Local-first: No cloud dependency
βœ… Multi-model support: Use specialized models for specific roles
βœ… Fast: Optimized for real-time dev workflows
βœ… Customizable: YAML-based configuration
βœ… Offline: Ideal for secure environments


By combining Continue with Ollama and powerful open-source models like Devstral en Qwen3, you unlock a next-level AI coding experience - all within VS Code and on your own machine.


🀝 Combining Roo Code and Continue

While Roo Code and Continue each have their own strengths, you can also use them together to create a powerful and flexible AI coding environment.

  • Use Continue for tightly integrated autocomplete, inline code edits, and multi-model roles within VS Code.
  • Use Roo Code as a complement for tasks like browser automation, terminal commands, and structured prompts via custom modes.

By installing both tools side by side in Visual Studio Code, you benefit from:

  • The best of fast, contextual coding assistance (Continue)
  • Advanced AI agent functionality and automation (Roo Code)

πŸ’‘ Note: Make sure your configuration files (such as continue.config.yaml and Roo’s settings) don’t conflict. Both tools run independently and can function in parallel. Sometimes it is also necessary to adjust the key combinations due to conflicts.


10 Coding Models for Local Use with Ollama (May 2025)

Model NamePrecise Model IdentifierNotes
DevStral 7Bdevstral:7b
DeepSeek Coder 6.7Bdeepseek-coder:6.7b
DeepSeek Coder 33B Q4deepseek-coder:33b-instruct-q4_K_MPersonal favorite
CodeLlama 13B Instructcodellama:13b-instruct
CodeLlama 34B Instructcodellama:34b-instructVRAM-intensive; feasible with quantization
Phind-CodeLlama-34B-v2phind-codellama:34b-v2
WizardCoder 34Bwizardcoder:34bWith Q4 or Q5 quantization
Mixtral 8x7B Instructmixtral:8x7b-instruct
StarCoder 15Bstarcoder:15bManual setup required
aiXcoder 7Baixcoder:7bManual setup required

Start building smarter, faster, and more securely, right from your editor.

Schrijf in voor onze Nieuwsbrief

Hebt u vragen of hulp nodig? Wij helpen u graag.

15+ jaar ervaring Preferred partner van Dell, HPE & Supermicro en meer Advies op maat binnen 1 werkdag Snelle levering & installatie Wereldwijde 24/7 onsite support Laagste prijsgarantie