Running Opencoder LLM in VS Code: A Local, Copilot Alternative

Experiencing a local AI assistant in VS Code with OpenCoder LLM.
Warp Terminal

Programming is the one area where AI is being used extensively. Most editors allow you to add AI agents like chatGPT, Microsoft's Copilot etc.

There are also several open source large language models specifically centered around coding like CodeGemma.

And then we have a new player entering the scene and everyone talking about it.

This new player is OpenCoder, an open-source Code LLM available in 1.5B and 8B models.

Since OpenCoder is getting popular, I decided to quickly test it out as a local AI assistent to help me code in VS Code.

With my experience here, you’ll also be able to integrate OpenCoder (or any other LLM) into VS Code with the help of CodeGPT extension and enjoy the perks of a local AI assistant.

📋
This tutorial is performed on Ubuntu 24.04 in a VM. Steps mentioned are mostly applicable on Ubuntu. But, feel free to replicate the same on other Linux distributions and operating systems.

Step 1: Install VS Code (if you don't have it already)

First, ensure that Visual Studio Code is installed on your system. If not, follow these steps to get it set up:

  • Download VS Code for Ubuntu here.
  • Once downloaded, open your terminal and install it by running:
sudo dpkg -i ./<downloaded-file>.deb

After installation, launch VS Code from your applications menu or by typing code in the terminal.

Step 2: Install the CodeGPT Extension

While you can integrate some major platforms like Ollama and LM Studio for free, CodeGPT isn’t entirely free. Basic usage comes at no cost, but if you're looking for a one-click solution with advanced features, CodeGPT offers paid plans that require creating an account.

CodeGPT is a powerful tool that I've found invaluable for boosting productivity and simplifying coding workflows.

As an extension for VS Code, it integrates seamlessly, providing instant code suggestions, completions, and debugging insights right where I need them.

Here’s how to set it up:

  1. Open VS Code.
  2. Click on the Extensions icon in the sidebar.
  3. Search for “CodeGPT” in the Extensions Marketplace.
  4. Select the extension and click “Install.”
CodeGPT extension in VS Code

After installation, the extension should appear in the sidebar for easy access.

CodeGPT pinned to the sidebar for easy access

This extension serves as the interface between your VS Code environment and OpenCoder, allowing you to request code suggestions, completions, and more.

Step 3: Install Ollama

Ollama is an essential tool for managing and deploying language models locally. It simplifies the process of downloading and running models locally, making it a crucial component for this setup.

Ollama provides an official installation script. Run the command below to install Ollama:

curl -fsSL https://ollama.com/install.sh | sh

Once the installation process completes, open a web browser and enter:

localhost:11434

It should show the message, "Ollama is running".

Ollama running locally

We have also covered Ollama installation steps in detail, in case you need that.

Running AI Locally Using Ollama on Ubuntu Linux
Running AI locally on Linux because open source empowers us to do so.

Step 4: Download Opencoder model with Ollama

With CodeGPT and Ollama installed, you’re ready to download the Opencoder model:

In the terminal window, type:

ollama run opencoder

This may take a few minutes, depending on your internet speed and hardware specifications.

Downloading opencoder model using ollama

Once complete, the model is now ready for use within CodeGPT.

Step 5: Run your Copilot AI with opencoder

Open a code file or project in VS Code (I'm using an empty docker-compose.yml file).

In the CodeGPT panel, ensure Ollama is selected as your active provider.

Inside CodeGPT, selecting Ollama as a local provider

Next, it'll ask you to select the model. I've downloaded both version of Opencoder with 1.5B & 8B parameters but I'll be using the former:

After selecting the provider, it'll prompt to select the LLM from the list

Start engaging with the model by typing your queries or use it for code completions, suggestions, or any coding help you need:

Once selected, you can now engage with your AI code assistant in the below chat box

Here’s a quick video of me interacting with the OpenCoder model inside VS Code using the CodeGPT extension.

Conclusion

After testing OpenCoder with Ollama on my Ubuntu VM setup, I found that while the performance wasn't as snappy as cloud-based AI services like ChatGPT, it was certainly functional for most coding tasks.

The responsiveness can vary, especially on modest hardware, so performance is a bit subjective depending on your setup and expectations.

However, if data privacy and local processing are priorities, this approach still offers a solid alternative.

For developers handling sensitive code who want an AI copilot without relying on cloud services, Ollama and CodeGPT in VS Code is a worthwhile setup, balancing privacy and accessibility right on your desktop.

About the author
Abhishek Kumar

Abhishek Kumar

I'm definitely not a nerd, perhaps a geek who likes to tinker around with whatever tech I get my hands on. Figuring things out on my own gives me joy. BTW, I don't use Arch.

Become a Better Linux User

With the FOSS Weekly Newsletter, you learn useful Linux tips, discover applications, explore new distros and stay updated with the latest from Linux world

itsfoss happy penguin

Great! You’ve successfully signed up.

Welcome back! You've successfully signed in.

You've successfully subscribed to It's FOSS.

Success! Check your email for magic link to sign-in.

Success! Your billing info has been updated.

Your billing was not updated.