Why Self-Host?
Some organizations can’t send code to external APIs — regulated industries, defense contractors, companies with proprietary algorithms. Self-hosted tools run the AI model on your own servers, keeping all code within your network. No data leaves your infrastructure.
Key Open-Source Tools
Tabby (33K+ GitHub stars) is a self-hosted coding assistant that runs on consumer GPUs with no external dependencies. Continue.dev is an open-source IDE extension supporting any model provider. Aider is a CLI agent supporting any LLM. All three let you bring your own model — including local models via Ollama.
The Tradeoff
Self-hosted tools give you full control and privacy, but the models you can run locally (7B–70B parameters) are significantly less capable than frontier cloud models (hundreds of billions of parameters). The gap is closing — but in 2026, cloud models still produce notably better code for complex tasks.
The connection: The open-source ecosystem connects directly to Ch 3 (Training Code Models). Models like StarCoder2, DeepSeek-Coder, and CodeLlama are open-weight, meaning you can download and run them locally. The training pipeline is transparent, so you know exactly what data your model learned from.