
Homelab AI stack 2026 — what to run and in what order
TL;DR Stop running your AI brain on someone else's servers. Here's the exact stack I run on my homelab — in the order that actually makes sense to deploy it. Why self-hosted AI in 2026? The models crossed a threshold. qwen2.5:32b running locally on a decent machine beats GPT-3.5 on most developer tasks. It's free, private, offline, and you own every token. Self-hosting your AI stack isn't a nerd flex anymore. It's good engineering hygiene. You wouldn't run prod on someone else's laptop. Why run your reasoning on their servers? The Stack (in order) 1. Traefik — everything behind HTTPS first Before anything else gets internet-exposed, Traefik goes in. Automatic TLS, reverse proxy, single entrypoint. docker run -d \ -p 80:80 -p 443:443 \ -v /var/run/docker.sock:/var/run/docker.sock \ traefik:v3.0 Don't skip this step. Everything else sits behind it. 2. Ollama — your local LLM engine curl -fsSL https://ollama.ai/install.sh | sh ollama run qwen2.5:32b Swap model names freely: gemma3 , mistr
Continue reading on Dev.to
Opens in a new tab




