Back to articles
Ollama behind a reverse proxy with Caddy or Nginx for HTTPS streaming
NewsDevOps

Ollama behind a reverse proxy with Caddy or Nginx for HTTPS streaming

via Dev.toRost

Running Ollama behind a reverse proxy is the simplest way to get HTTPS, optional access control, and predictable streaming behaviour. This post focuses on Caddy and Nginx ingress for the Ollama API, not on client code. If you already have Python or Go clients talking to Ollama, this post is the missing piece: ingress and transport for the same API. For how Ollama fits alongside vLLM, Docker Model Runner, LocalAI, and cloud hosting trade-offs, see LLM Hosting in 2026: Local, Self-Hosted & Cloud Infrastructure Compared . For request examples and client code, see Ollama CLI Cheatsheet . For UI and multi-user layers, see Open WebUI overview, quickstart and alternatives . For the bigger picture on self-hosting and data control, see LLM self-hosting and AI sovereignty . For a reproducible single-node Ollama service in Docker Compose (persistent volumes, OLLAMA_HOST , NVIDIA GPUs, upgrades), see Ollama in Docker Compose with GPU and Persistent Model Storage . Why you should proxy Ollama inste

Continue reading on Dev.to

Opens in a new tab

Read Full Article
6 views

Related Articles