
Use any OpenCode model from Open WebUI, LangChain, or the OpenAI SDK — with a one-line install
If you use OpenCode , you already have GitHub Copilot, Ollama, Anthropic, Gemini, and other providers configured in one place. The problem: every other tool in your workflow — Open WebUI, LangChain, Chatbox, Continue, Zed, your own scripts — needs the same models re-entered with their own API keys and base URLs. I built opencode-llm-proxy to fix this. It's an OpenCode plugin that starts a local HTTP server on http://127.0.0.1:4010 and translates between whatever API format your tool uses and OpenCode's model list. Install npm install opencode-llm-proxy Add to opencode.json : { "plugin" : [ "opencode-llm-proxy" ] } Start OpenCode — the proxy starts automatically. Supported API formats All formats support streaming. Format Endpoint OpenAI Chat Completions POST /v1/chat/completions OpenAI Responses API POST /v1/responses Anthropic Messages API POST /v1/messages Google Gemini POST /v1beta/models/:model:generateContent Examples OpenAI SDK (Python) — route to Ollama from openai import OpenAI
Continue reading on Dev.to JavaScript
Opens in a new tab




