
I Benchmarked 8 Cloud AI Models. The 397B One Lost to a 1.6s Model.
title: "I Benchmarked 8 Cloud AI Models. The 397B One Lost to a 1.6s Model." description: "Speed tests, tool calling, reasoning, and code generation across 8 cloud models. The results surprised me." tags: ["ai", "ollama", "benchmark", "cloud"] published: false I Benchmarked 8 Cloud AI Models. The 397B One Lost to a 1.6s Model. I run a self-hosted AI agent setup with OpenClaw, and I've been using qwen3.5:397b-cloud as my default model for months. It's big, it's powerful, it's from Alibaba. What more could you want? Turns out, you might want speed . And accuracy. Today I ran a comprehensive benchmark across 8 cloud models available through Ollama. The results were... humbling. My default 397B parameter model got beaten by a model that's 14x faster . The Setup I tested each model on three tasks: Math : Simple arithmetic (23×17+5) Code : Python string reverse one-liner Logic : The classic bat-and-ball puzzle (bat + ball = $1.10, bat costs $1 more than ball, what's the ball's price?) I also
Continue reading on Dev.to
Opens in a new tab



