Moonshot AI quietly rolled Kimi Code K2.6 out to all paid subscribers on April 13, and the pricing sheet alone is worth a look — you're paying about a sixth of what Claude Code costs for something that benchmarks in the same neighborhood. The catch isn't the model. It's how Moonshot ships it.

For backend and infra folks who've watched their Anthropic bill climb as agentic coding eats more of the workflow, this matters. So let's talk about what's actually there, and where the rough edge cuts.

What Shipped

K2.6 is the latest turn of Moonshot's trillion-parameter MoE — 32B active per forward pass, 256K context, and a stated 100 tokens/second output. The model powers Moonshot's terminal agent, Kimi Code CLI, which is the company's answer to Claude Code, Codex CLI, and Gemini CLI. You install it with one line:

curl -L code.kimi.com/install.sh | bash

The CLI reads and edits files, runs shell commands, fetches web pages, and autonomously plans multi-step tasks. If you've used Claude Code, the mental model transfers one-for-one. Moonshot ran a roughly week-long closed beta before the April 13 rollout, so the bigger shift here is that K2.6 is now the default model when you invoke the CLI, not a separate SKU you have to opt into.

Official K2.6 benchmark numbers are pending. The baseline — K2.5 — already scores 76.8% on SWE-Bench Verified, which is Sonnet-class. Beta testers report cleaner reasoning traces, better multi-step agent plans, and fewer tool-call execution failures on 2.6. That's three soft claims, not hard numbers, but the devs I've talked to weren't faking enthusiasm.

The Price Comparison That Matters

Here's the table everyone actually cares about:

Model Input ($/MTok) Output ($/MTok) SWE-Bench Verified
Claude Sonnet 4.6 $3.00 $15.00 ~77%
GPT-5.4 $1.25 $10.00 ~74%
Kimi Code K2.6 $0.60 $2.50 ~77% (K2.5 base)
GLM-5.1 $0.20 $1.10 80.2%

Honestly, GLM-5.1 undercuts everyone on raw cost, but its CLI story isn't as polished yet. If what you want is a drop-in Claude Code alternative with similar UX, this is the closest thing shipping today. Six-times-cheaper output pricing is real money when your team is running iteration-heavy refactors on a monorepo.

Agent Swarm — Moonshot's name for parallel sub-agent execution — coordinates up to 100 sub-agents and claims up to 4.5x speedup on parallelizable tasks like batch refactors. I haven't independently verified the 4.5x, and the 100-agent ceiling is mostly marketing (you'll never use that many in practice), but the underlying pattern is real: task decomposition into independent work units that run concurrently. It's the same thing Claude's managed agents and OpenAI's Swarm-style architectures are circling.

The Version Pinning Problem

Here's where Moonshot stumbled. The CLI exposes the model under a unified identifier — kimi-for-coding — which auto-resolves to whatever snapshot Moonshot considers "current." You cannot pin to a specific build. You cannot roll back. If they push a regression, you eat the regression, and there's no flag to hold on a known-good version.

For individual devs hacking on side projects, fine. For teams running coding agents inside CI pipelines or production bots, that's a real blocker. Reproducibility of agent runs matters — when a refactor PR starts looking weird, you want to know whether the model changed under you or the codebase did. Anthropic lets you pin claude-sonnet-4-6 for exactly this reason. OpenAI's model IDs freeze. Moonshot's does not.

I suspect they'll fix this because the big buyers — enterprise dev teams — won't adopt without it. But today, if your workflow depends on deterministic agent behavior, Kimi Code is a tool for your local terminal, not your CI runner.

Rate Limits Are the Other Footgun

Subscription tiers cap you at 300–1,200 API calls per rolling 5-hour window with 30 concurrent requests max. That sounds generous. It is, for interactive coding. It is not generous if you're fanning out parallel agents on a large codebase. Agent Swarm's 100-agent claim runs headfirst into the 30-concurrent cap. Read the fine print before you build your pipeline around the marketing number.

So Should You Switch?

Straight answer: try it for exploratory work, don't move your production agents yet.

The economics are genuinely attractive. If your team spends a few grand a month on Claude for everyday refactors, boilerplate generation, and test writing, this covers that workload at a small fraction of the price. Claude still pulls ahead on long multi-constraint agent chains and nuanced English prompting — the kind of tasks where the model has to hold five constraints in working memory while threading a refactor across twelve files. That's a narrower gap than it was six months ago, but it exists.

The version pinning and rate limit stories mean this is a tool you evaluate on a laptop first, not something you drop into a deploy pipeline on day one. Moonshot has the option of fixing both. They probably will. When they do, the question for Anthropic's coding-tool margin gets uncomfortable — because the quality gap is closing and the price gap isn't.

Tune in next release cycle. Or don't — you'll see it in your cloud bill either way.