Skip to main content

Google just quietly shipped Pied Piper.

TurboQuant compresses the KV cache 6x at 3 bits with no fine-tuning. Nobody is talking about it.

March 22, 2026

Nobody is talking about this and it is driving me a little insane.

On March 24th, Google Research published a paper called TurboQuant. It is going to ICLR 2026 next month. The internet noticed it for about 36 hours -- mostly to make Silicon Valley jokes about Pied Piper, which, yes, fair -- and then moved on.

Here is what actually happened: Google published a training-free, model-agnostic compression algorithm that shrinks the KV cache by 6x at 3-4 bits with near-zero quality loss on H100s. No fine-tuning. No calibration data. No model-specific configuration. You apply it to any transformer and it works.

That is the thing. Let me say it again more slowly.

You have a model. You are serving it in production. Your KV cache is eating your GPU memory. Every long-context request expands it. You are capacity-constrained on how many concurrent users you can serve. You cannot add context length without adding hardware.

You add TurboQuant. You change nothing else. Your KV cache now takes 6x less memory. You either handle 6x more concurrent users on the same hardware, or you double your context window on the same hardware, or some combination. Eight times faster attention logit computation on H100s as a bonus.

No retraining. No fine-tuning. No model changes.

I have written many times about the memory wall in inference -- the idea that decode is memory-bound, that the KV cache growing with context length is the structural bottleneck, that adding more Tensor Core compute does not fix a problem that lives in HBM bandwidth and capacity. TurboQuant is the first thing I have seen that attacks that problem from a direction I did not expect.

Here is how it actually works, because the "two-stage compression pipeline" summary everyone is using tells you nothing useful.

Stage one is PolarQuant. You take the KV vectors -- the key and value tensors sitting in HBM waiting to be attended over -- and you apply a random orthogonal rotation. What this does: it spreads the energy of the vector uniformly across all its coordinates. Before rotation, certain coordinates carry disproportionate information (the "outlier channel" problem that breaks naive quantization -- some coordinates are huge, some are tiny, standard quantizers hate this). After rotation, every coordinate follows a predictable Beta distribution. Now you can apply a Lloyd-Max scalar quantizer -- derived from probability theory, not learned from data -- and the codebook is the same for every vector in every model. No per-block normalization constants. No overhead.

Stage two is QJL -- Quantized Johnson-Lindenstrauss. You take the tiny residual error left over from the PolarQuant stage and you apply a Johnson-Lindenstrauss transform to it. This reduces each residual value to a single sign bit. One bit. That one bit eliminates the systematic bias in attention score computation that would otherwise accumulate at extreme compression ratios.

The result: 3 bits per KV element. Down from 16 bits full precision. 6x compression. Mathematically near-optimal -- provably close to the information-theoretic lower bound for this compression problem. The benchmarks across LongBench, Needle-in-a-Haystack, ZeroSCROLLS, RULER, and L-Eval show essentially no quality loss at 4 bits and acceptable quality loss at 3 bits for models above 3B parameters.

The thing I keep coming back to: this is not a soft result.

Most KV cache compression papers show you cherry-picked benchmarks on small models with quality degradation that becomes obvious in production. TurboQuant's needle-in-a-haystack numbers are perfect across all tested sequence lengths at 4-bit. The mathematical framing is not hand-wavy -- PolarQuant is provably optimal under its assumptions, QJL has tight theoretical bounds, and the whole pipeline approaches the coding theory lower bound.

"But the 6x is relative to FP16 and production systems are already quantiz--" Yes. Real gains over already-quantized production deployments are smaller. Int8 KV caches are common, int4 less so. The paper compares against existing quantization baselines and still wins. The honest number is probably 2-3x improvement over what you are running today if you are already doing basic KV quantization. That is still an enormous number in a world where KV cache is the binding constraint on serving cost.

There is no official open-source release from Google yet -- expected Q2 2026. Community ports exist already. Someone built an MLX implementation in 25 minutes using GPT-5.4, which is its own kind of news. There's a llama.cpp integration in active development -- turbo3, turbo4, asymmetric K/V quantization with Sparse V attention gating layered on top. Someone ran a 104B parameter model at 128K context on a MacBook with turbo3 and 74GB peak memory.

A MacBook.

Cloudflare's CEO called this Google's DeepSeek moment. Memory chip stocks dropped at open the morning after the paper dropped. Both of those reactions are approximately correct and also slightly missing the point.

DeepSeek was about training efficiency -- doing more with less compute during the expensive, capital-intensive phase. TurboQuant is about inference efficiency -- serving more users at lower cost during the phase that scales with every request. They attack different parts of the cost curve. Both matter. The inference cost curve is the one that compounds with adoption.

The part that actually matters for people who run serving infrastructure: if this holds up at 70B+ scale (the paper only benchmarked up to 8B, which is a real caveat), the implications for multi-tenant serving are significant. You are currently capacity-constrained by KV cache per user per session. 6x compression means you are serving 6x more concurrent users before you hit the memory wall. Or you are allowing 6x longer context per user before you hit the limit. Your inference cost per user drops. Your hardware utilization on the same fleet increases.

That is not a marginal efficiency gain. That is a qualitative change in what is economically feasible to serve.

I am watching the llama.cpp integration closely. The official Google implementation drops Q2. If the quality numbers hold at production model sizes, this is going in every serious inference stack within six months.

Nobody is talking about it because it dropped on a Tuesday and Twitter spent 36 hours doing Pied Piper jokes and then moved on to whatever Elon said.

It was a good Pied Piper joke though.

the memory wall in inference is real.

i have written about it before -- the KV cache grows with context, your HBM fills up, your serving capacity is bounded by memory not compute, adding more Tensor Cores does not help.

turboquant is the first thing in two years of watching this space that actually addresses that constraint from the right direction.

not by adding hardware. by making the math more efficient.

watch for the q2 official release. that's when this stops being a paper and starts being something you can deploy.

i write these when i have something worth saying. no schedule. no algorithm. if you want to know when the next one goes up -- leave your email.

no spam. no sequence. just the note, when it exists.