DEV Community

Cover image for It's Not Smarter Models — It's Cheaper Memory: TurboQuant's Real Impact, Wall Street Panic & Academic Storm
Yang Goufang
Yang Goufang

Posted on

It's Not Smarter Models — It's Cheaper Memory: TurboQuant's Real Impact, Wall Street Panic & Academic Storm

One-line summary: TurboQuant is a genuinely important engineering breakthrough — but Google's marketing, academic ethics controversy, and Wall Street's overreaction made the story far more dramatic than the technology itself.


0. What This Article Answers

Google Research published TurboQuant at ICLR 2026 (arXiv 2504.19874), claiming 6x memory compression, 8x speedup, and zero accuracy loss for LLM KV caches.

Then, in the same week:

  1. Global memory stocks lost over $90 billion in market cap
  2. An ETH Zürich researcher publicly accused the paper of academic plagiarism and experimental fraud
  3. Google released zero code — so the community reproduced it in days, with one person using Claude Code to read the math and build a full implementation in 7 days, adding his own research contributions on top

What kind of paper simultaneously blows up Wall Street, academia, and the open-source community?


1. Why KV Cache Is AI's Real Bottleneck

Before discussing TurboQuant, understand this: modern LLMs are not compute-bound — they're memory-bound.

When a model generates text, it must remember all prior conversation history (attention history). This intermediate result, called the KV Cache, grows linearly with context length.

Concrete numbers:

Model Context Length KV Cache Size
70B model 128K tokens ~40 GB
35B model 100K tokens ~20 GB

40 GB of KV Cache — larger than the model itself. This is what the industry calls the Memory Wall.

Your model may be "only" 8B parameters, but when you feed it a 100K-token codebase, VRAM gets devoured instantly. This is why memory is so expensive, and why HBM is AI hardware's scarcest resource.

TurboQuant's goal: not making models smarter, but making AI's "memory" extremely cheap.


2. Technical Breakdown: What TurboQuant Actually Does

TurboQuant is fundamentally two engineering techniques combined:

PolarQuant: Making Data "Compressible"

Traditional quantization's nemesis is outliers — extreme values that destroy compression precision.

PolarQuant applies a random rotation to data vectors, then converts to polar coordinates (angle + radius). Mathematically, this exploits the near-independence property of coordinates in high-dimensional space after random rotation, making the value distribution extremely stable.

Result: eliminates per-block normalization overhead, saving significant metadata space.

QJL (Quantized Johnson-Lindenstrauss): 1-Bit Error Correction

Compression is inherently lossy. QJL projects the quantization error and stores just a 1-bit sign (+/-) to correct it, ensuring attention inner-product computations stay on track.

One-line summary: rotate data to make it compressible, then use 1-bit to pull errors back.


3. Deconstructing the Hype: What Google Didn't Say Loudly

Google's headline claims: 6x memory reduction, 8x speed, zero accuracy loss.

As engineers, we need to unwrap the packaging.

"6x Memory Compression" — Roughly Correct, With Gaps

Source Compression Ratio
Google paper (3-bit) 6x
turboquant_plus community test (3-bit) 4.6–5.1x
turboquant_plus (4-bit) 3.8x
turboquant_plus (2-bit) 6.4x
tonbistudio PyTorch implementation ~5x

Verdict: ~4.6–5.7x at 3-bit, not exactly 6x. Directionally correct, but the marketing number runs high.

"8x Speedup" — The Number That Needs the Most Clarification

The 8x compares 4-bit vs FP32 attention logit computation on H100 — not end-to-end inference speed.

Community end-to-end benchmarks (llama.cpp / Metal):

Metric Result
Single-request TPS (Tokens Per Second) 7–24% slower than q8_0
System throughput 2–4x improvement (freed VRAM enables more concurrent requests)

Why does it get slower? Every token generated requires real-time dequantization of compressed KV cache on the GPU. We relieved the memory-bound bottleneck but shifted pressure to compute-bound.

This isn't a flaw — it's a trade-off: slight per-request TPS decrease for massive system-level scalability. But Google using "8x" as a headline number without clarifying it's attention-only is genuinely misleading.

"Zero Accuracy Loss" — Conditionally True

  • 3.5 bits: LongBench 50.06 matches FP32 baseline; Needle-in-Haystack: perfect 100 score (4K–104K) — genuinely lossless
  • 2.5 bits: The paper itself says "marginal degradation"
  • Extreme code reasoning scenarios: still needs observation

QJL's Real-World Performance: Community Pushback

This is the most important community finding: 6 independent teams confirmed that QJL (Algorithm 2 in the paper) actually degrades attention quality in practice.

Most community implementations have now dropped QJL entirely, using only MSE-optimal quantization (Algorithm 1). The paper's most elegant theoretical contribution turns out to be a net negative in production — a classic gap between academic claims and engineering reality.


4. Wall Street Panic: One Paper Evaporates $90 Billion

After Google promoted TurboQuant on its official blog on March 24, global memory stocks were hammered:

Stock Decline
Micron (MU) 6 consecutive down days, cumulative -20%, entered bear market
SK Hynix -6.23%
Samsung -4.8% (cumulative -20% over following weeks)
SanDisk -11% single day
Kioxia -6%
Total market cap evaporated >$90 billion

Citi cut Micron's price target. Korea's KOSPI fell from 6,300 to 5,000 in one month (TurboQuant was one of several factors).

But was the panic justified?

The Jevons Paradox from economics is worth considering: when a resource's efficiency improves and per-unit cost drops, total consumption explodes.

When long context becomes cheap, we won't buy less RAM. We'll run larger agent systems, longer context windows, more concurrent requests. Total memory demand will actually increase exponentially.

Multiple analysts maintained Buy ratings on memory stocks, arguing that efficiency gains have historically never reduced total demand — only accelerated adoption.


5. Academic Storm: ETH Zürich Accuses Plagiarism and Experimental Fraud

This is the most serious part of the entire story.

Jianyang Gao — ETH Zürich postdoctoral researcher and first author of RaBitQ — published a public statement identifying three problems:

Problem 1: Suspected Plagiarism

TurboQuant's core method (applying random rotation before quantization) has direct structural overlap with RaBitQ. The critical evidence:

TurboQuant's second author Majid Daliri proactively contacted the RaBitQ team in January 2025, requesting help debugging his own Python implementation based on RaBitQ.

This proves the TurboQuant team had detailed knowledge of RaBitQ's techniques. Yet the paper described RaBitQ as "grid-based PQ," deliberately omitting RaBitQ's shared random rotation step.

Problem 2: Theoretical Mischaracterization

The TurboQuant paper labels RaBitQ as "theoretically suboptimal" with "relatively coarse analysis."

However, RaBitQ's extended version, published at a top theoretical computer science conference, rigorously proves its error bounds reach asymptotic optimality (matching the Alon-Klartag bound).

Problem 3: Fabricated Experimental Comparison

This is the most egregious:

Test Subject Hardware
RaBitQ Single-core CPU + Python translation + multithreading disabled
TurboQuant NVIDIA A100 GPU

Then the paper reports "RaBitQ is several orders of magnitude slower." Daliri's own May 2025 email acknowledges: "we were using a single-core CPU instance, and multiprocessing was indeed disabled."

Timeline

Date Event
May 2024 RaBitQ posted to arXiv with full source code
Jan 2025 Daliri contacts Gao requesting debugging help
Apr 2025 TurboQuant appears on arXiv
May 2025 Gao emails detailed corrections; Daliri claims to inform co-authors, then stops responding
Nov 2025 Gao discovers unrevised paper submitted to ICLR
Jan 2026 ICLR accepts TurboQuant
Mar 2026 Google promotes paper; Gao goes public; Stanford NLP Group amplifies

TurboQuant team's response: Agreed to address Problems 2 and 3 only after ICLR concludes, but refused to discuss Problem 1 (methodological overlap), claiming "random rotation and JL transforms have become standard field techniques — it's infeasible to cite every method that employs them."


6. The Open-Source Counter-Strike: Claude Code Reproduction in 7 Days

Google released zero code. The community's response: we'll do it ourselves.

Dozens of independent implementations appeared within days. The most impressive: Tom Turney's turboquant_plus.

7 days, from scratch, reading math formulas with Claude Code. Not just reproduction — he added original research contributions:

Contribution Description
Sparse V Skips dequantization for 90% of low-weight V positions; +22.8% decode speed, zero accuracy loss
Temporal Decay Older tokens auto-downgrade precision, further compressing historical memory
Asymmetric K/V Allocation Keys at 4-bit, Values at 2-bit (because K/V norm disparities reach 4–182x)

Validated end-to-end on Qwen 3.5 35B-A3B (MoE) via llama.cpp Metal on Apple Silicon. 511+ tests, 100% coverage.

The significance goes beyond TurboQuant: when math formulas are clear enough, AI coding agents can go directly from paper to implementation. The "moat" of not releasing code is disappearing.


7. Real Impact on the AI Ecosystem

Agent Systems: From Reactive to Persistent

Agents are fundamentally "long memory + multi-step reasoning." Previously, agents would "forget" during long runs or costs would spike.

4–5x KV cache compression means:

  • Agents can retain extremely long task histories and sub-agent context
  • Multi-agent system costs drop dramatically — each agent previously consumed massive KV cache
  • Parallel agent count can multiply several-fold

Claude Code / Codex: Repository-Level Reasoning

Previously limited by KV cache, AI coding tools could only see partial code, constantly chunking. With cheaper memory, entire repos + git history fit in context without pain, enabling qualitative jumps in cross-file reasoning and large-scale refactoring.

Local AI: From Demo to Usable

People are already running 122B models on Apple Silicon with TurboQuant + llama.cpp for Claude Code-level tasks — no cloud, no API, no subscription. 35B + long-context inference on consumer hardware is now genuinely possible.

Structural Cost Shift in Inference

Before: cost ≈ model size
Now: cost ≈ KV cache × concurrency

With KV cache shrunk 4–5x, cloud providers can serve more users per machine. The next avalanche in API pricing is coming.


8. Conclusion: What's Next

TurboQuant's historical significance isn't that it makes AI smarter — it's that it changes the cost structure of using AI.

There's no free lunch — we traded slight TPS latency for freedom in context length and concurrency.

The predictable next step: LLM KV caches will evolve a L1/L2/L3 cache hierarchy similar to CPUs — hot data in uncompressed high-speed VRAM for TPS, cold historical data compressed via TurboQuant in slower tiers.

When memory is no longer a burden, AI is truly ready to take on complex engineering at scale.

But first, Google might want to address that academic ethics issue.


References


What's your take — is TurboQuant overhyped or underrated? Drop your thoughts below.

Top comments (0)