Google TurboQuant Compresses LLM Memory by 6x — and Rattled Memory Chip Stocks

Google Research has published TurboQuant, a new training-free algorithm that compresses the memory requirements of large language models by at least six times — and within 24 hours of its announcement, triggered a sell-off in memory chip stocks. Samsung, SK Hynix, and Micron all posted significant declines as investors worried the compression breakthrough could slow demand for High Bandwidth Memory (HBM) chips, the hardware category that has been the primary revenue driver for memory makers since the AI build-out began in 2023.
What TurboQuant Does
TurboQuant is a KV cache compression algorithm — the KV cache is the memory structure that LLMs use to store context during a conversation or inference run. As context windows have grown longer, KV cache size has become one of the primary bottlenecks in deploying large models at scale. Longer context means more memory, more bandwidth, and higher costs.
TurboQuant works in two stages. The first, called PolarQuant, uses random vector rotation to simplify data geometry before applying standard quantization. The second, called QJL (Quantized Johnson-Lindenstrauss), applies 1-bit residual compression to remaining errors using a mathematical error-checker to eliminate bias and improve accuracy. Together they compress KV cache from 16 bits down to approximately 3 bits — with no training or fine-tuning required. On NVIDIA H100 GPUs, 4-bit TurboQuant delivers up to 8x performance increase over 32-bit unquantized keys.
Why Memory Chip Investors Are Worried
The investor reaction reflects a long-running concern in the semiconductor market: algorithmic efficiency gains reducing the amount of hardware needed to run AI at scale. A 6x compression in KV cache memory translates directly to fewer HBM chips required per inference run — or the ability to run much larger models on existing hardware. Memory chip makers have benefited enormously from AI infrastructure spending, and any credible evidence that memory requirements could shrink threatens a key pillar of their growth thesis. This is the same dynamic driving Japan's massive bet on domestic chip manufacturing, as seen in Japan's USD 16.3 billion investment in Rapidus.
The algorithm was co-authored by Amir Zandieh and Vahab Mirrokni from Google Research and presented at ICLR 2026. Benchmarks were run on LongBench, Needle In A Haystack, ZeroSCROLLS, RULER, and L-Eval using Gemma and Mistral models. Internet commenters quickly nicknamed it "Pied Piper" — a reference to the fictional compression breakthrough in the Silicon Valley TV show.
Frequently Asked Questions
What is the KV cache and why does it matter?
The KV cache stores intermediate computation results that LLMs reference when generating text, allowing them to process long contexts without recomputing everything from scratch. As context windows have expanded to millions of tokens, KV cache memory requirements have grown dramatically, making it a key cost driver in AI inference.
Does TurboQuant require retraining existing models?
No. TurboQuant is training-free, meaning it can be applied to already-deployed models without any fine-tuning or additional training runs. This makes it practical to deploy on existing model weights immediately.
Has TurboQuant been released publicly?
An open-source release is expected around Q2 2026, aligned with the algorithm's presentation at ICLR 2026. The research builds on two prior Google publications: Quantized Johnson-Lindenstrauss (AAAI 2025) and PolarQuant (AISTATS 2026).
The Bottom Line
TurboQuant is the most credible algorithmic challenge to AI memory demand since the DeepSeek efficiency story earlier this year. A 6x compression ratio with zero accuracy loss and no retraining requirement is not a theoretical result — it is a practical tool deployable on existing models. Whether this translates to meaningfully lower memory chip demand depends on whether AI labs apply the savings toward lower costs or toward running even larger models. History suggests the latter: efficiency gains in AI have consistently been consumed by scaling ambition rather than reducing hardware spend.