Charts
DataOn-chain
VIP
Market Cap
API
Rankings
CoinOSNew
CoinClaw🦞
Language
  • 简体中文
  • 繁体中文
  • English
Leader in global market data applications, committed to providing valuable information more efficiently.

Features

  • Real-time Data
  • Special Features
  • AI Grid

Services

  • News
  • Open Data(API)
  • Institutional Services

Downloads

  • Desktop
  • Android
  • iOS

Contact Us

  • Chat Room
  • Business Email
  • Official Email
  • Official Verification

Join Community

  • Telegram
  • Twitter
  • Discord

© Copyright 2013-2026. All rights reserved.

简体繁體English
|Legacy

Google Shrinks AI Memory With No Accuracy Loss—But There's a Catch

CN
Decrypt
Follow
3 hours ago
AI summarizes in 5 seconds.

Google Research published TurboQuant on Wednesday, a compression algorithm that shrinks a major inference-memory bottleneck by at least 6x while maintaining zero loss in accuracy.


The paper is slated for presentation at ICLR 2026, and the reaction online was immediate.


Cloudflare CEO Matthew Prince called it Google's DeepSeek moment. Memory stock prices, including Micron, Western Digital, and Seagate, fell on the same day.





So is it real?


Quantization efficiency is a big achievement by itself. But "zero accuracy loss" needs context.


TurboQuant targets the KV cache—the chunk of GPU memory that stores everything a language model needs to remember during a conversation.


As context windows grow toward millions of tokens, those caches balloon into hundreds of gigabytes per session. That's the actual bottleneck. Not compute power but raw memory.


Traditional compression methods try to shrink those caches by rounding numbers down—from 32-bit floats to 16, to 8 to 4-bit integers, for example. To better understand it, think of shrinking an image from 4K, to full HD, to 720p and so. It’s easy to tell it’s the same image overall, but there’s more detail in 4K resolution.


The catch: they have to store extra "quantization constants" alongside the compressed data to keep the model from going stupid. Those constants add 1 to 2 bits per value, partially eroding the gains.


TurboQuant claims it eliminates that overhead entirely.


It does this via two sub-algorithms. PolarQuant separates magnitude from direction in vectors, and QJL (Quantized Johnson-Lindenstrauss) takes the tiny residual error left over and reduces it to a single sign bit, positive or negative, with zero stored constants.


The result, Google says, is a mathematically unbiased estimator for the attention calculations that drive transformer models.


In benchmarks using Gemma and Mistral, TurboQuant matched full-precision performance under 4x compression, including perfect retrieval accuracy on needle-in-haystack tasks up to 104,000 tokens.


For context on why those benchmarks matter, expanding a model's usable context without quality loss has been one of the hardest problems in LLM deployment.




Now, the fine print.


"Zero accuracy loss" applies to KV cache compression during inference—not to the model's weights. Compressing weights is a completely different, harder problem. TurboQuant doesn't touch those.


What it compresses is the temporary memory storing mid-session attention computations, which is more forgiving because that data can theoretically be reconstructed.


There's also the gap between a clean benchmark and a production system serving billions of requests. TurboQuant was tested on open-source models—Gemma, Mistral, Llama—not Google's own Gemini stack at scale.


Unlike DeepSeek's efficiency gains, which required deep architectural decisions baked in from the start, TurboQuant requires no retraining or fine-tuning and claims negligible runtime overhead. In theory, it drops straight into existing inference pipelines.


That's the part that spooked the memory hardware sector—because if it works in production, every major AI lab runs leaner on the same GPUs they already own.


The paper goes to ICLR 2026. Until it ships in production, the "zero loss" headline stays in the lab.


免责声明:本文章仅代表作者个人观点,不代表本平台的立场和观点。本文章仅供信息分享,不构成对任何人的任何投资建议。用户与作者之间的任何争议,与本平台无关。如网页中刊载的文章或图片涉及侵权,请提供相关的权利证明和身份证明发送邮件到support@aicoin.com,本平台相关工作人员将会进行核查。

油价一天翻倍,注册币安钱包抓差价!
广告
|
|
APP
Windows
Mac
Share To

X

Telegram

Facebook

Reddit

CopyLink

|
|
APP
Windows
Mac
Share To

X

Telegram

Facebook

Reddit

CopyLink

Selected Articles by Decrypt

2 hours ago
Australia Lays Groundwork for Tokenized Asset Markets After RBA Project
5 hours ago
Google Sets 2029 Deadline to Deal With Quantum Threat—Is It a Problem for Bitcoin?
6 hours ago
$15 Billion in Bitcoin Options Expire Friday as Trump\\\'s Iran Deadline Looms
View More

Table of Contents

|
|
APP
Windows
Mac
Share To

X

Telegram

Facebook

Reddit

CopyLink

Related Articles

avatar
avatarbitcoin.com
30 minutes ago
Crypto Market Maker Warning: Binance Lists Six Red Flags Traders Should Know
avatar
avatarbitcoin.com
1 hour ago
Strategy Elevates Bitcoin Security as Massive 762K BTC Holdings Raise Market Stakes
avatar
avatarbitcoin.com
2 hours ago
Grayscale Sees Crypto Valuations Recovering as Global Pressures Begin to Ease
avatar
avatarbitcoin.com
2 hours ago
Petroyuan Rising: Deutsche Bank Analysis Flags Iran War as Turning Point for US Dollar Dominance
avatar
avatarU.today
2 hours ago
\\\'We Engineer Volatility,\\\' Michael Saylor on Strategy\\\'s Performance
APP
Windows
Mac

X

Telegram

Facebook

Reddit

CopyLink