Google unveils TurboQuant, PolarQuant and more to cut LLM/vector search memory use, pressuring MU, WDC, STX & SNDK.
Forget the parameter race. Google's TurboQuant research compresses AI memory by 6x with zero accuracy loss. It's not available yet, but it points to where AI efficiency is headed.