logo
welcome
TechCrunch

TechCrunch

A popular technique to make AI more efficient has drawbacks | TechCrunch

TechCrunch
Summary
Nutrition label

84% Informative

Quantization is a technique to make AI models more efficient, but it has limits.

It may be better to just train a smaller model rather than cook down a big one, study says.

Google spent $191 million to train one of its flagship Gemini models, but could spend $6 billion a year training a model to answer half of all Google Search queries.

The study suggests that precisions lower than 7- or 8-bit may see a noticeable step down in quality.

The takeaway is that AI models are not fully understood, and known shortcuts that work in many kinds of computation don't work here.

Kumar acknowledges that his and his colleagues’ study was at relatively small scale — they plan to test more models in the future.