A Caltech Lab at PrismML Just Fit an 8 Billion Parameter AI Model Into 1.15 GB. Announcing a Breakthrough in AI Compression: ...
The idea of simplifying model weights isn’t a completely new one in AI research. For years, researchers have been experimenting with quantization techniques that squeeze their neural network weights ...