Skip to content

IST-DASLab/unified-sc-laws

Repository files navigation

Unified Scaling Laws for Compressed Repersentations

arXiv

Scaling laws have shaped recent advances in machine learning by enabling predictable scaling of model performance based on model size, computation, and data volume. Concurrently, the rise in computational cost for AI has motivated model compression techniques, notably quantization and sparsification, which have emerged to mitigate the steep computational demands associated with large-scale training and inference. This paper investigates the interplay between scaling laws and compression formats, exploring whether a unified scaling framework can accurately predict model performance when training occurs over various compressed representations, such as sparse, scalar-quantized, sparse-quantized or even vector-quantized formats. Our key contributions include validating a general scaling law formulation and showing that it is applicable both individually but also composably across compression types. Based on this, our main finding is demonstrating both theoretically and empirically that there exists a simple "capacity" metric -- based on the representation's ability to fit random Gaussian data -- which can robustly predict parameter efficiency across multiple compressed representations. On the practical side, we extend our formulation to directly compare the accuracy potential of different compressed formats, and to derive better algorithms for training over sparse-quantized formats.

Source Code

This repository is based on the epfml/schedules-and-scaling repository for their "Scaling Laws and Compute-Optimal Training Beyond Fixed Training Durations" and the IST-DASLab/QuEST "QuEST: Stable Training of LLMs with 1-Bit Weights and Activations" papers.

Cite This Work

@inproceedings{
unifiedscalinglawscompressed,
title={Unified Scaling Laws for Compressed Representations},
author={Andrei Panferov and Alexandra Volkova and Ionut-Vlad Modoranu and Vage Egiazarian and Mher Safaryan and Dan Alistarh},
booktitle={The Thirty-ninth Annual Conference on Neural Information Processing Systems},
year={2025},
url={https://openreview.net/forum?id=24wDPGiDzA}
}

About

This repository contains the code for the "Unified Scaling Laws for Compressed Representations" study.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors