What Makes Low-Bit Quantization-Aware Training Work for Reasoning LLMs? A Systematic Study

This systematic study investigates low-bit quantization-aware training for reasoning LLMs, revealing how calibration alignment and initialization strategies ...

Level: advanced

By Keyu Lv and 8 other authors

Category: research