Neural Weight Compression for Language Models

Explore Neural Weight Compression (NWC), a novel framework utilizing autoencoders to compress language model weights while maintaining FP16-level accuracy at...

Level: advanced

By Unknown

Category: research