AQLM: Extreme Compression of Large Language Models via Additive Quantization #5984
joseph777111
started this conversation in
Ideas
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Official Pytorch repository for Extreme Compression of Large Language Models via Additive Quantization
https://arxiv.org/pdf/2401.06118.pdf
https://github.com/Vahe1994/AQLM
Thoughts?
Beta Was this translation helpful? Give feedback.
All reactions