@_EldarKurtic
Our flagship paper on how far careful quantization can really go in practice got accepted as an oral at ACL 2025 (top 8%)! 🥳 Turns out, old-school methods like GPTQ, SmoothQuant, and RTN are quite good when tuned properly. All of the tricks are already in LLM-Compressor! https://t.co/lX1X0XGGeZ