Optimizing Large Language Models through Quantization: A Comparative Analysis of PTQ and QAT Techniques Paper • 2411.06084 • Published Nov 9, 2024 • 1
DL-QAT: Weight-Decomposed Low-Rank Quantization-Aware Training for Large Language Models Paper • 2504.09223 • Published Apr 12 • 1