LO-BCQ: Block Clustered Quantization for 4-bit (W4A4) LLM Inference
arXiv:2502.05376v2 Announce Type: replace Abstract: Post-training quantization (PTQ) is a promising approach to reducing the storage and computational requirements of large language models (LLMs) without...