Post-training quantization is standard for memory-efficient deployment of large language models.Recent work has shown that basic rounding-based quantization schemes like GGUF pose security risks.An attack has been introduced on GGUF quantization, exploiting quantization errors to construct malicious models.The attack demonstrated effectiveness on three popular large language models across various scenarios.