核心概念
Widely used quantization methods for deploying large language models (LLMs) on commodity hardware can be exploited to create models that behave maliciously when quantized, even if they appear benign in their full-precision form.
Egashira, K., Vero, M., Staab, R., He, J., & Vechev, M. (2024). Exploiting LLM Quantization. Advances in Neural Information Processing Systems, 38.
This paper investigates the security implications of LLM quantization, particularly the potential for malicious actors to exploit zero-shot quantization methods to introduce vulnerabilities.