Executive Summary: As medical AI systems—particularly those used in radiology—become increasingly integrated into clinical workflows, their susceptibility to adversarial manipulation grows. This paper examines a novel threat vector: the exploitation of 32-bit floating-point (FP32) quantized weights in deep learning models deployed for radiological diagnosis. We demonstrate that even subtle perturbations introduced during model quantization can be weaponized to induce systematic misclassification in X-ray and CT scan analysis. Our findings reveal that adversarial poisoning through quantized weight manipulation can bypass both traditional and modern defense mechanisms, posing a critical risk to patient safety and diagnostic integrity. This research serves as a call to action for healthcare institutions and AI developers to adopt robust, quantization-aware security measures in medical AI deployment.
The integration of deep learning models into radiology—spanning X-ray, CT, MRI, and PET imaging—has revolutionized diagnostic accuracy, workflow efficiency, and early disease detection. Models such as DenseNet, ResNet, and Vision Transformers (ViTs) trained on large-scale medical imaging datasets now approach or surpass human expert performance in certain tasks, including lung nodule detection and breast cancer screening.
However, these models are computationally intensive. To deploy them on edge devices or in resource-constrained environments, developers increasingly rely on quantization—a technique that reduces the precision of model weights and activations from 32-bit floating-point (FP32) to lower-bit representations such as 8-bit integers (INT8) or even 4-bit floating-point (FP4). Quantization reduces memory footprint, accelerates inference, and lowers power consumption—critical for portable radiology devices.
While quantization is well-studied for performance and efficiency, its security implications remain under-explored. Most quantization pipelines assume model weights are trustworthy, but this assumption is fragile in adversarial contexts.
We introduce a new attack class: Quantized Weight Poisoning (QWP). In QWP, an adversary with access to a model's FP32 checkpoint introduces subtle, structured perturbations into the weights before quantization. These perturbations are designed to:
The attack pipeline consists of four stages:
We evaluated QWP on a benchmark radiology dataset consisting of 12,400 chest X-rays (CheXpert subset) with five pathology classes: Atelectasis, Cardiomegaly, Effusion, Infiltration, and No Finding. A DenseNet-121 model was trained to 92% AUC on the validation set. We then applied targeted adversarial poisoning at the convolutional layer weights using a projected gradient descent (PGD) strategy constrained to alter only 0.5% of weights.
After quantization to INT8 using TensorRT, we observed:
Further analysis revealed that quantization acts as a non-linear filter, selectively amplifying adversarial components while suppressing benign noise—a phenomenon we term quantization-induced adversarial amplification.
The core insight lies in the interaction between adversarial perturbations and quantization noise. Adversarial perturbations are typically small in magnitude but highly structured, often aligned with the model's decision boundary gradients. When applied to FP32 weights, these perturbations are invisible to standard validation tools.
During quantization, each FP32 weight is mapped to the nearest representable quantized value. However, the rounding function is non-differentiable and non-linear. This non-linearity can:
Our ablation studies confirm that the attack fails if quantization is disabled or replaced with high-precision inference—highlighting the pivotal role of quantization in enabling the exploit.
The implications of QWP are profound:
Current medical AI validation protocols (e.g., DICOM conformance, model drift monitoring) do not account for quantization-level tampering, leaving a critical security blind spot.
To counter QWP, we propose a multi-layered defense framework:
Retrain models using quantization-aware training (QAT) with adversarial examples generated during the FP32 stage. This ensures robustness propagates through quantization. Tools like NVIDIA's TensorRT and Google's TFLite now support QAT, but adoption in medical AI is lagging.
Implement