跳至主導覽 跳至搜尋 跳過主要內容

Diffbias: Harnessing diffusion models’ prediction bias for adversarial patch defense

研究成果: Article同行評審

摘要

Adversarial patches pose a significant and real threat to deep neural networks, capable of inducing misclassification in realistic physical scenarios. Developing reliable and robust defense methods against these attacks is a critical application, and current research remains unsatisfactory. In this paper, we propose a novel framework that exploits the fact that unnatural perturbations introduced by adversarial patches can produce prediction biases significantly different from those of clean images during denoising. In the localization stage, our method focuses on the critical denoising steps through an adaptive temporal sampling strategy and introduces an energy metric that fuses kinetic and potential energy to quantify the degree of anomaly in the denoised trajectory. Furthermore, by combining this with the adaptive similarity weighting mechanism and the striding trajectory consistency analysis, our method effectively suppresses the interference of background noise, so as to achieve accurate locking of the patch area. In the restoration phase, the same diffusion model is applied to the patch region to restore the original visual content and integrity. This two-stage architecture shares a unified diffusion model, enabling the localization and inpainting processes to enhance the overall defense performance through information complementarity. Extensive experiments on the INRIA, COCO2017, and APRICOT datasets show that our approach achieves state-of-the-art detection performance under both digital and physical attack types without compromising the recognition accuracy of clean images.

原文English
文章編號133009
期刊Neurocomputing
676
DOIs
出版狀態Published - 1 5月 2026

指紋

深入研究「Diffbias: Harnessing diffusion models’ prediction bias for adversarial patch defense」主題。共同形成了獨特的指紋。

引用此