GPA: Enhancing Generalizable Physical Adversarial Attacks Across Multiple Vision Tasks

Abstract

Adversarial attacks pose a significant challenge in deep learning, as carefully crafted perturbations can severely degrade even the most advanced models. In real-world scenarios, where the target models are often unknown, previous works often focus on creating adversarial patterns for specific known models, with the goal of generalizing these patterns to other models. However, such attacks relies heavily on prior model information, leading in poor generalization. To overcome this, we propose a novel method called GPA. Our solution includes an attention extraction module based on a pre-trained vision encoder, which captures precise and generalizable features of model attention on objects. We also introduce attack loss functions that divert attention away from target objects. Compared to state-of-the-art methods, our approach achieves superior attack performance across various downstream vision tasks, including object detection, instance segmentation, and depth estimation. Moreover, the adversarial patterns generated by GPA maintain their effectiveness in real-world scenarios.

Publication
ICASSP 2025-2025 IEEE International Conference on Acoustics, Speech and Signal Processing
Mingye Xie
Mingye Xie
PhD Candidate

Life itself is the most wonderful fairy tale.