JOURNAL ARTICLE

Universal Adversarial Perturbations for Vision-Language Pre-trained Models

Abstract

Vision-language pre-trained (VLP) models have been the foundation of numerous vision-language tasks. Given their prevalence, it becomes imperative to assess their adversarial robustness, especially when deploying them in security-crucial real-world applications. Traditionally, adversarial perturbations generated for this assessment target specific VLP models, datasets, and/or downstream tasks. This practice suffers from low transferability and additional computation costs when transitioning to new scenarios.\n\nIn this work, we thoroughly investigate whether VLP models are commonly sensitive to imperceptible perturbations of a specific pattern for the image modality. To this end, we propose a novel black-box method to generate Universal Adversarial Perturbations (UAPs), which is so called the <u>E</u>ffective and <u>T</u>ransferable <u>U</u>niversal Adversarial Attack (ETU), aiming to mislead a variety of existing VLP models in a range of downstream tasks. The ETU comprehensively takes into account the characteristics of UAPs and the intrinsic cross-modal interactions to generate effective UAPs. Under this regime, the ETU encourages both global and local utilities of UAPs. This benefits the overall utility while reducing interactions between UAP units, improving the transferability. To further enhance the effectiveness and transferability of UAPs, we also design a novel data augmentation method named ScMix. ScMix consists of self-mix and cross-mix data transformations, which can effectively increase the multi-modal data diversity while preserving the semantics of the original data. Through comprehensive experiments on various downstream tasks, VLP models, and datasets, we demonstrate that the proposed method is able to achieve effective and transferrable universal adversarial attacks.

Keywords:
Adversarial system Computer science Artificial intelligence Natural language processing Computer vision

Metrics

13
Cited By
8.30
FWCI (Field Weighted Citation Impact)
24
Refs
0.96
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Adversarial Robustness in Machine Learning
Physical Sciences →  Computer Science →  Artificial Intelligence
Multimodal Machine Learning Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Domain Adaptation and Few-Shot Learning
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

Rethinking Textual Adversarial Defense for Pre-Trained Language Models

Jiayi WangRongzhou BaoZhuosheng ZhangHai Zhao

Journal:   IEEE/ACM Transactions on Audio Speech and Language Processing Year: 2022 Vol: 30 Pages: 2526-2540
JOURNAL ARTICLE

A Survey of Vision-Language Pre-Trained Models

Yifan DuZikang LiuJunyi LiWayne Xin Zhao

Journal:   Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence Year: 2022 Pages: 5436-5443
© 2026 ScienceGate Book Chapters — All rights reserved.