Robust Contrastive Language-Image Pre-training against Data Poisoning and Backdoor Attacks

被引:0
|
作者
Yang, Wenhan [1 ]
Gao, Jingdong [1 ]
Mirzasoleiman, Baharan [1 ]
机构
[1] UCLA, Comp Sci Dept, Los Angeles, CA 90024 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Contrastive vision-language representation learning has achieved state-of-the-art performance for zero-shot classification, by learning from millions of image-caption pairs crawled from the internet. However, the massive data that powers large multimodal models such as CLIP, makes them extremely vulnerable to various types of targeted data poisoning and backdoor attacks. Despite this vulnerability, robust contrastive vision-language pre-training against such attacks has remained unaddressed. In this work, we propose ROCLIP, the first effective method for robust pre-training multimodal vision-language models against targeted data poisoning and backdoor attacks. ROCLIP effectively breaks the association between poisoned image-caption pairs by considering a relatively large and varying pool of random captions, and matching every image with the text that is most similar to it in the pool instead of its own caption, every few epochs.It also leverages image and text augmentations to further strengthen the defense and improve the performance of the model. Our extensive experiments show that ROCLIP renders state-of-the-art targeted data poisoning and backdoor attacks ineffective during pre-training CLIP models. In particular, ROCLIP decreases the success rate for targeted data poisoning attacks from 93.75% to 12.5% and that of backdoor attacks down to 0%, while improving the model's linear probe performance by 10% and maintains a similar zero shot performance compared to CLIP. By increasing the frequency of matching, ROCLIP is able to defend strong attacks, which add up to 1% poisoned examples to the data, and successfully maintain a low attack success rate of 12.5%, while trading off the performance on some tasks (1).
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Understanding Chinese Video and Language via Contrastive Multimodal Pre-Training
    Lei, Chenyi
    Luo, Shixian
    Liu, Yong
    He, Wanggui
    Wang, Jiamang
    Wang, Guoxin
    Tang, Haihong
    Miao, Chunyan
    Li, Houqiang
    [J]. PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 2567 - 2576
  • [42] Leveraging Contrastive Language–Image Pre-Training and Bidirectional Cross-attention for Multimodal Keyword Spotting
    Liu, Dong
    Mao, Qirong
    Gao, Lijian
    Wang, Gang
    [J]. Engineering Applications of Artificial Intelligence, 2024, 138
  • [43] Knowledge Boosting: Rethinking Medical Contrastive Vision-Language Pre-training
    Chen, Xiaofei
    He, Yuting
    Xue, Cheng
    Ge, Rongjun
    Li, Shuo
    Yang, Guanyu
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT I, 2023, 14220 : 405 - 415
  • [44] Contrastive Learning for Self-Supervised Pre-Training of Point Cloud Segmentation Networks With Image Data
    Janda, Andrej
    Wagstaff, Brandon
    Ng, Edwin G.
    Kelly, Jonathan
    [J]. 2023 20TH CONFERENCE ON ROBOTS AND VISION, CRV, 2023, : 145 - 152
  • [45] Sculpting Holistic 3D Representation in Contrastive Language-Image-3D Pre-Training
    Gao, Yipeng
    Wang, Zeyu
    Zheng, Wei-Shi
    Xie, Cihang
    Zhou, Yuyin
    [J]. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2024, : 22998 - 23008
  • [46] Data-Efficient Contrastive Language-Image Pretraining: Prioritizing Data Quality over Quantity
    Joshi, Siddharth
    Jain, Arnav
    Payani, Ali
    Mirzasoleiman, Baharan
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [47] Conditional Embedding Pre-Training Language Model for Image Captioning
    Li, Pengfei
    Zhang, Min
    Lin, Peijie
    Wan, Jian
    Jiang, Ming
    [J]. NEURAL PROCESSING LETTERS, 2022, 54 (06) : 4987 - 5003
  • [48] Conditional Embedding Pre-Training Language Model for Image Captioning
    Pengfei Li
    Min Zhang
    Peijie Lin
    Jian Wan
    Ming Jiang
    [J]. Neural Processing Letters, 2022, 54 : 4987 - 5003
  • [49] Combining Defences Against Data-Poisoning Based Backdoor Attacks on Neural Networks
    Milakovic, Andrea
    Mayer, Rudolf
    [J]. DATA AND APPLICATIONS SECURITY AND PRIVACY XXXVI, DBSEC 2022, 2022, 13383 : 28 - 47
  • [50] Fine-grained knowledge about manipulable objects is well-predicted by contrastive language image pre-training
    Walbrin, Jon
    Sossounov, Nikita
    Mahdiani, Morteza
    Vaz, Igor
    Almeida, Jorge
    [J]. ISCIENCE, 2024, 27 (07)