PRADA: Practical Black-box Adversarial Attacks against Neural Ranking Models

被引:4
|
作者
Wu, Chen [1 ,2 ]
Zhang, Ruqing [1 ,2 ]
Guo, Jiafeng [1 ,2 ]
De Rijke, Maarten [3 ]
Fan, Yixing [2 ,4 ]
Cheng, Xueqi [2 ,4 ]
机构
[1] Inst Comp Technol Acad Sci, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, 6 Kexueyuan South Rd, Beijing 100190, Peoples R China
[3] Univ Amsterdam, NL-1012WX Amsterdam, Netherlands
[4] Chinese Acad Sci, Inst Comp Technol, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Adversarial attack; decision-based black-box attack setting; neural ranking models; SPAM DETECTION;
D O I
10.1145/3576923
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Neural ranking models (NRMs) have shown remarkable success in recent years, especially with pre-trained language models. However, deep neural models are notorious for their vulnerability to adversarial examples. Adversarial attacks may become a new type of web spamming technique given our increased reliance on neural information retrieval models. Therefore, it is important to study potential adversarial attacks to identify vulnerabilities of NRMs before they are deployed. In this article, we introduce the Word Substitution Ranking Attack (WSRA) task against NRMs, which aims at promoting a target document in rankings by adding adversarial perturbations to its text. We focus on the decision-based black-box attack setting, where the attackers cannot directly get access to the model information, but can only query the target model to obtain the rank positions of the partial retrieved list. This attack setting is realistic in real-world search engines. We propose a novel Pseudo Relevance-based ADversarial ranking Attack method (PRADA) that learns a surrogate model based on Pseudo Relevance Feedback (PRF) to generate gradients for finding the adversarial perturbations. Experiments on two web search benchmark datasets show that PRADA can outperform existing attack strategies and successfully fool the NRM with small indiscernible perturbations of text.
引用
收藏
页数:27
相关论文
共 50 条
  • [1] Topic-oriented Adversarial Attacks against Black-box Neural Ranking Models
    Liu, Yu-An
    Zhang, Ruqing
    Guo, Jiafeng
    de Rijke, Maarten
    Chen, Wei
    Fan, Yixing
    Cheng, Xueqi
    [J]. PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 1700 - 1709
  • [2] Multi-granular Adversarial Attacks against Black-box Neural Ranking Models
    Liu, Yu-An
    Zhang, Ruqing
    Guo, Jiafeng
    de Rijke, Maarten
    Fan, Yixing
    Cheng, Xueqi
    [J]. PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 1391 - 1400
  • [3] Black-Box Adversarial Attacks against Audio Forensics Models
    Jiang, Yi
    Ye, Dengpan
    [J]. SECURITY AND COMMUNICATION NETWORKS, 2022, 2022
  • [4] Boundary Defense Against Black-box Adversarial Attacks
    Aithal, Manjushree B.
    Li, Xiaohua
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2349 - 2356
  • [5] Black-box Adversarial Attacks on Video Recognition Models
    Jiang, Linxi
    Ma, Xingjun
    Chen, Shaoxiang
    Bailey, James
    Jiang, Yu-Gang
    [J]. PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 864 - 872
  • [6] Practical Black-Box Attacks against Machine Learning
    Papernot, Nicolas
    McDaniel, Patrick
    Goodfellow, Ian
    Jha, Somesh
    Celik, Z. Berkay
    Swami, Ananthram
    [J]. PROCEEDINGS OF THE 2017 ACM ASIA CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (ASIA CCS'17), 2017, : 506 - 519
  • [7] Simple Black-Box Adversarial Attacks on Deep Neural Networks
    Narodytska, Nina
    Kasiviswanathan, Shiva
    [J]. 2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, : 1310 - 1318
  • [8] Simple Black-box Adversarial Attacks
    Guo, Chuan
    Gardner, Jacob R.
    You, Yurong
    Wilson, Andrew Gordon
    Weinberger, Kilian Q.
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [9] Heuristic Black-Box Adversarial Attacks on Video Recognition Models
    Wei, Zhipeng
    Chen, Jingjing
    Wei, Xingxing
    Jiang, Linxi
    Chua, Tat-Seng
    Zhou, Fengfeng
    Jiang, Yu-Gang
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 12338 - 12345
  • [10] Black-box attacks against log anomaly detection with adversarial examples
    Lu, Siyang
    Wang, Mingquan
    Wang, Dongdong
    Wei, Xiang
    Xiao, Sizhe
    Wang, Zhiwei
    Han, Ningning
    Wang, Liqiang
    [J]. INFORMATION SCIENCES, 2023, 619 : 249 - 262