DEFENDING ACTIVE LEARNING AGAINST ADVERSARIAL INPUTS IN AUTOMATED DOCUMENT CLASSIFICATION

被引:0
|
作者
Pi, Lei [1 ]
Lu, Zhuo [2 ]
Sagduyu, Yalin [3 ]
Chen, Su [1 ]
机构
[1] Univ Memphis, Memphis, TN 38152 USA
[2] Univ S Florida, Tampa, FL 33620 USA
[3] Intelligent Automat Inc, Rockville, MD 20855 USA
关键词
active learning; document classification; security and attacks; malicious inputs;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Business and government operations generate large volumes of documents to be categorized through machine learning techniques before dissemination and storage. One prerequisite in such classification is to properly choose training documents. Active learning emerges as a technique to achieve better accuracy with fewer training documents by choosing data to learn and querying oracles for unknown labels. In practice, such oracles are usually human analysts who are likely to make mistakes or, in some cases, even intentionally introduce erroneous labels for malicious purposes. We propose a risk-factor based strategy to defend active-learning based document classification against human mistakes or adversarial inputs. We show that the proposed strategy can substantially alleviate the damage caused by malicious labeling. Our experimental results demonstrate the effectiveness of our defense strategy in terms of maintaining accuracy against adversaries.
引用
收藏
页码:257 / 261
页数:5
相关论文
共 50 条
  • [31] DEEP ADVERSARIAL ACTIVE LEARNING WITH MODEL UNCERTAINTY FOR IMAGE CLASSIFICATION
    Zhu, Zheng
    Wang, Hongxing
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 1711 - 1715
  • [32] Defending Against Universal Perturbations With Shared Adversarial Training
    Mummadi, Chaithanya Kumar
    Brox, Thomas
    Metzen, Jan Hendrik
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 4927 - 4936
  • [33] Dynamic and Diverse Transformations for Defending Against Adversarial Examples
    Chen, Yongkang
    Zhang, Ming
    Li, Jin
    Kuang, Xiaohui
    Zhang, Xuhong
    Zhang, Han
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, 2022, : 976 - 983
  • [34] Defending Against Adversarial Attacks in Deep Neural Networks
    You, Suya
    Kuo, C-C Jay
    [J]. ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS, 2019, 11006
  • [35] Defending Against Adversarial Attacks in Speaker Verification Systems
    Chang, Li-Chi
    Chen, Zesheng
    Chen, Chao
    Wang, Guoping
    Bi, Zhuming
    [J]. 2021 IEEE INTERNATIONAL PERFORMANCE, COMPUTING, AND COMMUNICATIONS CONFERENCE (IPCCC), 2021,
  • [36] DifFilter: Defending Against Adversarial Perturbations With Diffusion Filter
    Chen, Yong
    Li, Xuedong
    Hu, Peng
    Peng, Dezhong
    Wang, Xu
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 6779 - 6794
  • [37] DCAL: A New Method for Defending Against Adversarial Examples
    Lin, Xiaoyu
    Cao, Chunjie
    Wang, Longjuan
    Liu, Zhiyuan
    Li, Mengqian
    Ma, Haiying
    [J]. ARTIFICIAL INTELLIGENCE AND SECURITY, ICAIS 2022, PT II, 2022, 13339 : 38 - 50
  • [38] DEFENDING GRAPH CONVOLUTIONAL NETWORKS AGAINST ADVERSARIAL ATTACKS
    Ioannidis, Vassilis N.
    Giannakis, Georgios B.
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 8469 - 8473
  • [39] Minority Reports Defense: Defending Against Adversarial Patches
    McCoyd, Michael
    Park, Won
    Chen, Steven
    Shah, Neil
    Roggenkemper, Ryan
    Hwang, Minjune
    Liu, Jason Xinyu
    Wagner, David
    [J]. APPLIED CRYPTOGRAPHY AND NETWORK SECURITY WORKSHOPS, ACNS 2020, 2020, 12418 : 564 - 582
  • [40] Defending Against Model Inversion Attack by Adversarial Examples
    Wen, Jing
    Yiu, Siu-Ming
    Hui, Lucas C. K.
    [J]. PROCEEDINGS OF THE 2021 IEEE INTERNATIONAL CONFERENCE ON CYBER SECURITY AND RESILIENCE (IEEE CSR), 2021, : 551 - 556