DEFENDING ACTIVE LEARNING AGAINST ADVERSARIAL INPUTS IN AUTOMATED DOCUMENT CLASSIFICATION

被引:0
|
作者
Pi, Lei [1 ]
Lu, Zhuo [2 ]
Sagduyu, Yalin [3 ]
Chen, Su [1 ]
机构
[1] Univ Memphis, Memphis, TN 38152 USA
[2] Univ S Florida, Tampa, FL 33620 USA
[3] Intelligent Automat Inc, Rockville, MD 20855 USA
关键词
active learning; document classification; security and attacks; malicious inputs;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Business and government operations generate large volumes of documents to be categorized through machine learning techniques before dissemination and storage. One prerequisite in such classification is to properly choose training documents. Active learning emerges as a technique to achieve better accuracy with fewer training documents by choosing data to learn and querying oracles for unknown labels. In practice, such oracles are usually human analysts who are likely to make mistakes or, in some cases, even intentionally introduce erroneous labels for malicious purposes. We propose a risk-factor based strategy to defend active-learning based document classification against human mistakes or adversarial inputs. We show that the proposed strategy can substantially alleviate the damage caused by malicious labeling. Our experimental results demonstrate the effectiveness of our defense strategy in terms of maintaining accuracy against adversaries.
引用
收藏
页码:257 / 261
页数:5
相关论文
共 50 条
  • [1] CardioDefense: Defending against adversarial attack in ECG classification with adversarial distillation training
    Shao, Jiahao
    Geng, Shijia
    Fu, Zhaoji
    Xu, Weilun
    Liu, Tong
    Hong, Shenda
    [J]. BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 91
  • [2] On the Effectiveness of Adversarial Training in Defending against Adversarial Example Attacks for Image Classification
    Park, Sanglee
    So, Jungmin
    [J]. APPLIED SCIENCES-BASEL, 2020, 10 (22): : 1 - 16
  • [3] Defending Deep Learning Models Against Adversarial Attacks
    Mani, Nag
    Moh, Melody
    Moh, Teng-Sheng
    [J]. INTERNATIONAL JOURNAL OF SOFTWARE SCIENCE AND COMPUTATIONAL INTELLIGENCE-IJSSCI, 2021, 13 (01): : 72 - 89
  • [4] Evidential classification for defending against adversarial attacks on network traffic
    Beechey, Matthew
    Lambotharan, Sangarapillai
    Kyriakopoulos, Konstantinos G.
    [J]. INFORMATION FUSION, 2023, 92 : 115 - 126
  • [5] Defending against Adversarial Attacks in Federated Learning on Metric Learning Model
    Gu, Zhipin
    Shi, Jiangyong
    Yang, Yuexiang
    He, Liangzhong
    [J]. 2023 IEEE 22ND INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, BIGDATASE, CSE, EUC, ISCI 2023, 2024, : 197 - 206
  • [6] Defending non-Bayesian learning against adversarial attacks
    Su, Lili
    Vaidya, Nitin H.
    [J]. DISTRIBUTED COMPUTING, 2019, 32 (04) : 277 - 289
  • [7] Defending non-Bayesian learning against adversarial attacks
    Lili Su
    Nitin H. Vaidya
    [J]. Distributed Computing, 2019, 32 : 277 - 289
  • [8] Defending Network IDS against Adversarial Examples with Continual Learning
    Kozal, Jedrzej
    Zwolinska, Justyna
    Klonowski, Marek
    Wozniak, Michal
    [J]. 2023 23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW 2023, 2023, : 60 - 69
  • [9] Making Machine Learning Robust Against Adversarial Inputs
    Goodfellow, Ian
    McDaniel, Patrick
    Papernot, Nicolas
    [J]. COMMUNICATIONS OF THE ACM, 2018, 61 (07) : 56 - 66
  • [10] Defending Against Adversarial Attacks on Time-series with Selective Classification
    Kuehne, Joana
    Guehmann, Clemens
    [J]. 2022 PROGNOSTICS AND HEALTH MANAGEMENT CONFERENCE, PHM-LONDON 2022, 2022, : 169 - 175