Decaf: Data Distribution Decompose Attack Against Federated Learning

被引:0
|
作者
Dai, Zhiyang [1 ,2 ,3 ]
Gao, Yansong [4 ]
Zhou, Chunyi [5 ]
Fu, Anmin [1 ,2 ,3 ]
Zhang, Zhi [4 ]
Xue, Minhui [6 ]
Zheng, Yifeng [7 ]
Zhang, Yuqing [8 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Cyber Sci & Engn, Nanjing 210094, Peoples R China
[2] Xidian Univ, State Key Lab Integrated Serv Networks, Xian 710071, Peoples R China
[3] Minist Educ, Key Lab Cyberspace Secur, Zhengzhou 450001, Peoples R China
[4] Univ Western Australia, Dept Comp Sci & Software Engn, Perth, WA 6009, Australia
[5] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou 310058, Peoples R China
[6] CSIRO Data61, Sydney, NSW 2122, Australia
[7] Harbin Inst Technol, Shenzhen 518055, Peoples R China
[8] Univ Chinese Acad Sci, Natl Comp Network Intrus Protect Ctr, Beijing 101408, Peoples R China
基金
中国国家自然科学基金;
关键词
Data models; Training; Data privacy; Servers; Privacy; Generative adversarial networks; Distributed databases; Load modeling; Federated learning; Training data; privacy attack; data distribution decompose;
D O I
10.1109/TIFS.2024.3516545
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In contrast to prevalent Federated Learning (FL) privacy inference techniques such as generative adversarial networks attacks, membership inference attacks, property inference attacks, and model inversion attacks, we devise an innovative privacy threat: the Data Distribution Decompose Attack on FL, termed Decaf. This attack enables an honest-but-curious FL server to meticulously profile the proportion of each class owned by the victim FL user, divulging sensitive information like local market item distribution and business competitiveness. The crux of Decaf lies in the profound observation that the magnitude of local model gradient changes closely mirrors the underlying data distribution, including the proportion of each class. Decaf addresses two crucial challenges: accurately identify the missing/null class(es) given by any victim user as a premise and then quantify the precise relationship between gradient changes and each remaining non-null class. Notably, Decaf operates stealthily, rendering it entirely passive and undetectable to victim users regarding the infringement of their data distribution privacy. Experimental validation on five benchmark datasets (MNIST, FASHION-MNIST, CIFAR-10, FER-2013, and SkinCancer) employing diverse model architectures, including customized convolutional networks, standardized VGG16, and ResNet18, demonstrates Decaf's efficacy. Results indicate its ability to accurately decompose local user data distribution, regardless of whether it is IID or non-IID distributed. Specifically, the dissimilarity measured using $L_{\infty }$ distance between the distribution decomposed by Decaf and ground truth is consistently below 5% when no null classes exist. Moreover, Decaf achieves 100% accuracy in determining any victim user's null classes, validated through formal proof.
引用
收藏
页码:405 / 420
页数:16
相关论文
共 50 条
  • [1] Data distribution inference attack in federated learning via reinforcement learning support
    Yu, Dongxiao
    Zhang, Hengming
    Huang, Yan
    Xie, Zhenzhen
    HIGH-CONFIDENCE COMPUTING, 2025, 5 (01):
  • [2] LDIA: Label distribution inference attack against federated learning in edge computing
    Gu, Yuhao
    Bai, Yuebin
    JOURNAL OF INFORMATION SECURITY AND APPLICATIONS, 2023, 74
  • [3] Defending Against Data Poisoning Attack in Federated Learning With Non-IID Data
    Yin, Chunyong
    Zeng, Qingkui
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (02) : 2313 - 2325
  • [4] Defense against backdoor attack in federated learning
    Lu, Shiwei
    Li, Ruihu
    Liu, Wenbin
    Chen, Xuan
    COMPUTERS & SECURITY, 2022, 121
  • [5] A Data Reconstruction Attack against Vertical Federated Learning Based on Knowledge Transfer
    Suimon, Takumi
    Koizumi, Yuki
    Takemasa, Junji
    Hasegawa, Toni
    IEEE INFOCOM 2024-IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS, INFOCOM WKSHPS 2024, 2024,
  • [6] Securing federated learning: a defense strategy against targeted data poisoning attack
    Ansam Khraisat
    Ammar Alazab
    Moutaz Alazab
    Tony Jan
    Sarabjot Singh
    Md. Ashraf Uddin
    Discover Internet of Things, 5 (1):
  • [7] Practical Attribute Reconstruction Attack Against Federated Learning
    Chen, Chen
    Lyu, Lingjuan
    Yu, Han
    Chen, Gang
    IEEE TRANSACTIONS ON BIG DATA, 2024, 10 (06) : 851 - 863
  • [8] VagueGAN: A GAN-Based Data Poisoning Attack Against Federated Learning Systems
    Sun, Wei
    Gao, Bo
    Xiong, Ke
    Lu, Yang
    Wang, Yuwei
    2023 20TH ANNUAL IEEE INTERNATIONAL CONFERENCE ON SENSING, COMMUNICATION, AND NETWORKING, SECON, 2023,
  • [9] Mitigate Data Poisoning Attack by Partially Federated Learning
    Dam, Khanh Huu The
    Legay, Axel
    18TH INTERNATIONAL CONFERENCE ON AVAILABILITY, RELIABILITY & SECURITY, ARES 2023, 2023,
  • [10] Data Reconstruction Attack with Label Guessing for Federated Learning
    Jang, Jinhyeok
    Oh, Yoonju
    Ryu, Gwonsang
    Choi, Daeseon
    JOURNAL OF INTERNET TECHNOLOGY, 2023, 24 (04): : 893 - 903