Preserving Data Utility in Differentially Private Smart Home Data

被引:0
|
作者
Stirapongsasuti, Sopicha [1 ]
Tiausas, Francis Jerome [1 ]
Nakamura, Yugo [2 ]
Yasumoto, Keiichi [1 ,3 ]
机构
[1] Nara Inst Sci & Technol, Ikoma, Nara 6300192, Japan
[2] Kyushu Univ, Dept Informat Sci & Elect Engn, Fukuoka 8190395, Japan
[3] RIKEN, Ctr Adv Intelligence Project AIP, Tokyo 1030027, Japan
关键词
Differential privacy; machine learning; privacy; smart home; PRESERVATION; EFFICIENT; SYSTEM; CARE;
D O I
10.1109/ACCESS.2024.3390039
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The development of smart sensors and appliances can provide a lot of services. Nevertheless, the act of aggregating data containing sensitive information related to privacy in a single location poses significant issues. Such information can be misused by a malicious attacker. Also, some previous studies attempted to apply privacy mechanisms, but they decreased data utility. In this paper, we propose privacy protection mechanisms to preserve privacy-sensitive sensor data generated in a smart home. We leverage R & eacute;nyi differential privacy (RDP) to preserve privacy. However, the preliminary result showed that using only RDP still significantly decreases the utility of data. Thus, a novel scheme called feature merging anonymization (FMA) is proposed to preserve privacy while maintaining data utility by merging feature dataframes of the same activities from other homes. Also, the expected trade-off is defined so that data utility should be greater than the privacy preserved. To evaluate the proposed techniques, we define privacy preservation and data utility as inverse accuracy of person identification (PI) and accuracy of activity recognition (AR), respectively. We trained the AR and PI models for two cases with and without FMA, using 2 smart-home open datasets i.e. the HIS and Toyota dataset. As a result, we could lower the accuracy of PI in the HIS and Toyota dataset to 73.85% and 41.18% with FMA respectively compared to 100% without FMA, while maintaining the accuracy of AR at 94.62% and 87.3% with FMA compared to 98.58% and 89.28% without FMA in the HIS and Toyota dataset, respectively. Another experiment was conducted to explore the feasibility of implementing FMA in a local server by partially merging frames of the original activity with frames of other activities at different merging ratios. The results show that the local server can still satisfy the expected trade-off at some ratios.
引用
收藏
页码:56571 / 56581
页数:11
相关论文
共 50 条
  • [41] PCKV: Locally Differentially Private Correlated Key-Value Data Collection with Optimized Utility
    Gu, Xiaolan
    Li, Ming
    Cheng, Yueqiang
    Xiong, Li
    Cao, Yang
    PROCEEDINGS OF THE 29TH USENIX SECURITY SYMPOSIUM, 2020, : 967 - 984
  • [42] Examining the Utility of Differentially Private Synthetic Data Generated using Variational Autoencoder with TensorFlow Privacy
    Tai, Bo-Chen
    Li, Szu-Chuang
    Huang, Yennun
    Wang, Pang-Chieh
    2022 IEEE 27TH PACIFIC RIM INTERNATIONAL SYMPOSIUM ON DEPENDABLE COMPUTING (PRDC), 2022, : 236 - 241
  • [43] Privacy-preserving data infrastructure for smart home appliances based on the Octopus DHT
    Fabian, Benjamin
    Feldhaus, Tobias
    COMPUTERS IN INDUSTRY, 2014, 65 (08) : 1147 - 1160
  • [44] PrivPfC: differentially private data publication for classification
    Dong Su
    Jianneng Cao
    Ninghui Li
    Min Lyu
    The VLDB Journal, 2018, 27 : 201 - 223
  • [45] Differentially Private Publication of Vertically Partitioned Data
    Tang, Peng
    Cheng, Xiang
    Su, Sen
    Chen, Rui
    Shao, Huaxi
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2021, 18 (02) : 780 - 795
  • [46] A differentially private method for crowdsourcing data submission
    Zhang, Lefeng
    Xiong, Ping
    Ren, Wei
    Zhu, Tianqing
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2019, 31 (19):
  • [47] Research on Differentially Private Trajectory Data Publishing
    Feng Dengguo
    Zhang Min
    Ye Yutong
    JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY, 2020, 42 (01) : 74 - 88
  • [48] Differentially Private Feature Selection for Data Mining
    Anandan, Balamurugan
    Clifton, Chris
    IWSPA '18: PROCEEDINGS OF THE FOURTH ACM INTERNATIONAL WORKSHOP ON SECURITY AND PRIVACY ANALYTICS, 2018, : 43 - 53
  • [49] Differentially Private Distance Learning in Categorical Data
    Battaglia, Elena
    Celano, Simone
    Pensa, Ruggero G.
    DATA MINING AND KNOWLEDGE DISCOVERY, 2021, 35 (05) : 2050 - 2088
  • [50] Differentially Private Learning with Small Public Data
    Wang, Jun
    Zhou, Zhi-Hua
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 6219 - 6226