Secure Neural Network Inference as a Service with Resource-Constrained Clients

被引:0
|
作者
de Vries, Rik [1 ]
Mann, Zoltan Adam [1 ]
机构
[1] Univ Amsterdam, Amsterdam, Netherlands
关键词
Machine learning as a service; neural network; privacy-preserving machine learning; inference; edge computing; edge intelligence; multi-party computation; homomorphic encryption;
D O I
10.1145/3603166.3632132
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Applying services computing to neural networks, a service provider may provide inference with a pre-trained neural network as a service. Clients use the service to get the neural network's output on their input. To protect sensitive data, secure neural network inference (SNNI) entails that only the client learns the output; the input remains the client's secret and the neural network's parameters remain the service provider's secret. Several SNNI approaches were proposed and evaluated in environments where both service providers and clients used powerful computers. In many real settings, for instance in edge computing, client devices are resource-constrained. This paper is the first to investigate the impact of client-side resource constraints on SNNI. We perform experiments with two state-of-the-art SNNI approaches and three neural networks. We vary the compute and memory capacity of the client device and measure the impact on inference time. Our findings show that client-side resource constraints significantly impact the performance and even the applicability of SNNI approaches. The results indicate the limits of current SNNI approaches for resource-constrained clients. Based on the results, we identify research directions to improve SNNI for resource-constrained clients.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Iterative neural networks for adaptive inference on resource-constrained devices
    Leroux, Sam
    Verbelen, Tim
    Simoens, Pieter
    Dhoedt, Bart
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (13): : 10321 - 10336
  • [2] Iterative neural networks for adaptive inference on resource-constrained devices
    Sam Leroux
    Tim Verbelen
    Pieter Simoens
    Bart Dhoedt
    [J]. Neural Computing and Applications, 2022, 34 : 10321 - 10336
  • [3] Adaptive Sparse Deep Neural Network Inference on Resource-Constrained Cost-Efficient GPUs
    Dun, Ming
    Zhang, Xu
    Cao, Huawei
    Zhang, Yuan
    Huang, Junying
    Ye, Xiaochun
    [J]. 2023 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE, HPEC, 2023,
  • [4] Video Traffic Modeling for Resource-Constrained Clients
    Osborne, Roland
    Villasenor, John
    [J]. GLOBECOM 2006 - 2006 IEEE GLOBAL TELECOMMUNICATIONS CONFERENCE, 2006,
  • [5] Resource-Constrained Classification Using a Cascade of Neural Network Layers
    Leroux, Sam
    Bohez, Steven
    Verbelen, Tim
    Vankeirsbilck, Bert
    Simoens, Pieter
    Dhoedt, Bart
    [J]. 2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [6] A neural network based heuristic for resource-constrained project scheduling
    Shou, YY
    [J]. ADVANCES IN NEURAL NETWORKS - ISNN 2005, PT 1, PROCEEDINGS, 2005, 3496 : 794 - 799
  • [7] A Secure Anonymous Authentication Protocol for Roaming Service in Resource-Constrained Mobility Environments
    Madhusudhan, R.
    Shashidhara, R.
    [J]. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2020, 45 (04) : 2993 - 3014
  • [8] A Secure Anonymous Authentication Protocol for Roaming Service in Resource-Constrained Mobility Environments
    R. Madhusudhan
    R. Shashidhara
    [J]. Arabian Journal for Science and Engineering, 2020, 45 : 2993 - 3014
  • [9] Secure Communications for Resource-Constrained IoT Devices†
    Taha, Abd-Elhamid M.
    Rashwan, Abdulmonem M.
    Hassanein, Hossam S.
    [J]. SENSORS, 2020, 20 (13) : 1 - 18
  • [10] FeatherNet: An Accelerated Convolutional Neural Network Design for Resource-constrained FPGAs
    Morcel, Raghid
    Hajj, Hazem M.
    Saghir, Mazen A. R.
    Akkary, Haitham
    Artail, Hassan
    Khanna, Rahul
    Keshavamurthy, Anil
    [J]. ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2019, 12 (02)