论文标题
将数据隐私增强和计算减少的分配学习
Binarizing Split Learning for Data Privacy Enhancement and Computation Reduction
论文作者
论文摘要
Split Learning(SL)通过允许客户在不共享原始数据的情况下协作培训深度学习模型来实现数据隐私。但是,SL仍然存在限制,例如潜在的数据隐私泄漏和客户的高计算。在这项研究中,我们建议将SL局部层进行二进制,以更快地计算(在移动设备上的训练和推理阶段的前向传播时间少17.5倍)和减少的内存使用情况(最多减少了记忆和带宽要求的32倍)。更重要的是,二进制的SL(B-SL)模型可以减少SL被粉碎数据中的隐私泄漏,而模型精度的降级很小。为了进一步增强隐私保护,我们还提出了两种新颖的方法:1)培训额外的局部泄漏损失,2)应用差异隐私,可以单独或同时集成到B-SL模型中。与多种基准模型相比,不同数据集的实验结果肯定了B-SL模型的优势。还说明了B-SL模型针对功能空间劫持攻击(FSHA)的有效性。我们的结果表明,B-SL模型对于具有高隐私保护要求(例如移动医疗保健应用程序)的轻巧的物联网/移动应用程序很有希望。
Split learning (SL) enables data privacy preservation by allowing clients to collaboratively train a deep learning model with the server without sharing raw data. However, SL still has limitations such as potential data privacy leakage and high computation at clients. In this study, we propose to binarize the SL local layers for faster computation (up to 17.5 times less forward-propagation time in both training and inference phases on mobile devices) and reduced memory usage (up to 32 times less memory and bandwidth requirements). More importantly, the binarized SL (B-SL) model can reduce privacy leakage from SL smashed data with merely a small degradation in model accuracy. To further enhance the privacy preservation, we also propose two novel approaches: 1) training with additional local leak loss and 2) applying differential privacy, which could be integrated separately or concurrently into the B-SL model. Experimental results with different datasets have affirmed the advantages of the B-SL models compared with several benchmark models. The effectiveness of B-SL models against feature-space hijacking attack (FSHA) is also illustrated. Our results have demonstrated B-SL models are promising for lightweight IoT/mobile applications with high privacy-preservation requirements such as mobile healthcare applications.