SplitFed


  • SplitFed: When Federated Learning Meets Split Learning

  • Metadata

      • Tags: #/unread
      • Authors: [[Chandra Thapa]], [[Pathum Chamikara Mahawaga Arachchige]], [[Seyit Camtepe]], [[Lichao Sun]]
  • Abstract

Federated learning (FL) and split learning (SL) are two popular distributed machine learning approaches. Both follow a model-to-data scenario; clients train and test machine learning models without sharing raw data. SL provides better model privacy than FL due to the machine learning model architecture split between clients and the server. Moreover, the split model makes SL a better option for resource-constrained environments. However, SL performs slower than FL due to the relay-based training across multiple clients. In this regard, this paper presents a novel approach, named splitfed learning (SFL), that amalgamates the two approaches eliminating their inherent drawbacks, along with a refined architectural configuration incorporating differential privacy and PixelDP to enhance data privacy and model robustness. Our analysis and empirical results demonstrate that (pure) SFL provides similar test accuracy and communication efficiency as SL while significantly decreasing its computation time per global epoch than in SL for multiple clients. Furthermore, as in SL, its communication efficiency over FL improves with the number of clients. Besides, the performance of SFL with privacy and robustness measures is further evaluated under extended experimental settings.

Note

Abstract

Split learning
Prons:

  • Model architecture split between clients and the server.
  • Split learning 更适用于 resource-constrained.
    Cons:
  • Slower, relay-based training across multiple cllients, 接力训练

Proposed Framework

Main server,high-performance computing resources.
Fed server, conducts FedAvg on the clientside local updates

Workflow

Clients side
- Forward propagation, parallel
- Pass smashed data to the main server

Main server

  • Forward propagation and back-propagation
  • Sends gradients of the smashed data to the respective clients
  • Updata its main server model

Clients side

  • Receive gradients, back-propagation, computes gradients
  • DP + gradients , send to the fed server

Fed server

  • Conducts the FedAvg of their local updates
  • Sends aggregation to all participants

Variants of SFL
Based on Server-side Aggregation
variants of SFL: removing the model aggregation part in the server-side computation module

The forward-backward propagations of the server-side model sequentially with respect to the client’s smashed data

Based on data label sharing

  1. Share data labels to the server
  2. Without sharing any data labels to the server
    Clients will process 2 client-side model portions, one with the first few layers, another with the laset few layers. Middle layers will be computed at the server-side.

Privacy Protection at the Client-side

  1. Differential privacy to the client-side model training
  2. PixelDP noise layer in the client-side model.

Privacy Protection at the Fed Server
Clipped the gradients

Add noise

client 对加噪后的梯度进行 update

Privacy Protection on Main Server.
Smashed data to the main server 可能泄露信息
Integrate a noise layer in the client-side model based on the concepts of PixelDP

Conclusion

Faster than SL by parallel processing across clients.
SFL has similar performance compared to SL
Hybrid approach that supports machine learning with resource-constrained devices.


文章作者: lunan
版权声明: 本博客所有文章除特別声明外,均采用 CC BY 4.0 许可协议。转载请注明来源 lunan !
  目录