Mathematics (Aug 2024)

EVFL: Towards Efficient Verifiable Federated Learning via Parameter Reuse and Adaptive Sparsification

  • Jianping Wu,
  • Chunming Wu,
  • Chaochao Chen,
  • Jiahe Jin,
  • Chuan Zhou

DOI
https://doi.org/10.3390/math12162479
Journal volume & issue
Vol. 12, no. 16
p. 2479

Abstract

Read online

Federated learning (FL) demonstrates significant potential in Industrial Internet of Things (IIoT) settings, as it allows multiple institutions to jointly construct a shared learning model by exchanging model parameters or gradient updates without the need to transmit raw data. However, FL faces risks related to data poisoning and model poisoning. To address these issues, we propose an efficient verifiable federated learning (EVFL) method, which integrates adaptive gradient sparsification (AdaGS), Boneh–Lynn–Shacham (BLS) signatures, and fully homomorphic encryption (FHE). The combination of BLS signatures and the AdaGS algorithm is used to build a secure aggregation protocol. These protocols verify the integrity of parameters uploaded by industrial agents and the consistency of the server’s aggregation results. Simulation experiments demonstrate that the AdaGS algorithm significantly reduces verification overhead through parameter sparsification and reuse. Our proposed algorithm achieves better verification efficiency compared to existing solutions.

Keywords