posted on 2025-08-01, 16:44authored byJ Wang, J Hu, J Mills, G Min, M Xia, N Georgalas
Federated learning (FL) is a privacy-preserving distributed machine learning paradigm that enables collaborative training
among geographically distributed and heterogeneous devices without gathering their data. Extending FL beyond the supervised learning
models, federated reinforcement learning (FRL) was proposed to handle sequential decision-making problems in edge computing
systems. However, the existing FRL algorithms directly combine model-free RL with FL, thus often leading to high sample complexity
and lacking theoretical guarantees. To address the challenges, we propose a novel FRL algorithm that effectively incorporates modelbased RL and ensemble knowledge distillation into FL for the first time. Specifically, we utilise FL and knowledge distillation to create
an ensemble of dynamics models for clients, and then train the policy by solely using the ensemble model without interacting with
the environment. Furthermore, we theoretically prove that the monotonic improvement of the proposed algorithm is guaranteed. The
extensive experimental results demonstrate that our algorithm obtains much higher sample efficiency compared to classic model-free
FRL algorithms in the challenging continuous control benchmark environments under edge computing settings. The results also highlight
the significant impact of heterogeneous client data and local model update steps on the performance of FRL, validating the insights
obtained from our theoretical analysis.
Funding
101008297
EP/X019160/1
EP/X038866/1
Engineering and Physical Sciences Research Council (EPSRC)