Federated multi-objective reinforcement learning

Research output: Contribution to journalArticlepeer-review

7 Scopus citations

Abstract

Multi-objective reinforcement Learning (MORL) has significant potential for solving complex decision problems with conflicting objectives. Desiring sufficient training samples, it is promising to achieve federated MORL in large-scale distributed settings. However, itstill suffers from poor efficiency and high privacy risks. To mitigate the inefficiency issue, we first propose a novel probablistic algorithm PMORL that can seek an optimal policy via the expectation maximization (EM) algorithm with high efficiency. To extend PMORL to distributed settings with privacy protection, we then present the first federated MORL algorithm Fed-PMORL with client-level differential privacy (DP). In Fed-PMORL, personalized actors are trained and maintained at local clients whereas critics are aggregated and sanitized at the central server. Extensive experimental results in benchmark MORL environments demonstrate that Fed-PMORL under DP guarantees can achieve superior performance with high efficiency. In particular, compared with the state-of-the-art methods, PMORL and Fed-PMORL can save up to 50% training episodes for achieving the same model utility. With a sufficient number of clients (e.g., 1000 clients), Fed-PMORL with a formal DP guarantee shows utility comparable to that of the non-private algorithm.

Original languageEnglish
Pages (from-to)811-832
Number of pages22
JournalInformation Sciences
Volume624
DOIs
StatePublished - May 2023

Keywords

  • Differential privacy
  • Federated learning
  • Graph model
  • Multi-objective optimization
  • Reinforcement learning

Fingerprint

Dive into the research topics of 'Federated multi-objective reinforcement learning'. Together they form a unique fingerprint.

Cite this