Multi-agent Performative Prediction with Greedy Deployment and Consensus Seeking Agents

From MaRDI portal
Publication:6410122

arXiv2209.03811MaRDI QIDQ6410122FDOQ6410122


Authors: Qiang Li, Chung-Yiu Yau, Hoi-To Wai Edit this on Wikidata


Publication date: 8 September 2022

Abstract: We consider a scenario where multiple agents are learning a common decision vector from data which can be influenced by the agents' decisions. This leads to the problem of multi-agent performative prediction (Multi-PfD). In this paper, we formulate Multi-PfD as a decentralized optimization problem that minimizes a sum of loss functions, where each loss function is based on a distribution influenced by the local decision vector. We first prove the necessary and sufficient condition for the Multi-PfD problem to admit a unique multi-agent performative stable (Multi-PS) solution. We show that enforcing consensus leads to a laxer condition for the existence of Multi-PS solution with respect to the distributions' sensitivities, compared to the single agent case. Then, we study a decentralized extension to the greedy deployment scheme [Mendler-D"unner et al., 2020], called the DSGD-GD scheme. We show that DSGD-GD converges to the Multi-PS solution and analyze its non-asymptotic convergence rate. Numerical results validate our analysis.













This page was built for publication: Multi-agent Performative Prediction with Greedy Deployment and Consensus Seeking Agents

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6410122)