Abstract
In this paper, we propose a non-disclosure incentive mechanism for federated learning via alternating-offers bargaining which does not utilize device information of clients while improving learning performance and approximating an optimal resource allocation. By defining utility functions of clients that reflect their self estimated performances and determining minimum compensation information, clients can be selected for participation without revealing their device information such as loss of their local model or dataset size. To determine the compensation resource allocation, server and clients perform alternating-offers bargaining, which does not require sharing their utility functions. Experiment results show that the proposed incentive mechanism can speed up convergence, improve test accuracy, and induce compensation resource allocation near the Nash bargaining soution while not revealing the device information of clients.
| Original language | English |
|---|---|
| Pages (from-to) | 1343-1352 |
| Number of pages | 10 |
| Journal | Journal of Korean Institute of Communications and Information Sciences |
| Volume | 50 |
| Issue number | 9 |
| DOIs | |
| State | Published - Sep 2025 |
Bibliographical note
Publisher Copyright:© 2025, Korean Institute of Communications and Information Sciences. All rights reserved.
Keywords
- Alternating-offers bargaining
- Federated learning
- Incentive mechanism