What is it about?

Federated Learning (FL) is a way for many devices (like phones or computers) to work together to train a shared machine learning model without sharing their private data. Instead of sending data to one place, each device trains a small piece of the model on its own data and only shares updates to the model. Normally, FL uses a setup where all devices talk to a central server. But in real life, that’s not always ideal because: 1. Devices have different speeds and capabilities (some are fast, some are slow). 2. Their data is often very different from each other. 3. Some devices may lag or drop out. 4. The central server can become a weak spot—if it fails, everything stops. To fix these problems, we created a new approach: a peer-to-peer system where devices talk directly to each other instead of relying on one server. This system works better when devices are different and not all are available at the same time. We also added a smart method to decide how much each device’s update should count, based on how much progress it has made. This helps balance things when some devices are slower or have less power. Finally, when we tested it using popular image datasets (CIFAR-10 and CIFAR-100), this new method got up to 37.7% better accuracy than older methods—even when limiting how often devices send updates.

Featured Image

Why is it important?

This work is important because it makes Federated Learning more practical and effective in real-world situations. In many cases, we want to train machine learning models using data from personal devices (like smartphones or wearables), but we also want to protect user privacy by keeping the data on the device. FL is designed for that—but traditional methods rely too much on a central server and assume that all devices are equally capable and always available, which isn't true in real life. The new peer-to-peer approach solves these problems by removing the central server and letting devices communicate directly. This makes the system more robust, especially when devices are very different or drop in and out. Also, the way it adjusts how much each device’s update counts—based on how much progress it has made—helps make the training fairer and more accurate, even when some devices are slower or have less computing power. Ultimately, this makes FL more useful for real applications like healthcare, finance, or mobile services, where data privacy is crucial and devices are very diverse. It also means faster, more reliable, and more accurate models, which benefits everyone.

Read the Original

This page is a summary of: Towards Asynchronous Peer-to-Peer Federated Learning for Heterogeneous Systems, March 2025, ACM (Association for Computing Machinery),
DOI: 10.1145/3721146.3721952.
You can read the full text:

Read

Contributors

The following have contributed to this page