Jump to content
Main menu
Main menu
move to sidebar
hide
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Edge Computing Wiki
Search
Search
Appearance
Create account
Log in
Personal tools
Create account
Log in
Pages for logged out editors
learn more
Contributions
Talk
Editing
Federated Learning
(section)
Page
Discussion
British English
Read
Edit
View history
Tools
Tools
move to sidebar
hide
Actions
Read
Edit
View history
General
What links here
Related changes
Upload file
Special pages
Page information
Appearance
move to sidebar
hide
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
= 5.8 Open Challenges = Federated Learning (FL) helps protect user data by training models directly on devices. While useful, FL faces many challenges that make it hard to use in real-world systems.<sup>[1]</sup><sup>[2]</sup><sup>[3]</sup><sup>[4]</sup><sup>[5]</sup><sup>[6]</sup> * '''System heterogeneity''': Devices used in FL—like smartphones, sensors, or edge servers—have different speeds, memory, and battery life. Some may be too slow or lose power during training, causing delays. Solutions include using smaller models, pruning, partial updates, and early stopping.<sup>[1]</sup><sup>[6]</sup> * '''Communication bottlenecks''': FL requires frequent communication between devices and a central server. Sending full model updates can overwhelm slow or unstable networks. To address this, update compression methods like quantization, sparsification, and knowledge distillation are used.<sup>[2]</sup><sup>[3]</sup> * '''Statistical heterogeneity (Non-IID data)''': Each device collects different data based on user behavior and environment, leading to non-IID distributions. This can reduce global model accuracy. Solutions include clustered FL, personalized models, and meta-learning approaches.<sup>[2]</sup><sup>[5]</sup> * '''Privacy and security''': Even if raw data stays on devices, model updates can leak sensitive information. Malicious clients may poison the model. Mitigation strategies include secure aggregation, differential privacy, and homomorphic encryption, though these may add computational and communication overhead.<sup>[2]</sup><sup>[3]</sup> * '''Scalability''': FL must support thousands or millions of clients—many of which may drop out or go offline. Techniques like reliable client selection and hierarchical FL (with edge aggregators) improve scalability.<sup>[4]</sup> * '''Incentive mechanisms''': Clients may not participate without benefits, especially if training drains battery or bandwidth. Solutions like blockchain-based tokens, credit systems, or reputation scores are being explored, but adoption remains low.<sup>[2]</sup> * '''Lack of standardization and benchmarks''': FL lacks widely accepted benchmarks and datasets. Simulations often ignore real-world issues like device failure, non-stationarity, or network variability. Frameworks like LEAF, FedML, and Flower offer partial solutions, but more real-world testing is needed.<sup>[2]</sup><sup>[3]</sup> * '''Concept drift and continual learning''': User data changes over time, which can make models outdated. Continual learning techniques like memory replay, adaptive learning rates, and early stopping help models stay relevant.<sup>[1]</sup><sup>[6]</sup> * '''Deployment complexity''': Devices run different operating systems, have different hardware specs, and operate on varying network conditions. Managing model updates and ensuring consistent performance is difficult across such diverse platforms. * '''Reliability and fault tolerance''': Devices may crash, lose connection, or send corrupted updates. FL systems must be resilient and able to recover from partial failures without degrading the global model. * '''Monitoring and debugging''': In FL, training occurs across thousands of devices, making it hard to monitor model behavior or identify bugs. New tools are needed to improve observability and debugging in distributed systems.
Summary:
Please note that all contributions to Edge Computing Wiki may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see
Edge Computing Wiki:Copyrights
for details).
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)