[Paper] Federated Unlearning in Edge Networks: A Survey of Fundamentals, Challenges, Practical Applications and Future Directions

Published: (January 14, 2026 at 08:39 PM EST)
4 min read
Source: arXiv

Source: arXiv - 2601.09978v1

Overview

Federated Unlearning (FUL) is emerging as a crucial capability for edge‑centric AI systems that must honor data‑deletion requests while still benefiting from collaborative learning. This survey paper consolidates the state‑of‑the‑art in FUL, laying out its core concepts, the technical hurdles of making unlearning work over heterogeneous devices, and the avenues where it can be deployed today.

Key Contributions

  • Comprehensive taxonomy of FUL techniques, organized around three implementation pillars: communication efficiency, resource allocation, and security/privacy.
  • Critical analysis of existing frameworks, comparing their assumptions (e.g., trusted server vs. fully decentralized) and the trade‑offs they make on model accuracy, latency, and bandwidth.
  • Mapping of real‑world application domains (IoT, autonomous vehicles, healthcare, smart cities) where FUL can turn regulatory compliance into a competitive advantage.
  • Identification of open research challenges such as verifiable unlearning, incentive‑compatible client participation, and cross‑silo heterogeneity.
  • Road‑map for future work, proposing benchmark datasets, standardized evaluation metrics, and integration with emerging privacy tools (e.g., differential privacy, secure aggregation).

Methodology

The authors performed a systematic literature review covering papers from major ML, security, and networking venues (NeurIPS, ICML, IEEE IoT, ACM CCS, etc.). They:

  1. Defined inclusion criteria – works that explicitly address removal of client contributions from federated models.
  2. Classified papers along the three challenge axes (communication, resources, security) and further by unlearning strategy (exact retraining, approximate influence‑based pruning, cryptographic revocation).
  3. Synthesized findings into a unified framework, highlighting common assumptions, algorithmic primitives, and evaluation practices.
  4. Validated the taxonomy through expert interviews with researchers and industry practitioners working on FL platforms.

Results & Findings

AspectInsight
Communication costMost practical FUL schemes rely on partial model updates or gradient masking to avoid full model retransmission, cutting bandwidth by 30‑70 % compared with naïve retraining.
Resource allocationAdaptive client selection (e.g., “unlearning‑aware” scheduling) reduces on‑device compute overhead, making unlearning feasible on low‑power edge nodes without sacrificing global convergence.
Security & privacyCombining secure aggregation with cryptographic proof of deletion enables auditors to verify that a client’s data influence has been removed, mitigating malicious rollback attacks.
Accuracy impactApproximate unlearning methods typically incur <2 % drop in test accuracy, while exact retraining can cause >10 % degradation if not carefully regularized.
Deployment readinessOnly a handful of open‑source FL toolkits (e.g., TensorFlow Federated, PySyft) currently expose unlearning APIs; most surveyed frameworks are still prototype‑level.

Practical Implications

  • Regulatory compliance made actionable – Companies can embed FUL modules into their FL pipelines to automatically honor GDPR/CCPA “right‑to‑be‑forgotten” requests without re‑training from scratch.
  • Cost savings for edge deployments – By avoiding full model recomputation, service providers can reduce cloud‑edge traffic and extend battery life on IoT devices.
  • Trust & market differentiation – Transparent, verifiable unlearning can become a selling point for privacy‑focused products (e.g., smart home assistants, health‑monitoring wearables).
  • Integration pathways – Existing FL orchestration platforms can adopt the surveyed “unlearning‑aware” client selection and secure aggregation primitives with modest code changes.
  • Tooling roadmap – The paper’s benchmark suggestions (standardized unlearning datasets, performance metrics) give developers a concrete starting point for building and testing FUL capabilities.

Limitations & Future Work

  • Benchmark scarcity – The community lacks widely accepted datasets and reproducible pipelines for measuring unlearning efficiency and model utility.
  • Verification overhead – Cryptographic proofs of deletion add latency; scalable, lightweight verification remains an open problem.
  • Heterogeneity handling – Most surveyed methods assume relatively homogeneous model architectures; extending FUL to mixed‑precision or heterogeneous model families is under‑explored.
  • Incentive mechanisms – Designing economic models that reward clients for participating in unlearning (e.g., compensating extra communication) is still nascent.
  • Long‑term dynamics – The impact of repeated unlearning cycles on model drift and fairness has not been systematically studied.

Bottom line: This survey stitches together the fragmented research on federated unlearning, offering developers a clear map of what works today, where the gaps lie, and how to start building privacy‑compliant, edge‑friendly AI services.

Authors

  • Jer Shyuan Ng
  • Wathsara Daluwatta
  • Shehan Edirimannage
  • Charitha Elvitigala
  • Asitha Kottahachchi Kankanamge Don
  • Ibrahim Khalil
  • Heng Zhang
  • Dusit Niyato

Paper Information

  • arXiv ID: 2601.09978v1
  • Categories: cs.DC
  • Published: January 15, 2026
  • PDF: Download PDF
Back to Blog

Related posts

Read more »