Train Without Strain: Adaptive Pruning and Hypernetwork Personalization for Federated Transformers

Moqbel Hamood, Abdullatif Albaseer, Mohamed Abdallah, Ala Al-Fuqaha, Bechir Hamdaoui

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Deploying transformer models in Personalized Federated Learning (PFL) over wireless networks is challenging due to their large size, which leads to high communication overhead, increased latency, and excessive energy consumption. Traditional pruning and sparsification methods, designed mainly for conventional deep learning architectures, are ineffective for transformers and can cause divergence or degrade performance - especially when applied to self-attention layers or through direct federated averaging. To address these challenges, we propose a novel dual approach called PFL-TPS (PFL with Transformer Pruning and Sparsification). Our approach efficiently reduces communication and computation costs while maintaining model performance, making it suitable for resource-constrained wireless networks. Specifically, we apply adaptive pruning with trainable thresholds to the transformer's Feed-Forward Layers (FFLs), and only these trainable thresholds are shared with the server, resulting in minimal uploaded data. For the Self-Attention Layers (SALs), instead of transmitting bandwidth-intensive model parameters, we employ a server-side hypernetwork that generates personalized parameters based on device-specific embedding vectors sent by the devices, significantly reducing communication overhead and maintaining personalization. Extensive experiments show that PFL-TPS reduces energy consumption by up to 50%, decreases training time by 60.44%, and improves model accuracy by 49.87% compared to baselines in wireless networks.

Original languageEnglish
Title of host publicationICC 2025 - IEEE International Conference on Communications
EditorsMatthew Valenti, David Reed, Melissa Torres
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages3575-3580
Number of pages6
ISBN (Electronic)9798331505219
DOIs
Publication statusPublished - 2025
Event2025 IEEE International Conference on Communications, ICC 2025 - Montreal, Canada
Duration: 8 Jun 202512 Jun 2025

Publication series

NameIEEE International Conference on Communications
ISSN (Print)1550-3607

Conference

Conference2025 IEEE International Conference on Communications, ICC 2025
Country/TerritoryCanada
CityMontreal
Period8/06/2512/06/25

Keywords

  • learnable Thresholds
  • Personalized Federated Learning (PFL)
  • Pruning
  • Resource Optimization
  • Sparse Models
  • Transformers

Fingerprint

Dive into the research topics of 'Train Without Strain: Adaptive Pruning and Hypernetwork Personalization for Federated Transformers'. Together they form a unique fingerprint.

Cite this