TY - JOUR
T1 - Robust Aggregation in Over-the-Air Computation with Federated Learning
T2 - A Semantic Anti-Interference Approach
AU - Ji, Jun Cheng
AU - Lam, Chan Tong
AU - Wang, Ke
AU - Ng, Benjamin K.
N1 - Publisher Copyright:
© 2025 by the authors.
PY - 2026/1
Y1 - 2026/1
N2 - Over-the-air federated learning (AirFL) enables distributed model training across wireless edge devices, preserving data privacy and minimizing bandwidth usage. However, challenges such as channel noise, non-identically distributed data, limited computational resources, and small local datasets lead to distorted model updates, inconsistent global models, increased training latency, and overfitting, all of which reduce accuracy and efficiency. To address these issues, we propose the Semantic Anti-Interference Aggregation (SAIA) framework, which integrates a semantic autoencoder, component-wise median aggregation, validation accuracy weighting, and data augmentation. First, a semantic autoencoder compresses model parameters into low-dimensional vectors, maintaining high signal quality and reducing communication costs. Second, component-wise median aggregation minimizes noise and outlier impact, ideal for AirFL as it avoids mean-based aggregation’s noise sensitivity and complex methods’ high computation. Third, validation accuracy weighting aligns updates from non-identically distributed data to ensure consistent global models. Fourth, data augmentation doubles dataset sizes, mitigating overfitting and reducing variance. Experiments on MNIST demonstrate that SAIA achieves an accuracy of approximately 96% and a loss of 0.16, improving accuracy by 3.3% and reducing loss by 39% compared to conventional federated learning approaches. With reduced computational and communication overhead, SAIA ensures efficient training on resource constrained IoT devices.
AB - Over-the-air federated learning (AirFL) enables distributed model training across wireless edge devices, preserving data privacy and minimizing bandwidth usage. However, challenges such as channel noise, non-identically distributed data, limited computational resources, and small local datasets lead to distorted model updates, inconsistent global models, increased training latency, and overfitting, all of which reduce accuracy and efficiency. To address these issues, we propose the Semantic Anti-Interference Aggregation (SAIA) framework, which integrates a semantic autoencoder, component-wise median aggregation, validation accuracy weighting, and data augmentation. First, a semantic autoencoder compresses model parameters into low-dimensional vectors, maintaining high signal quality and reducing communication costs. Second, component-wise median aggregation minimizes noise and outlier impact, ideal for AirFL as it avoids mean-based aggregation’s noise sensitivity and complex methods’ high computation. Third, validation accuracy weighting aligns updates from non-identically distributed data to ensure consistent global models. Fourth, data augmentation doubles dataset sizes, mitigating overfitting and reducing variance. Experiments on MNIST demonstrate that SAIA achieves an accuracy of approximately 96% and a loss of 0.16, improving accuracy by 3.3% and reducing loss by 39% compared to conventional federated learning approaches. With reduced computational and communication overhead, SAIA ensures efficient training on resource constrained IoT devices.
KW - federated learning
KW - Internet of Things
KW - low power consumption
KW - over-the-air computation
KW - semantic encoding–decoding framework
KW - spiking neuron
UR - https://www.scopus.com/pages/publications/105027995901
U2 - 10.3390/math14010124
DO - 10.3390/math14010124
M3 - Article
AN - SCOPUS:105027995901
SN - 2227-7390
VL - 14
JO - Mathematics
JF - Mathematics
IS - 1
M1 - 124
ER -