Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Commun. Netw., 09 February 2026

Sec. Wireless Communications

Volume 7 - 2026 | https://doi.org/10.3389/frcmn.2026.1748815

Data- and distance-aware clustering for scalable wireless federated learning



Zhenning Chen

Zhenning Chen 1*Zihe XuZihe Xu2Yihan DingYihan Ding3Youren Wang
Youren Wang1
  • 1 College of Automation Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, China
  • 2 Jiangsu Key Laboratory of Wireless Communications and IoT, Nanjing University of Posts and Telecommunications, Nanjing, China
  • 3 School of Information Science and Technology, Dalian Maritime University, Dalian, China

Introduction: Federated learning (FL) enables model training on edge devices using local data while aggregating model updates at a central server without exchanging raw data, thereby preserving privacy. However, achieving satisfactory convergence accuracy with low communication energy remains challenging. This work investigates a three-tier clustered FL (CFL) architecture to improve global training performance and communication efficiency through joint device clustering and resource scheduling.

Methods: We analyze how clustering strategies influence learning convergence and communication energy consumption. Based on this analysis, we propose a clustering method that jointly accounts for gradient cosine similarity and communication distance. A simplified procedure is further developed for device association and cluster-head selection, with the goals of improving intra-cluster data balance and reducing the overall communication distance to the server.

Results: Simulations demonstrate that the proposed method consistently improves model accuracy while reducing communication energy consumption compared with random clustering and similarity-based clustering baselines.

Discussion: These results indicate that jointly considering update similarity and communication distance in CFL can effectively balance learning quality and communication cost, offering a practical approach for energy-efficient federated training in edge networks.

1 Introduction

The explosive growth of mobile edge devices and their local storage of data has challenged traditional centralized machine learning paradigms. Centralized approaches necessitate frequent data transmission to the central server, which results in significant privacy risks and high communication overhead (Xia et al., 2020). Federated learning (FL) has emerged as a promising distributed learning paradigm, which enables model training by aggregating locally trained models at the central server without directly transmitting raw data, thereby preserving data privacy (Konečnỳ et al., 2015). However, FL still relies on frequent communication between devices and the server, which becomes inefficient in wireless environments with heterogeneous data distributions (Lu et al., 2024).

During the evolution of FL, extensive efforts have been devoted to mitigating communication overhead from various perspectives. Hierarchical FL reduces communication frequency through staged aggregation, thereby lowering communication costs (Briggs et al., 2020; Tran et al., 2025); however, its rigid hierarchical design often limits adaptability in dynamic environments. Prototype-based clustered FL is a representative branch of personalized FL, which aligns client features through global or local prototypes, thereby effectively reducing communication costs by minimizing the need for frequent model exchanges (Yang et al., 2024; Tan et al., 2022). However, such methods often compromise global consistency, thereby limiting scalability when tasks are only partially shared. Moreover, resource-aware FL introduces client selection under bandwidth and computation constraints to improve communication efficiency but typically decouples resource optimization from model aggregation, neglecting their intrinsic interplay (Nishio and Yonetani, 2019). Against this backdrop, clustered FL (CFL) has become a more direct and effective solution. CFL reduces communication costs and improves convergence stability by performing intra- and inter-cluster aggregation after grouping clients based on their statistical or geographical characteristics (Ghosh et al., 2021; Zeng et al., 2023). For example, Yan et al. (2024) proposed adaptive clustering strategies that enable flexible client participation and dynamic cluster formation, thereby reducing redundant communication under non-independent and identically distributed (Non-IID) conditions. Similarly, Gao et al. (2023) clustered clients based on the similarity of their local data distributions and used acceleration algorithms to shorten training time and lower communication overhead. In addition, Zhang et al. (2024) developed an adaptive CFL framework that adjusts cluster size and communication intervals through online similarity measurement, thereby improving both robustness and communication efficiency. Despite these advances, existing CFL approaches still suffer from several limitations: 1) existing studies rarely explore how the clustering strategy influences both data richness and convergence dynamics. Over-reliance on data similarity for clustering may reduce intra-cluster diversity, thereby weakening model generalization. 2) Prior CFL methods primarily rely on gradient similarity or geographic proximity for clustering but often ignore joint optimization of learning performance and resource efficiency.

To bridge this research gap, we propose a data- and distance-aware clustering scheme. The proposed scheme exploits data distribution characteristics and geographical information to optimize cluster head selection and device association scheduling prior to the training process. Based on this clustering result, the CFL training procedure is subsequently carried out. The main contributions of this study are summarized as follows.

• We propose a CFL framework that collectively considers learning performance and communication cost. On the learning side, a convergence analysis is conducted to theoretically demonstrate that enhancing the diversity and representativeness of intra-cluster data effectively improves the convergence behavior of CFL under data heterogeneity. On the communication side, the communication cost is modeled in terms of the transmission distance. Based on this model, a combined optimization problem for cluster head selection and device association scheduling is formulated, which simultaneously accounts for learning performance and communication cost.

• Based on the formulated joint optimization objective, we develop an iterative algorithm to efficiently solve the cluster head selection and device association scheduling problem. The proposed algorithm decomposes the original NP-hard problem into two tractable subproblems, which are solved in an alternating optimization manner.

• Simulation results demonstrate that the proposed method consistently outperforms the three baseline algorithms in terms of model accuracy and communication energy efficiency, thereby validating the effectiveness of the proposed framework.

2 System model

We consider a wireless CFL system, as illustrated in Figure 1, which consists of a central server and a set of devices U = { 1,2 , , U } . Each device u U has a local dataset D u = ( x u i , y u i ) i = 1 D u , where D u = | D u | . The communication distances are denoted by d u , u U × U for the links between devices and d u , cs U × 1 for the links between devices and the server. Before training, a subset of high-performance devices N = { 1,2 , , N } can be selected from the set of devices as candidates for cluster heads, and all devices are grouped through two scheduling strategies.

Figure 1
Network diagram illustrating a hierarchical federated learning framework where multiple device clusters with designated cluster heads communicate local and aggregated models to a central server, following indicated pathways for data upload and model broadcasting as defined in the legend.

Figure 1. System model.

A binary variable a n { 0,1 } indicates whether the device n is selected as a cluster head. The set of cluster heads is N c h , satisfying a n = 1 for all n N c h . The number of clusters is N c h . We also introduce a binary variable b u , n { 0,1 } that indicates whether the device u is associated with the cluster head n . Each cluster is defined as C n = { u U b u , n = 1 } , and the overall cluster set is C = { C 1 , C 2 , , C N } , where C n C n = for n n and | C n | 1 . The total number of samples in cluster n is D n = u U b u , n D u .

The global loss function is defined in Equation 1:

F w = n N a n D n D u U b u , n D u D n F u w . ( 1 )

Here, F u ( w ) = 1 D u i = 1 D u ( x u i , y u i , w ) is the local loss at device u , and D = u U D u is the total number of training samples. The intra-cluster aggregation loss at the cluster head n is defined as F n ( w ) = u U b u , n D u D n F u ( w ) .

2.1 CFL process

The architecture of CFL consists of three layers: intra-cluster devices, cluster heads, and a central server. A synchronous aggregation scheme is used. The overall process comprises the following steps, where Step 1 is executed during the initialization phase and the remaining steps are iteratively performed throughout the training process.

1. C l u s t e r i n g : Before the formal training process begins, each device performs pre-training using the global model w broadcasted by the central server, computes local gradients with its local dataset, and uploads F u ( w ) to the central server. The central server performs data similarity analysis and communication distance calculation based on the gradients and location information uploaded by all devices. These metrics are used to select cluster heads and establish device associations.

2. Local training : The central server broadcasts the global model w to all devices. The devices receive the model and unfold the local training. In training round t , each device u updates its own model with its local data using the SGD algorithm, as shown in Equation 2:

w u t = w u t 1 γ g u w u t 1 , ( 2 )

where w u t represents the local model parameters obtained during the t th round of local iteration, g u ( w u t ) = g ( ζ u t ; w u t ) denotes the stochastic gradient computed on a randomly sampled mini-batch ζ u t of local training data D u , and γ represents the learning rate.

1. Intra - cluster model aggregations: Devices perform intra-cluster model aggregation during local iteration t { τ 1 , 2 τ 1 , 3 τ 1 , , T } , where T denotes the predefined total number of global training rounds. All devices within the cluster upload their latest models to their respective cluster heads. Each cluster head then aggregates the local models within the cluster according to Equation 3:

w n t = u U b u , n D u D n w u t . ( 3 )

After intra-cluster model aggregation is completed, each cluster head broadcasts the aggregated model to its associated devices for continued local training.

1. Inter - cluster model aggregations: Devices conduct inter-cluster model aggregation during local iteration t { τ 2 , 2 τ 2 , 3 τ 2 , , T } , where τ 2 is an integer multiple of τ 1 . Equation 4 is provided as follows:

w t = n N a n D n D w n t . ( 4 )

Subsequently, the central server broadcasts the latest global model to all devices, which is then used for the next round of local training.

2.2 Problem formula

The pairwise cosine similarity between the gradients of devices is computed, as shown in Equation 5:

c u , u w = F u w , F u w F u w F u w . ( 5 )

Assuming that the spectrum is divided into U orthogonal sub-channels and all devices share similar channel conditions (e.g., equal transmission power and bandwidth), the main factor influencing communication energy consumption becomes the distance between devices and their associated cluster heads, along with distance between cluster heads and the central server. Therefore, minimizing the global communication energy consumption can be transformed into minimizing the total communication distance. Let d n , cs denote the distance between the cluster head n and the central server. The total communication distance across all clusters is defined as shown in Equation 6:

d C = n N a n u U b u , n d u , n + d n , cs . ( 6 )

The objective is to determine the optimal clustering strategy that minimizes the weighted sum of the final global training loss and the overall communication cost. The optimization problem is formulated as shown in Equation 7:

min A , B α F w T + β d C s.t. C 1 :  n N b u , n = 1 , u U , C 2 :  u U b u , n V n , n N , C 3 :  b u , n a n , u U , n N , C 4 :  a n 0,1 , n N , C 5 :  b u , n 0,1 , u U , n N . ( 7 )

Here, A = { a n n N } , and B = { b u , n u U , n N } denote the cluster head selection and device association strategies, respectively. The weights α and β are used to balance model performance and communication cost, and α + β = 1 . Constraints C1–C5 ensure valid clustering assignments and binary decisions, where V n imposes a capacity limit on the number of devices that a cluster head can serve due to constrained communication resources. Since F ( w T ) depends implicitly on the decision variables, direct optimization is intractable. We, therefore, analyze the convergence behavior to derive a tractable formulation.

3 Convergence analysis

To evaluate how clustering strategies influence learning performance, we conduct a convergence analysis to understand how they influence the convergence performance of CFL. To obtain the expected convergence rate of CFL, we first make the following assumptions (Wang et al., 2020; Wan et al., 2021).

• Assumption 1: Assume that the global loss function is differentiable, its gradient is uniformly Lipschitz continuous, and there exists a positive constant L that satisfies Equation 8:

F w F w L w w . ( 8 )

The equation is equivalent to Equation 9:

F w F w + w w T F w + L 2 w w 2 2 . ( 9 )

Here, 2 denotes the calculation of the Euclidean norm.

• Assumption 2: Global divergence is bounded, as shown in Equation 10:

n N a n D n D F n w F w 2 ϵ 2 , n , w . ( 10 )

• Assumption 3: Local divergence is bounded, as presented in Equation 11:

u U b u , n D u D n F u w F n w 2 ϵ n 2 . n , w . ( 11 )

• Assumption 4: Local variance is bounded, as shown in Equation 12:

E g u w F u w 2 σ 2 , u , w . ( 12 )

Based on the aforementioned assumptions and the description of the global model, we present the following convergence results. Given the optimal global model w * , the learning rate is set to satisfy γ 1 2 6 G L . Then, it should satisfy 0 < 4 γ 2 L 2 τ 2 2 1 12 γ 2 L 2 τ 2 2 1 3 and 1 < 1 1 12 γ 2 L 2 τ 1 2 < 1 1 12 γ 2 L 2 τ 2 2 2 . The optimality gap between the expected global loss function value and the optimal global loss function value is presented in Equation 13:

1 T t = 1 T E f w t 2 2 f w 1 f w * γ T + 36 γ 2 L 2 τ 2 2 ϵ 2 + 3 n N a n D n D 2 u U b u , n D u D n 2 σ 2 + 48 γ 2 L 2 τ 1 2 n N a n D n D ϵ n 2 + 24 γ 2 L 2 τ 2 n N a n D n D 1 D n D u U b u , n D u D n 2 σ 2 + 16 γ 2 L 2 τ 1 n N a n D n D u U b u , n D u D n 1 b u , n D u D n 2 σ 2 . ( 13 )

From Equation 13, we can analyze how each key parameter would affect the convergence of the proposed CFL algorithm. The learning rate γ determines the gradient step size, while the smoothness constant L characterizes the curvature of the loss function; increasing either parameter would enlarge the convergence upper bound, whereas decreasing them would tighten it, thereby improving convergence. Participation weights a n and b u , n , along with data sizes D n and D u , determine each device’s contribution to the aggregated gradient; uneven participation or disproportionate data sizes would further increase the bound. The total number of communication rounds T reduces the contribution of the initial optimality gap, so more rounds would tighten the bound. The parameter τ 1 specifies how many times each device updates its model before a single intra-cluster aggregation; a larger τ 1 would amplify the effects of ϵ n 2 and σ 2 , thus increasing the bound. Similarly, the parameter τ 2 specifies how many times each device updates its model before a single inter-cluster aggregation; a larger τ 2 value would amplify the effects of ϵ 2 and σ 2 , thus also increasing the bound.

Notably, by combining the convergence bound in Equation 13 with assumptions 2 and 3, it can be observed that client drift—defined as the deviation of local gradient updates from the global gradient direction—increases the convergence upper bound, thereby degrading learning performance. In particular, assumptions 2 and 3 characterize such drift through the deviation measures ϵ and ϵ n , which quantify the discrepancy between cluster-level gradients and the global gradient, along with between-individual client gradients and their corresponding cluster-level gradients. Since the derived convergence upper bound increases monotonically with both ϵ and ϵ n , stronger client drift results in a looser convergence bound and, consequently, slower convergence. We further observe that reducing ϵ 2 , which measures the discrepancy between the aggregated data distribution within each cluster and the global data distribution, yields a more significant improvement in learning performance than reducing ϵ n 2 . This is because the period of cluster-level updates, τ 2 , is typically larger than the period of device-level updates, τ 1 , causing the effect of intra-cluster heterogeneity to accumulate more strongly over τ 2 steps. This indicates that enhancing intra-cluster data representativeness plays a dominant role in mitigating client drift and motivates our focus on cluster-level optimization. Consequently, we focus on minimizing the following objective min n N a n D n D F n ( w ) F ( w ) 2 . For analytical convenience, we assume uniform data sizes across devices and simple averaging during global aggregation. Under these assumptions, the above divergence objective can be reformulated using norm properties and further bounded via the Cauchy–Schwarz inequality. This results in an upper bound that relates the discrepancy between local and global gradients to the pairwise similarity among local gradients. As a result, minimizing the sum of intra-cluster gradient cosine similarity serves as a tractable surrogate for reducing client drift induced by data heterogeneity. Accordingly, based on the definition of cosine similarity in Equation 2, the objective is transformed into the maximization problem shown in Equation 14:

min 1 N c h n N c h 1 C n 2 u , u C n c u , u w . ( 14 )

4 Algorithm design

Based on the above convergence analysis, the cluster-level data representativeness metric is defined as follows (Equation 15):

c C = 1 N c h n N c h 1 C n 2 u , u C n c u , u w . ( 15 )

By substituting the original global loss function with c ( C ) , the optimization problem in Equation 4 can be reformulated as presented in Equation 16:

min A , B β d C + α c C s.t. C 1 –C 5 , ( 16 )

where Equation 14 is NP-hard and is addressed by decomposing it into device association and cluster-head selection strategies based on cosine similarity and communication distance.

Given an initial cluster head set N c h N , the device association problem is formulated as shown in Equation 17:

min B β n N c h u U b u , n d u , n + α 1 N c h n N c h 1 C n 2 u , u C n c u , u w s.t. C 1 , C 2 , C 5 , ( 17 )

which is an integer nonlinear programming problem. We adopt the Gurobi solver for optimal device association under a fixed cluster head set. The overall clustering utility is defined in Equation 18:

Ψ N = ψ N + n N a n d n , cs , ( 18 )

where ψ ( N ) is the device association utility obtained by solving Equation 17 with the given cluster head set. The cluster head selection problem can be formulated as shown in Equation 19:

min A Ψ N s.t. C 4 . ( 19 )

To minimize Ψ ( N c h ) , a greedy iterative strategy is used to adjust the cluster head set, which includes the following three main update operations:

Cluster head addition: For any device n N \ N c h , compute the utility Ψ ( N c h + { n } ) ; if it improves the objective, add n to N c h .

Cluster head exchange: For n N \ N c h and n N c h , evaluate Ψ ( N c h + { n } { n } ) ; replace n with n if utility is improved.

Cluster head removal. For n N c h , compute Ψ ( N c h { n } ) ; if this reduces the objective, remove n from the set.

Therefore, in the greedy iterative strategy for cluster head selection, the number of cluster heads varies dynamically until convergence. We summarize the alternating optimization process of device association and cluster head selection in Algorithm 1. The overall complexity is mainly determined by two components: (i) solving the device association problem using the Gurobi solver for a fixed cluster head set and (ii) the greedy iterative updates for cluster head selection, including addition, exchange, and removal operations. In particular, let N = | N | be the total number of candidate nodes and U = | U | be the number of devices. Computing the device association utility ψ ( N c h ) via Gurobi involves all devices in U , with complexity O ( U ) per evaluation. During each iteration of the greedy cluster head updates, the addition step requires up to N | N c h | evaluations of Ψ ( N c h ) , the exchange step requires up to ( N | N c h | ) | N c h | evaluations, and the removal step requires up to | N c h | evaluations. Hence, the per-iteration complexity can be expressed as O ( N | N c h | ) + ( N | N c h | ) | N c h | + | N c h | U . Since | N c h | N , this can be simplified and upper-bounded by O ( N 2 U ) per iteration. Therefore, for a total of T iterations until convergence, the overall computational complexity of the alternating optimization algorithm is O ( T N 2 U ) , which clearly highlights its scalability with respect to the number of candidate nodes and the number of devices.

Algorithm 1
www.frontiersin.org

Algorithm 1. Clustering strategy based on device association and cluster head selection.

5 Numerical results

We simulate a wireless FL system consisting of a central server and 100 devices uniformly distributed within a circular area of 100 m radius. To model long-range communication, the server is positioned 1 km away from the device cluster center. The total number of global training rounds is fixed at 200, and the intra- and inter-cluster model aggregation periods are set to τ 1 = 1 and τ 2 = 2 , respectively. Each device is allocated to a communication bandwidth of 1 Mbps and transmits at a power of 10 mW. The noise power spectral density is set to 173 dBm/Hz, and the path loss model follows 128.1 + 37.6 log 10 ( d km ) . The model size is 28 kB, and each local training batch consists of 32 samples, with a learning rate of 0.01. Communication energy consumption is calculated using the Shannon formula, considering both device-to-cluster-head and cluster-head-to-server distances (Chen et al., 2020; Taïk et al., 2022). For simplicity, we assume identical transmission power and bandwidth across all devices so that both communication energy and delay depend solely on transmission distance and exhibit the same trend. In addition, we consider the single-hop communication model between devices and their associated cluster heads, which is commonly adopted in cluster-based wireless FL architectures. Under this assumption, routing overhead is negligible and does not affect the relative performance comparison. Experiments are conducted on the Fashion-MNIST and CIFAR-10 datasets. To emulate a statistically heterogeneous non-IID environment, data are partitioned across devices using a Dirichlet distribution, with a concentration parameter set to 0.5, which corresponds to a moderate-to-high level of data heterogeneity and has been widely adopted in prior federated learning studies to evaluate robustness under non-IID settings (Meng et al., 2023). We set the weights to α = β = 0.5 .

To evaluate the proposed method, three baseline algorithms are considered for comparison: (i) random clustering, where devices are grouped based on the geographical proximity and cluster heads are selected randomly within each cluster; (ii) similarity-based clustering, which groups devices with similar local data distributions using statistical distance metrics, with heads randomly assigned; and (iii) FedAvg, the conventional FL scheme without clustering, where all devices communicate directly with the central server in each round.

To verify the effectiveness of the proposed clustering algorithm in enhancing learning performance in FL, we conduct test accuracy comparison experiments on the Fashion-MNIST and CIFAR-10 datasets. The experiments evaluate the impact of different device clustering algorithms, along with the classical FedAvg, on the model’s training accuracy. Figures 2, 3 illustrate the evolution of test accuracy during training, and the corresponding test accuracies at convergence are summarized in Table 1. As shown, the proposed algorithms consistently achieve the highest test accuracies for a given number of training rounds and maintain significant advantages throughout the training process. The similarity-based clustering algorithm ranks second, suggesting that adjusting data within clusters to a more balanced distribution—i.e., aligning the data distribution across clusters with the global distribution—can lead to better convergence performance than clustering purely based on intra-cluster data similarity, which is in line with expectations. The randomized clustering algorithm ranks third because it does not account for data distribution or similarity within clusters, leading to less balanced clusters and consequently slower convergence. The FedAvg algorithm exhibits the worst performance, primarily due to the heterogeneity of local data distributions across devices under α = 0.5 , which leads to inconsistent update directions in local models and partially counteracts gradients during global model aggregation, thereby reducing convergence efficiency. Overall, these results further highlight the effectiveness of the proposed clustering algorithm in improving the training of devices in heterogeneous environments.

Figure 2
Line chart comparing four algorithms—Proposed, Similarity, Random, and FedAvg—showing accuracy percentage over 200 rounds. Proposed achieves the highest accuracy, followed by Similarity, Random, and FedAvg. Legend identifies each colored line.

Figure 2. Test accuracy on Fashion-MNIST vs. local training rounds.

Figure 3
Line graph comparing model accuracy versus training rounds for four methods: Proposed, Similarity, Random, and FedAvg. Proposed achieves the highest final accuracy, followed by Similarity, Random, and FedAvg. Each curve consistently improves with more rounds.

Figure 3. Test accuracy on CIFAR-10 vs. local training rounds.

Table 1
www.frontiersin.org

Table 1. Test accuracy of different algorithms on Fashion-MNIST and CIFAR-10 datasets.

To verify the effectiveness of the proposed algorithm in reducing global communication energy consumption, energy simulations are conducted on the Fashion-MNIST dataset under different numbers of devices, as shown in Figure 4. FedAvg incurs the highest energy consumption due to frequent long-distance communication with the server. In contrast, the other three algorithms use intra-cluster aggregation, which shortens communication distances and reduces upload frequency, thereby lowering energy usage. The proposed method performs best by jointly optimizing inter-device distances and balancing intra-cluster data, further reducing global communication energy.

Figure 4
Bar chart comparing communication energy consumption (joules) across four methods—Proposed, Similarity, Random, and FedAvg—at device numbers 60, 80, 100, and 120. FedAvg consistently shows the highest energy consumption, while Proposed method has the lowest across all device numbers.

Figure 4. Global energy consumption vs. number of devices.

As shown in Figures 2, 4, the proposed algorithm consistently achieves high training accuracy with low communication energy consumption, highlighting its advantage in maintaining model performance while reducing communication cost. In comparison, similarity-based and random clustering exhibit slower convergence and higher energy consumption. The results demonstrate that balancing intra-cluster data enhances cluster representativeness, mitigates aggregation conflicts, and, when combined with geographical proximity, contributes to reducing overall energy consumption.

Table 2 presents the learning performance and communication energy under different values of the weighting factor α . As α increases, giving more importance to model accuracy in the optimization problem, the learning performance improves, reaching up to 85.1% for α = 1 . However, this comes at the cost of higher communication energy, which increases from 101 J to 168 J over the same range. These results clearly illustrate the trade-off controlled by the weighting factor: larger α prioritizes learning accuracy at the expense of communication efficiency, while smaller α reduces energy consumption but yields lower model performance. This demonstrates that α serves as an effective tuning parameter to balance learning quality and communication cost in the proposed CFL framework.

Table 2
www.frontiersin.org

Table 2. Impact of the weighting factor α on learning performance and communication energy.

6 Conclusion

This study investigates the trade-off between learning performance and communication energy consumption in CFL, focusing on how cluster head selection and device association affect model training and energy overhead. To collectively optimize model performance and communication efficiency in wireless FL, we first conduct a convergence analysis linking global loss to inter-cluster data imbalance and use cosine similarity to quantify distributional dissimilarity. An optimization model of training loss is then constructed based on cosine gradient similarity, while communication energy is modeled as a function of transmission distance. Finally, a clustering algorithm is proposed to jointly schedule cosine similarity and communication distance for solving the reformulated combined optimization problem. The simulation results show that the proposed method markedly reduces communication energy while improving model accuracy.

In future work, we aim to introduce data-size–aware weighting mechanisms to further optimize client selection and matching, along with adaptive channel allocation strategies to extend the applicability of the method to heterogeneous devices and non-uniform wireless environments. These directions aim to improve both the scalability and robustness of the CFL framework, providing a more comprehensive solution for real-world federated learning scenarios.

Data availability statement

The datasets generated during this study are not publicly available due to the following restrictions: 1. the simulation data and model parameters are integral to the proprietary research methodology developed in this work. 2. The training data consist of standard benchmark datasets that are already publicly available through their original sources. 3. Raw gradient information and device-specific data cannot be shared as they may contain sensitive information about the federated learning system architecture. Requests to access the datasets should be directed to Zhenning Chen, bGlua19jaGVuQHllYWgubmV0.

Author contributions

ZC: Writing – original draft. ZX: Writing – review and editing. YD: Data curation, Investigation, Validation, Writing – review and editing. YW: Methodology, Supervision, Validation, Writing – review and editing.

Funding

The author(s) declared that financial support was not received for this work and/or its publication.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence, and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Briggs, C., Fan, Z., and Andras, P. (2020). “Federated learning with hierarchical clustering of local updates to improve training on non-iid data,” in 2020 international joint conference on neural networks (IJCNN), 1–9. doi:10.1109/IJCNN48605.2020.9207469

CrossRef Full Text | Google Scholar

Chen, M., Yang, Z., Saad, W., Yin, C., Poor, H. V., and Cui, S. (2020). A joint learning and communications framework for federated learning over wireless networks. IEEE Transactions Wireless Communications 20, 269–283. doi:10.1109/twc.2020.3024629

CrossRef Full Text | Google Scholar

Gao, Z., Xiong, Z., Zhao, C., and Feng, F. (2023). “Clustered federated learning framework with acceleration based on data similarity,” in International conference on algorithms and architectures for parallel processing, 80–92.

Google Scholar

Ghosh, A., Chung, J., Yin, D., and Ramchandran, K. (2021). An efficient framework for clustered federated learning. doi:10.48550/arXiv.2006.04088

CrossRef Full Text | Google Scholar

Konečnỳ, J., McMahan, B., and Ramage, D. (2015). Federated optimization: distributed optimization beyond the datacenter. arXiv Preprint arXiv:1511.03575. doi:10.48550/arXiv.1511.03575

CrossRef Full Text | Google Scholar

Lu, Z., Pan, H., Dai, Y., Si, X., and Zhang, Y. (2024). Federated learning with non-iid data: a survey. IEEE Internet Things J. 11, 19188–19209. doi:10.1109/JIOT.2024.3376548

CrossRef Full Text | Google Scholar

Meng, X., Li, Y., Lu, J., and Ren, X. (2023). An optimization method for non-iid federated learning based on deep reinforcement learning. Sensors 23, 9226. doi:10.3390/s23229226

PubMed Abstract | CrossRef Full Text | Google Scholar

Nishio, T., and Yonetani, R. (2019). “Client selection for federated learning with heterogeneous resources in Mobile edge,” in ICC 2019 - 2019 IEEE International Conference on Communications (ICC), 1–7. doi:10.1109/ICC.2019.8761315

CrossRef Full Text | Google Scholar

Taïk, A., Mlika, Z., and Cherkaoui, S. (2022). Clustered vehicular federated learning: process and optimization. IEEE Trans. Intelligent Transp. Syst. 23, 25371–25383. doi:10.1109/TITS.2022.3149860

CrossRef Full Text | Google Scholar

Tan, Y., Long, G., Liu, L., Zhou, T., Lu, Q., Jiang, J., et al. (2022). Fedproto: Federated prototype learning across heterogeneous clients. Proc. AAAI Conference Artificial Intelligence 36, 8432–8440. doi:10.1609/aaai.v36i8.20819

CrossRef Full Text | Google Scholar

Tran, D. T., Ha, N. B., Nguyen, V.-D., and Wong, K.-S. (2025). Sherl-fl: when representation learning meets split learning in hierarchical federated learning. arXiv Preprint arXiv:2508.08339. doi:10.48550/arXiv.2508.08339

CrossRef Full Text | Google Scholar

Wan, S., Lu, J., Fan, P., Shao, Y., Peng, C., and Letaief, K. B. (2021). Convergence analysis and system design for federated learning over wireless networks. IEEE J. Sel. Areas Commun. 39, 3622–3639. doi:10.1109/jsac.2021.3118351

CrossRef Full Text | Google Scholar

Wang, J., Wang, S., Chen, R.-R., and Ji, M. (2020). Local averaging helps: hierarchical federated learning and convergence analysis. arXiv Preprint arXiv:2010. doi:10.48550/arXiv.2010.12998

CrossRef Full Text | Google Scholar

Xia, W., Quek, T. Q. S., Guo, K., Wen, W., Yang, H. H., and Zhu, H. (2020). Multi-armed bandit-based client scheduling for federated learning. IEEE Trans. Wirel. Commun. 19, 7108–7123. doi:10.1109/TWC.2020.3008091

CrossRef Full Text | Google Scholar

Yan, Y., Tong, X., and Wang, S. (2024). Clustered federated learning in heterogeneous environment. IEEE Trans. Neural Netw. Learn. Syst. 35, 12796–12809. doi:10.1109/TNNLS.2023.3264740

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, M., Xu, J., Ding, W., and Liu, Y. (2024). Fedhap: federated hashing with global prototypes for cross-silo retrieval. IEEE Trans. Parallel Distributed Syst. 35, 592–603. doi:10.1109/TPDS.2023.3324426

CrossRef Full Text | Google Scholar

Zeng, D., Hu, X., Liu, S., Yu, Y., Wang, Q., and Xu, Z. (2023). Stochastic clustered federated learning. doi:10.48550/arXiv.2303.00897

CrossRef Full Text | Google Scholar

Zhang, Y., Chen, H., Lin, Z., Chen, Z., and Zhao, J. (2024). Fedac: an adaptive clustered federated learning framework for heterogeneous data. doi:10.48550/arXiv.2403.16460

CrossRef Full Text | Google Scholar

Keywords: device clustering, energy efficiency, federated learning, gradient similarity, wireless communications

Citation: Chen Z, Xu Z, Ding Y and Wang Y (2026) Data- and distance-aware clustering for scalable wireless federated learning. Front. Commun. Netw. 7:1748815. doi: 10.3389/frcmn.2026.1748815

Received: 18 November 2025; Accepted: 12 January 2026;
Published: 09 February 2026.

Edited by:

Osama Amin, King Abdullah University of Science and Technology, Saudi Arabia

Reviewed by:

Ahmad Bazzi, New York University Abu Dhabi, United Arab Emirates
Selvaraj Kandasamy, PSNA College of Engineering and Technology, India

Copyright © 2026 Chen, Xu, Ding and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Zhenning Chen, bGlua19jaGVuQHllYWgubmV0

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.