Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Artif. Intell., 04 February 2026

Sec. Machine Learning and Artificial Intelligence

Volume 9 - 2026 | https://doi.org/10.3389/frai.2026.1718193

This article is part of the Research TopicAdvanced Machine Learning Techniques for Single or Multi-Modal Information ProcessingView all 8 articles

A new clustered federated learning algorithm for heterogeneous data in high-precision wireless sensing

Zongrui TianZongrui TianJiasheng Tian
Jiasheng Tian*
  • School of Electronic Information and Communications, Huazhong University of Science and Technology, Wuhan, China

Introduction: This article studies a new clustering-based federated learning algorithm that leverages Kullback-Leibler (KL) divergence to tackle heterogeneous data in wireless sensing environments.

Methods: Firstly, highdimensional heterogeneous data is subjected to principal component analysis to generate dimension-reduced representations, thereby reducing computational complexity. Secondly, the KL divergence distances between each pair of clients are calculated, followed by clustering according to the minimum threshold. The new KL divergence distance between the aggregated clients and others is taken as the average of the two. Finally, the federated learning training is conducted within each cluster to obtain a personalized model based on the classic wireless datasets.

Results and Discussion: After the personalized models are tested, clients are reclustered and the models are updated—that is, a series of iterative operations, the optimal number of clusters and recognition accuracy are obtained. The test results show that the proposed algorithm based on KL divergence has higher recognition accuracy than several reported ones.

1 Introduction

With the rapid development of Internet of Things (IoT) technology and 6G network architecture, wireless sensing technology has been widely applied in scenarios such as smart home, industrial monitoring, health care and environmental perception. These applications generate massive amounts of sensing data that contain valuable information for decision-making. However, these raw data scattered across various devices are directly transmitted to cloud computing centers for centralized processing, resulting in high system communication overhead and a potential threat in safeguarding user data privacy.

However, traditional federated learning (FL) algorithms (e.g., FedAvg) assume that data across clients follows an Independent and Identically Distributed (IID) distribution, which is difficult to satisfy in practical wireless sensing environments. Due to differences in sensing device types, deployment locations, and working conditions, data among clients often exhibits significant heterogeneity (i.e., Non-IID), including distribution shifts, feature space discrepancies, and label distribution imbalance. This heterogeneity leads to performance degradation of the globally shared model.

To tackle the Non-IID problem in FL, personalized federated learning (PFL) has emerged as a research hotspot. Existing personalized methods can be roughly divided into three categories: (1) regularization-based methods, which introduce additional regularization terms to constrain the local model updates, balancing the consistency between local and global models; (2) model adaptation-based methods, which adjust specific components of the global model or fine-tune local model parameters to adapt to local data; (3) clustering-based methods, which group clients with similar data characteristics into clusters, and train cluster-specific models to achieve personalization. Among these, clustering-based methods are particularly suitable for wireless sensing scenarios due to their low computational overhead and strong scalability. However, existing clustering strategies in FL mainly rely on similarity metrics such as cosine similarity or Euclidean distance, which focus on feature space distance but fail to effectively capture the distribution differences between heterogeneous sensing data. This limitation leads to inaccurate clustering results, thereby affecting the performance of personalized models.

Kullback–Leibler (KL) divergence is a classic metric for measuring the difference between two probability distributions, which can quantitatively characterize the distribution deviation of heterogeneous data. Compared with cosine similarity (which focuses on direction consistency) and Euclidean distance (which focuses on feature value difference), KL divergence is more suitable for describing the intrinsic heterogeneity of sensing data. Motivated by this, this article proposes a KL divergence-based PFL method for wireless sensing environments.

The main contributions of this article are summarized as follows: (1) a clustering strategy based on KL divergence is proposed to effectively capture the distribution differences of heterogeneous sensing data, improving the accuracy of client clustering compared with traditional similarity metrics; (2) a PFL framework for wireless sensing is designed, which realizes model customization through cluster-specific training and iterative optimization, adapting to the data heterogeneity of sensing human behaviors; (3) extensive experiments are conducted on two classic wireless sensing datasets to verify the effectiveness of the proposed PFL algorithm based on KL divergence. The results demonstrate that the algorithm outperforms serval reported algorithms in terms of recognition accuracy.

2 Related work

Federated learning (FL) offers a solution to the challenge of scattered data hindering the centralized learning. In McMahan et al. (2017), first proposed the concepts of FL and verified its effectiveness and feasibility for collaborative model training without aggregating user data to a central server (Sattler et al., 2019). However, in data-heterogeneous environments, FL algorithms often suffer from significant performance degradation (Zhao et al., 2018). To address the challenge of data heterogeneity (Pang et al., 2025a,b), many researchers have proposed a range of improved methods such as personalized federated learning (PFL) algorithms (Tan et al., 2022; Arivazhagan et al., 2019).

One personalized approach involves designating some layers of a neural network as personalized layers and the rest as globally shared layers (Arivazhagan et al., 2019; Liang et al., 2022). In Arivazhagan et al. (2019), proposed the FedPer algorithm, which adopted a “base layers + personalized layers” design. In Liang et al. (2022), proposed LG-FedAvg, where the last several layers of the neural network were designated as personalized components. However, how to properly divide base layers and personalized layers in these algorithms remains an area requiring further research. In Collins et al. (2021), proposed the FedRep algorithm and designed the classification head of a neural network as the personalized component, while all other layers were designed for global federated training. Nevertheless, FedRep’s performance depends on the effectiveness of global representations; moreover, in real-world scenarios, if shared features across data from different clients are either less prominent or difficult to learn, this algorithm is at the risk of performance degradation.

Another personalized approach is that each personalized model exhibits a certain “degree of personalization” relative to the global model (Dinh et al., 2020; Deng et al., 2003; Li et al., 2021; Zhang et al., 2020). In Dinh et al. (2020), proposed the PFL algorithm “pFedMe” and introduced a regularized loss term to balance the trade off between personalization and generalization. However, pFedMe faced challenges in properly selecting hyperparameters to quantify this degree of personalization. In the same year, Deng et al. (2003) introduced an adaptive weight adjustment mechanism to dynamically tune the weight ratio between the global model and local personalized model in the final model. However, the adaptive weights are dynamically determined based on the loss, leading to insignificant improvement effects. In Li et al. (2021), proposed the Ditto algorithm, which employed the traditional FedAvg method for global model optimization. During synchronous training, FedAvg adopted a relatively global regularized model as the local personalized model. Nevertheless, using FedAvg for global training is unfavorable for convergence in data-heterogeneous scenarios, and Ditto also faces challenges in selecting hyperparameters to quantify the degree of personalization.

The aforementioned PFL algorithms generally focus on the personalized components of individual models relative to the global model, but do not directly consider the connections between two personalized models-specifically, the similarity among models across multiple clients (Zhang et al., 2020; Huang et al., 2021).

In Zhang et al. (2020), proposed the FedFomo algorithm, which achieved personalized updates by computing the optimal weighted combination of models for each client. Each client determined its aggregation weights based on the local loss of other clients’ models, resulting in models with lower losses being assigned larger weights. This algorithm improves model performance only to a certain extent. In Huang et al. (2021), put forward the FedAMP algorithm, which emphasized attention mechanisms to enhance pairwise collaboration among clients with similar data distributions. However, these personalized models accounting for the connections between personalized models tend to exhibit notable similarity as a result of such collaborative interactions.

Additionally, numerous studies Briggs et al. (2020), Islam et al. (2024), Ghosh et al. (2020), and Sattler et al. (2020) have focused on clustering based on personalization in FL. For this category of algorithms, the server initially randomly constructed K global models based on a certain type of similarity (e.g., distance similarity, cosine similarity), each associated with a distinct cluster. However, the clustering algorithm (Ghosh et al., 2020) required predefining the number of clusters K and involved frequent communication for transmitting model parameters or frequently varying them. Secondly, many relevant hyperparameters such as thresholds and cluster partitioning conditions were involved, leading to a linear increase in complexity (Sattler et al., 2020).

Therefore, this article studies a new clustered FL algorithm to address the challenge of data heterogeneity—specifically by leveraging Kullback–Leibler (KL) divergence to measure the similarity among multiple clients and performing effective clustering via iterative loops. This new algorithm enables efficient identification of distribution similarity, avoids the server’s arbitrary initial determination of K global models, and both difficulties in hyperparameter selection and the need for frequent communication.

3 Materials and methods

This section will introduce the new clustered federated learning (CFL) algorithm and datasets.

3.1 The new CFL algorithm

The New CFL algorithm includes the Principal Component Analysis (PCA), KL divergence calculation, clustering and federated learning training, etc.

3.1.1 Principal Component Analysis

Principal Component Analysis (PCA) is a statistical method that projects data onto a low-dimensional space via linear transformation, while maximizing the retention of the variance of the original data. In the new CFL algorithm, each client derives a principal component vector matrix using PCA based on its local data; this matrix is regarded as the client’s data feature and is used to measure the distance between different clients. Specifically, for each client’s data matrix XRM × d, after standardization and subsequent PCA processing (including covariance calculation, eigenvalue computation, and eigenvector derivation), a principal component vector matrix URd × c is obtained. Here, M denotes the number of samples, d represents the feature dimension of a single sample, and c is the selected number of principal components (with c < d). c is a user-specified value, which is used to determine the dimension after dimensionality reduction. This process can be expressed as:

U = PCA ( X , c )     (1)

After finishing PCA, we obtain the data features for each client. These data features are typically in the form of matrices, so the matrix vectorization for these data is required to simplify the subsequent KL divergence calculations. Specifically, we vectorize the matrix URd × c into UR1 × (d × c) (either a column vector or a row vector). Following the matrix vectorization, each client is assigned a corresponding characteristic vector U. We then take the absolute value of U and normalized it, as can be given by

U i = | U i j | j = 1 d × c | U i j |     (2)

Where i = 1, 2, …, N represents the number of the client numbers. And thus U is converted into the discrete probability distribution U.

Since the heterogeneous wireless sensing data X including high dimensionality, excessive redundant information, and noise interference, it is necessary for PCA, a classical dimensionality reduction and data preprocessing technique, to map high-dimensional variables to a low-dimensional principal component space via orthogonal transformation while preserving key variation information within the dataset. The low-dimensional data after PCA dimensionality reduction can not only reduce the computational cost of KL divergence and improve clustering efficiency, but also enhance the recognition sensitivity of KL divergence to “category differences in heterogeneous data.”

3.1.2 Calculation of KL divergence

Kullback–Leibler (KL) Divergence is a method for measuring the difference between two probability distributions. The value of KL divergence is non-negative: it equals zero if the two probability distributions are identical, and the smaller the value, the more similar the distributions. The preprocessed vectors U are treated as discrete probability density distribution, and thus two client’s KL divergence is calculated by

KL ( U m ( x ) , U n ( y ) ) = i = 1 c x i log ( x i y i )     (3)

where Um(x) ∈ Rc and Un(y) ∈ Rc are two (m, n) client’s PCA vectors, and xi, yi denote the i-th component of the m-th client vector Um(x) and the n-th client vector Un(y), respectively.

3.1.3 Clustering based on KL divergence

The hierarchical clustering algorithm is a clustering method that constructs a hierarchical structure of clusters through the gradual merging of existing clusters. Specifically, it determines the clustering relationships among clients based on the inter-client distance adjacency matrix B, as given by

B = [ KL 11 KL 12 KL 1 n KL 21 KL 22 KL 2 n KL m 1 KL m 2 KL m n ]     (4)

where m = n generally is the number of clients. First, each client is treated as an individual cluster. Next, using the similarity (or distance) adjacency matrix, locate the two clusters with the highest similarity and merge them into a new cluster. Then, update the similarity adjacency matrix to accurately reflect the structure of the newly formed cluster. Finally, repeat the processes of cluster merging and similarity adjacency matrix updating until either the preset number of clusters or hyperparameter cr (clustered ratio defined as the ratio of the number of clustering operations to the total number of clients N) is attained or all values in the adjacency matrix are above the designated threshold. If KL12 is the minimum value among all KL mn , client 1 and client 2 are merged into a new client 1, and the number of clients is reduced to m − 1(=n − 1). The new KL divergence distance KL1j between the new clustering client 1 and other clients j will be replaced by the average value of the KL divergence distance between the original two client (1, 2) and other clients j(i). Equation 4 can be updated to

B = [ KL 11 KL 12 KL 1 ( n 1 ) KL 21 KL 22 KL 2 ( n 1 ) KL ( m 1 ) 1 KL ( m 1 ) 2 KL ( m 1 ) ( n 1 ) ]     (5)

where

KL 11 = 0     (6)
KL 1 j = ( KL 1 ( j + 1 ) + KL 2 ( j + 1 ) ) / 2 , j = 2 , , n 1     (7)
KL i 1 = ( KL ( i + 1 ) 1 + KL ( i + 1 ) 2 ) / 2 , i = 2 , , m 1     (8)

and

KL i j = KL ( i + 1 ) ( j + 1 ) , i = 2 , , m 1 ; j = 2 , , n 1     (9)

If KL i j (i = 1, 2, …, m−1; j = 1, 2, …, n−1) is still less than the minimum threshold T, the operation of clustering from Equation 4 to Equation 5 will be carried out once again. This course of clustering can be expressed by.

C = { CLU i | i = 1 , 2 , 3 , } = cluster ( B , T , max num )     (10)

which is the clustering set. And T is the threshold, maxnum is the maximum number of clustering operation. CLU i is also a set that contains the id of the clients in this cluster. It can be given as

CLU i = { cid i , j | j = 1 , 2 , 3 , }     (11)

3.1.4 Federated learning training based on KL

The traditional federated learning algorithm FedAvg is executed within each cluster and the personalized federated learning algorithm is carried out among different clusters, meaning there is no longer any interaction of model parameters between different clusters, with only clients within the same cluster interacting with each other’s model parameters.

For client i, the model (wi) is updated locally, the updating formula is expressed by

w i t + 1 = arg min w i t [ f i ( w i t ) ]     (12)

where fi (*) is loss function. When the gradient descent method is adopted, the iteration model

w i t , e + 1 = w i t , e η f i ( w i t , e ) w i t , e     (13)

will be executed repeatedly. And η is learning rate, e stands for epoch E. The server performs a weighted average (pi)based on the updated models of each client within the same cluster (cluid). The updated formula w c l u i d t + 1 is as follows:

w c l u i d t + 1 = i p i w i t + 1     (14)

where t is the number of updating.

The server sends the global model w c l u i d t + 1 to the clients; each client updates the model parameters and trains the model using local data; then send their local models to the server, which aggregates them. The algorithm can be carried out as the following steps: Algorithm description titled "Clustering Federated Learning Algorithm Based on KL Divergence." It outlines the input, output, and initialization process, followed by step-by-step instructions for calculating model parameters. Steps include data upload, PCA, matrix vectorization, KL divergence calculation, clustering, parameter updates, and weighted averaging within clusters over training rounds.

According to Equations 114, we can extract stable feature representations from WiFi CSI (Channel State Information) data or wireless dataset and obtain low-dimensional feature vectors by applying PCA/KL transform. And then we calculate the Kullback–Leibler (KL) divergence of CSI features between clients to generate a similarity matrix. Finally, we obtain a federated learning architecture for collaborative training among clients within each cluster. Each cluster trains a dedicated model to adapt to the signal propagation characteristics of specific areas. The model or local models tailored to a specific sub-distribution of clients could be used to better handle target-oriented sensing tasks where data distribution might be highly localized (e.g., in-area monitoring, like Target-Oriented WiFi Sensing for Respiratory Healthcare: from Indiscriminate Perception to In-Area Sensing), resulting in improving the recognition accuracy.

3.2 The wireless datasets

3.2.1 ARWF wireless dataset

To evaluate the new personalized federated learning (PFL) algorithm, this study conducts simulations and experiments using two wireless datasets, with the target task being human activity recognition for wireless sensing in the future. The first wireless dataset is ARWF (Wang et al., 2019), which comprises 1,116 training samples and 278 test samples. Each sample features a spatial–temporal dimension of 52 × 192. In this dimension, 52 represents the number of subcarriers (a key parameter in wireless communication), and 192 corresponds to the number of time sampling points. For labeling, each sample is assigned 2 types of initial labels: 6 categories of human behavior and 16 categories of positions.

3.2.2 Widar 3.0 dataset

The second wireless dataset, Widar3.0 (Zhang et al., 2021), is a typical wireless dataset that leverages Channel State Information (CSI) for sensing. Each sample in this dataset has a 3-dimensional structure of 2 × 1,000 × 90: specifically, 2 denotes the number of receiving antennas, 1,000 represents the number of time steps, and 90 corresponds to the number of subcarriers-all of which align with the inherent characteristics of CSI data. The dataset contains 6 gesture labels and is divided into 18 clients according to position to simulate the distributed nature of federated learning.

4 Results and discussion

In this section, we primarily evaluate the recognition accuracy of the proposed new clustered FL algorithm based on KL divergence (KLCFL) and its sensitivity to key factors based on wireless datasets.

4.1 Analysis of influencing factors of the new algorithm

The experiments primarily evaluate the effectiveness of the clustering and the effects of three key parameters on model training: B (the batch size of training data), E (local training rounds), and hyperparameter (clustering ratio) cr.

4.1.1 Impact of batch size B

For the ARWF dataset, under the conditions of a fixed number of clients N = 32, the number of global rounds R = 100, and E = 1, the test results are presented in Figure 1.

Figure 1
Graph showing test accuracy over 100 global rounds for different batch sizes (B=5, 10, 20, 50) in a KLFL setup with N=32 and E=1 for ARWF. Accuracy improves as batch size increases, with B=50 achieving the highest accuracy.

Figure 1. The effects of B on test accuracy based on ARWF.

When B = 5, 10 and 20, the final test (recognition) accuracy of the KLCFL algorithm remains largely stable between 70 and 80% and exhibits relatively small fluctuations, as shown in Figure 1. When B = 5, the test accuracy of KLCFL increases rapidly at R = 10 or more, fluctuates slightly, and finally stabilizes at 80% at R = 100, as illustrated by the blue solid line in Figure 1. When B = 10, the test accuracy of KLCFL increases rapidly at R = 20 or more, fluctuates slightly, and finally stabilizes at 80% at R = 100, as illustrated by the red solid line. When B = 20, the test accuracy of KLCFL increases rapidly at R = 45 or more, fluctuates and finally stabilizes at 80% or so at R = 100, as shown by the blue star-solid line. However, when B = 50, the test accuracy of KLCFL begins increasing only at R = 90 and arrives at 55% or so at R = 100, as illustrated by the red star-solid line in Figure 1. In fact, B = 50 exceeds the total number of samples per client, which essentially amounts to using all samples in a single batch. This tends to make KLCFL overly fitted to the training set, ultimately leading to overfitting. Furthermore, a larger B typically results in smaller gradient variations, increasing the likelihood of KLCFL getting trapped in local optimal points. In contrast, a smaller B leads to larger gradient fluctuations, thereby reducing the likelihood of KLCFL getting stuck in local optimal points. It is also clear that the smaller the batch size B, the faster KLCFL converges, which in turn contributes to high recognition accuracy.

For the Widar3.0 dataset, the number of clients N = 18. When N = 18 and E = 1, we investigated the impact of B (batch size) on KLCFL’s test (recognition) accuracy. Specifically, when B = 8, 16, and 32 (limited by the sample size of Widar 3.0 dataset), the corresponding test accuracies vary with global rounds R, as illustrated by Figure 2A. Although the trend affected by B is somewhat similar to that as mentioned earlier shown in Figure 1—the larger B is, the faster the convergence—the impact is insignificant and not obvious. Such discrepancies can be neglected, showing strong robustness. It is clear that as R increases, the recognition accuracy for different B values converges quickly and achieves excellent performance, exceeding 99%. At R = 100, the test accuracies are 0.9984, 0.9983, and 0.9921, respectively, as shown in Figure 2B. Although the test accuracy is relatively high—partly attributed to the high quality of the dataset—the influence pattern of B remains consistent: the smaller B value tends to relatively higher accuracy and faster convergence, and this kind of feature is affected by the quality of the dataset to a certain extent.

Figure 2
Two graphs illustrating test accuracy for KLCFL with different batch sizes for Widar 3.0. Graph (a) is a line chart showing accuracy improvement over global rounds for batch sizes 8, 16, and 32. Graph (b) is a bar chart depicting high accuracy percentages: 99.84%, 99.83%, and 99.21% for batch sizes 8, 16, and 32, respectively.

Figure 2. The effects of B on test accuracy based on Widar3.0 dataset (a/b).

For these two datasets, this algorithm can converge quickly under different B values (B < 50), demonstrating strong robustness. For a given global round R (e.g., R = 100), the recognition accuracy of the algorithm KLCFL varies with different datasets, and the trend where the convergence speed tends to decrease as B increases also differs to varying degrees. Moreover, the convergence stability and amplitude fluctuation remain roughly consistent, with negligible differences within a certain range.

4.1.2 Impact of local training round E

For the ARWF dataset, let N = 32 and B = 5, we investigate the impact of E on KLCFL test accuracy. The results are presented in Figure 3.

Figure 3
Line graph showing test accuracy versus global rounds for different values of E (1, 3, 5, 10) in a KLCFL model. Accuracy increases with more rounds, leveling off near 0.8.

Figure 3. The effects of E on test accuracy based on ARWF.

The tested results demonstrate that as E increases (e.g., E = 1, 3, 5, 10), the recognition/test accuracy of KLCFL exhibits increasingly rapidly convergence with global rounds R until the recognition accuracy stabilizes at 0.8 or so. Additionally, the magnitude of fluctuations decreases progressively, while the overall performance remains stable, as shown in Figure 3. Under three different values of E, the differences in convergence accuracy and fluctuation amplitude are relatively small. However, there is a significant variation in convergence speed—with faster convergence observed as E increases (e.g., R < 20). Overall, local training rounds E exert little impact on recognition accuracy, indicating a certain degree of robustness.

Similarly, for the Widar3.0 dataset, when N = 18 and B = 5, the KLCFL algorithm converges rapidly, as shown in Figure 4A. Additionally, it converges faster as E increases (e.g., E = 1, 3, 5, 10 at R < 20), although negligible differences still exist. Moreover, fluctuation magnitudes decrease progressively with R increasing and these differences can be negligible for E = 1, 3, 5.10. For a given global round R = 100, across all tested values of local training rounds E (e.g., E = 1, 3, 5, 10), the KLCFL algorithm achieves a high recognition accuracy exceeding 98%. This illustrates that the impact of local training rounds E on recognition accuracy is minimal and can be neglected, as shown in Figure 4B. In Figure 4B, when R = 100, the test accuracy of the KLCFL algorithm is 98.96, 99.78, 99.83 and 99.62% for E = 1, 3, 5.10, respectively. In terms of being affected by E, the KLCFL algorithm is less affected by it and exhibits greater robustness.

Figure 4
Graph (a) shows test accuracy against global rounds for different epochs (E=1, 3, 5, 10) in the Widar 3.0 test, displaying increasing accuracy trends as rounds progress. Graph (b) is a bar chart showing test accuracy percentages at different epochs (1, 3, 5, 10), all achieving over 98.96%. Both graphs pertain to KLCFL with N=18 and B=5.

Figure 4. The effects of E on test accuracy based on Widar3.0 (a, b).

4.1.3 Impact of clustering ratio cr

For the ARWF dataset, with N = 32, E = 1, and B = 5, we conducted simulation experiments by varying the clustering ratio (cr) of the KLCFL algorithm- defined as the ratio of the number of clustering operations to the total number of clients N. Specifically, cr × N denotes the number of clustering operation, with 9 clusters obtained when cr = 0.3. Figure 5 presents the curves illustrating test accuracy variations under different cr settings.

Figure 5
Line graph showing test accuracy over 100 global rounds for ARWF with varying compression ratios (cr). Five lines indicate cr values: 0.1, 0.3, 0.5, 0.7, and 0.9. Accuracy improves rapidly at first, then stabilizes around 60 rounds.

Figure 5. Curves of test accuracy variations of the KLCFL with different CR based on the ARWF dataset.

The new clustered federal learning algorithm KLCFL was tested with clustering ratio cr set to 0.1, 0.3, 0.5, 07, and 0.9 based on ARWF dataset. The test results indicate that KLCFL’s recognition/test accuracy increases as cr rises—particularly for cr <0.5—and maintains a relatively high level of approximately 80%, as shown in Figure 5. However, when cr = 0.7 and 0.9, the recognition accuracy begins to decline, dropping significantly to roughly 50%. This indicates that an excessively large cr exerts a substantial negative impact on model performance. A higher number of clusters means many clients with highly heterogeneous data may be grouped into the same cluster. This will inevitably lead to the same model being used to predict heterogeneous data, resulting in a decrease in recognition accuracy. It is clear that the performance of KLCFL degrades noticeably when the clustering ratio cr is excessively large. Therefore, the reasonable selection of the clustering ratio is of crucial importance, as it directly affects the algorithm’s recognition accuracy. From Figure 5, it is clear that the KLCFL algorithm can achieve satisfactory recognition accuracy when cr falls within the typical range of 0.01–0.5, demonstrating that the KLCFL algorithm exhibits moderate robustness in a certain degree.

For the Widar3.0 dataset with N = 18, as cr increases, the differences in the algorithm’s convergence speed become increasingly pronounced, as shown in the Figure 6A. The convergence speed of the algorithm becomes increasingly slow as the clustering ratio (cr) increases, and the accuracy also decreases—for example, the recognition accuracy of the KLCFL algorithm is 86.83% at cr = 0.9, and is more than 97% at cr <0.7, which is consistent with the previous simulation results. Therefore, the reasonable selection of the clustering ratio cr is of crucial importance, as it directly affects the algorithm’s recognition accuracy and convergence speed, as shown in Figure 6A. However, when R = 100, the algorithm achieves relatively high recognition accuracy (as high as 97%) for this dataset. Specifically, KLCFL achieves a recognition accuracy of 98% across all tested cr values (cr = 0.1, 0.3, and 0.5). Only when cr = 0.9 does the test accuracy decrease slightly, dropping to around 86%, as shown in Figure 6B.

Figure 6
Two-part image showing test accuracy in a KLCFL study for Widar 3.0. (a) Line graph depicting test accuracy over 100 global rounds for clustering ratios (cr) of 0.1, 0.3, 0.5, 0.7, and 0.9, with accuracy increasing as cr decreases. (b) Bar graph showing test accuracy percentages for different cr values: 99.85% at 0.1, 99.83% at 0.3, 98.72% at 0.5, 97.63% at 0.7, and 86.83% at 0.9.

Figure 6. Curves of test accuracy variations of the KLCFL with different hyperparameters based on Widar3.0 (a, b).

Obviously, regardless of whether the dataset is of high or low quality, the clustering ratio largely affects the convergence speed and test accuracy. An excessively large clustering ratio will reduce the convergence speed and recognition accuracy. Therefore, the reasonable selection of the clustering ratio is crucial. For a dataset with unknown characteristics, it is generally feasible to refer to the test results of traversing multiple clustering ratios and select the optimal one by comprehensively considering the convergence speed and final recognition accuracy. Of course, for the two typical datasets selected in this study, choosing a clustering ratio below 0.5 is generally appropriate, which also ensures a certain degree of robustness.

4.2 Performance comparison between the new algorithm and several representative algorithms

To evaluate the performance of the new CFL algorithms (KLCFL) proposed in this study, we conduct a simulation-based performance comparison of KLCFL against several representative benchmark algorithms, including FedAvg (McMahan et al., 2017), FedRep (Collins et al., 2021), FedPer (Arivazhagan et al., 2019), PACFL (Vahidian et al., 2023; Aslam, 2023), and pFedMe (Dinh et al., 2020).

For the wireless dataset ARWF, we again observe that the KLCFL algorithm converges rapidly to notably high recognition accuracies: 83.05% (N = 16 defined by 16 positions) and 77.33% (N = 32, due to two clients at each position) at R = 100, as shown in Table 1. Among all the algorithms selected for comparison, the KLCFL algorithm proposed in this paper basically achieves the optimal/highest recognition accuracy and is worthy of application in practice. Table 1 presents the recognition accuracies of other six additional benchmark algorithms at R = 100 and N = 32, all of which are significantly lower than that of the KLCFL algorithm: (McMahan et al., 2017) a conventional FL algorithm, lacks sufficient consideration of personalization. FedPer (Arivazhagan et al., 2019) incorporates personalization, dividing the neural network model into base layers and personalized layers. The base layers are shared among all clients and updated by aggregation (e.g., by using FedAvg) on the server side. The personalized layers are trained only locally using the client’s own data, and do not participate in server-side aggregation. This architecture enables the model to both benefit from global collaboration and adapt to the specific data distribution of each client. The number of personalized layers (denoted as KP) is adjustable, which effectively addresses the data heterogeneity (Non-Independently and Identically Distributed-Non-IID) issue in federated learning. However, for FedPer (Arivazhagan et al., 2019) the precise partitioning of the “base layer and personalized layer” remains an open research question: the fixed division lacks theoretical basis and cannot be adjusted according to the training process or data characteristics. Moreover, the optimal KP value varies across different datasets and model architectures. To improve the recognition accuracy of FedPer, the basis for division and adaptive dynamic adjustment of KP should be performed according to changes in the training process and data characteristics. The algorithm pFedMe (Liang et al., 2022) encounters difficulties in selecting appropriate hyperparameters to quantify the degree of personalization. FedFomo (Zhang et al., 2020) biases personalized weights toward models with smaller losses. FedRep (Collins et al., 2021) fully addresses personalization (by designating the classification head as the personalized component) and effectively mitigates data heterogeneity in a certain degree. Its recognition accuracy is the closest to that of the KLCFL algorithm. The PACFL algorithm (Wang et al., 2019) primarily identifies distributional similarity by analyzing the principal angles between client data subspaces, which is called as cosine similarity.

Table 1
www.frontiersin.org

Table 1. Test accuracy of KLCFL and several algorithms based on ARWF dataset (R = 100).

After PCA vectors are normalized (e.g., low-dimensional representations), the Kullback–Leibler (KL) divergence can directly capture the “degree of information difference” between distributions. Cosine similarity (CS) only focuses on the consistency of vector directions, ignoring the difference in distribution intensity. In addition, in relation to the KL divergence, Euclidean distance (ED) merely measures the geometric spatial distance of vectors, failing to account for the probabilistic characteristics of distributions. Neither of CS and ED can address the core of the Non-IID problem: distribution deviation.

Additionally, KL divergence also has the advantage of effectively handling heterogeneous data. KL divergence is sensitive to the asymmetry and tail characteristics of distributions, effectively distinguishing between “distribution shift” and “pure numerical differences” represented by PCA vectors. Cosine similarity is insensitive to scale and ignores the distributional significance of the magnitudes of principal components in PCA vectors (e.g., the principal component magnitudes of a customer’s model concentrated in a few dimensions may correspond to specific distribution patterns). Euclidean distance is significantly affected by feature scales; even after PCA standardization, it may misclassify non-distributional numerical differences as heterogeneity. Moreover, the quantitative result of KL divergence directly reflects the “difficulty of distribution alignment,” enabling more accurate screening of customers with shareable model parameters after clustering. Cosine similarity may group customers “with consistent directions but significant differences in distribution shapes” into one cluster, while Euclidean distance may misclassify customers “that are geometrically close but distributional heterogeneous.”

Figure 7 compares the convergence processes and recognition accuracies of the new algorithm KLCFL with several representative algorithms. It is evident that KLCFL basically outperforms those alternatives in four key aspects: recognition accuracy, convergence speed, stability, and amplitude fluctuation. In Figure 7, although the convergence speed of the KLCFL algorithm ranks the second (not the fastest) and its recognition accuracy is not the highest, its convergence process shows that the algorithm has a relatively fast convergence speed, a smaller jitter amplitude, and a relatively high final recognition accuracy, which is also relatively consistent with the ideal recognition accuracy. However, most other algorithms (such as FedAvg, FedFomo, and pFedMe) converge slowly and achieve much lower final recognition accuracy.

Figure 7
Line graph showing test accuracy over 100 global rounds with different federated learning algorithms indicated by varied line styles and colors. Performance generally increases, with KLCLFL reaching the highest accuracy. Key indicates specific algorithms.

Figure 7. Test (recognition) accuracy of KLCFL and several representative algorithms based on ARWF dataset N = 32.

Furthermore, to further verify the performance of the new algorithm, we again conducted tests using the Widar3.0 dataset. The test results demonstrate that by adjusting parameter cr (cr = 0.1), the KLCFL algorithm can effectively address data heterogeneity and achieve high recognition accuracy, as shown in Figure 8. Figure 8 presents a comparison of the recognition accuracies between the KLCFL algorithm and several other representative algorithms. It is clear that the recognition accuracy and stability of the KLCFL algorithm are significantly higher than that of the other several algorithms.

Figure 8
Line graph showing test accuracy over global rounds, labeled R, with N equals sixteen. The graph compares methods: FedAvg, FedFomo, pFedMe, PACFL, FedRep, FedPer, and KLCFL. FedAvg leads with highest accuracy, followed by other methods with varying performance.

Figure 8. Test (recognition) accuracy of KLCFL and several representative algorithms based on Widar3.0 dataset.

In Figure 8, the pFedMe algorithm converges relatively fast and achieves high recognition accuracy (0.965), but its amplitude fluctuation is somewhat significant—it still fluctuates when R reaches 100. The KL-based algorithm ranks the third fastest in convergence, with the smallest amplitude fluctuation, and stably converges to the final ideal recognition accuracy (0.998). FedFomo converges relatively fast, but its recognition accuracy is unsatisfactory (around 80%). FedRep exhibits a moderate convergence speed and good stability, with an acceptable final recognition accuracy (0.975). As for the other three algorithms, by comparison, they converge slowly and yield low recognition accuracy (FedFomo: 0.897, Fedavg:0.867, Fedper0.799) at R = 100.

In addition, on ARWF dataset, PACFL converges the fastest with small amplitude fluctuation and achieves excellent final accuracy. However, on Widar 3.0 dataset, it converges slowly with somewhat significant fluctuation, and the final recognition accuracy is not satisfactory. Similarly, pFedMe exhibits a moderate convergence speed on ARWF dataset but with relatively large fluctuation, and its final recognition accuracy is unsatisfactory and relatively low at R = 100. That is to say, these two algorithms show considerable variations in recognition performance across different datasets, indicating poor robustness. In contrast, the KLCFL algorithm proposed in this paper demonstrates the most outstanding recognition performance and stability on both datasets, boasting strong robustness and the ability to stably and effectively handle heterogeneous data.

Of course, the recognition accuracy of other algorithms may be improved to some extent through parameter optimization, but the KLCFL algorithm in this paper already achieves a relatively ideal high recognition accuracy. Even if one or two other algorithms are optimized and their accuracy may exceed that of this KLCFL algorithm, the improvement will be limited. In particular, compared with PACFL, a clustering algorithm of the same type with optimized parameters, the proposed algorithm exhibits higher accuracy and better convergence performance, effectively addresses the data heterogeneity issue, and possesses a certain degree of robustness.

5 Conclusion

Wireless data from different regions typically exhibits high heterogeneity, with limited labeled data available (details are reported in a separate study). Traditional FL struggles to achieve efficient and fast distributed model training under such circumstances. How to develop efficient training and adaptation methods for distributed wireless sensing models has become a major challenge in the development of 6G integrated communication and sensing networks. To tackle the challenges brought by heterogeneous wireless data, this study proposes an improved Personalized Federated Learning (PFL) algorithm.

This KLCFL algorithm incorporates the Kullback–Leibler (KL) divergence distance between each pair of clients, thereby enabling flexible clustering of clients with similar characteristics and significantly improving the recognition accuracy of wireless sensing in a great degree. For the ARWF dataset and Widar 3.0 dataset, KLCFL can all converges rapidly with small amplitude fluctuation and achieves excellent final accuracy, ranking among the excellent clustering algorithms.

Firstly, this paper carried out the PCA of the two datasets (ARWF and Widar3.0) and obtain a principal component vector matrix. And then it finished calculating the Kullback–Leibler (KL) divergence distance for clustering. Secondly, these two sets of wireless datasets were used to study the impacts of variations in batch size (B), local training epochs (E) and clustering ratio cr on the recognition accuracy of KLCFL. For the ARWF dataset, at B = 5 or 10, E = 2 or <4, and cr = 0.3 or <0.5, the KLCFL algorithm can achieve optimal recognition performance (77% ~ 83%). With respect to the Widar 3.0 Dataset, since it features relatively low interference levels and high data quality, the three key influencing factors exert minimal impact on the KLCFL algorithm’s recognition accuracy—thus enabling the KLCFL algorithm to exhibit good robustness. Generally, when B < 30, E < 10, and cr < 0.5, the KLCFL algorithm achieves a recognition accuracy of 98% or higher. Finally, the recognition accuracy of the KLCFL algorithm was compared and analyzed with that of several reported algorithms. For the ARWF Wireless Dataset, the recognition accuracy of the KLCFL algorithm reaches 83.05% (N = 16) and 79.53% (N = 32) when the global round R = 100, which is higher than those of other algorithms, as shown in Figure 7, Table 1. Moreover, KLCFL converges faster than other algorithms and exhibits smaller fluctuations in amplitude. For the Widar3.0 wireless dataset, the recognition accuracy of the KLCFL algorithm is 99%, outperforming several reported algorithms, as shown in Figure 8.

From Figures 7, 8, this KLCFL algorithm not only demonstrates high recognition accuracy and fast convergence, but also exhibits small fluctuations and high stability.

Based on these datasets KLCFL not only achieves higher recognition accuracy than the current reported algorithms, but also exhibits a faster convergence speed, smaller fluctuation amplitude, and greater stability during the convergence process.

It is clear that the proposed KLCFL is an excellent PFL algorithm that can effectively address the heterogeneity of wireless data and achieve high recognition accuracy.

Firstly, this KLCFL algorithm can quantify differences among heterogeneous distributions and provide theoretic support for the rationality of clustering: In federated scenarios, data from individual clients often exhibits Non-IID (Non-Independently and Identically Distributed) characteristics. KL divergence can accurately measure the asymmetric differences in data distributions across various clients, provide a quantitative basis for cross-client data clustering, overcome the limitations of traditional similarity metrics (e.g., Euclidean distance) regarding data distribution assumptions, and theoretically validate the feasibility of heterogeneous data clustering.

Secondly, the KLCFL algorithm can guide the direction of collaborative optimization for heterogeneous data: Based on clustering results derived from KL divergence, the distribution patterns of data heterogeneity in federated systems (such as the degree of local data deviation, variations in category distribution, and so on) can be revealed. This provides theoretical guidance for data partitioning and the collaborative updating of model parameters in federated training.

Finally, the KLCFL algorithm can improve the theoretical framework of federated clustering: Most existing federated clustering methods rely on the assumption of independent and identical data distribution. By integrating distribution difference measurement into the federated clustering framework, KL divergence clustering fills the theoretical gap in federated clustering for heterogeneous data and offers a referable theoretical paradigm for the design of clustering strategies in subsequent heterogeneous federated learning.

However, the Kullback–Leibler (KL) divergence is inherently asymmetric, which gives rise to unstable optimization trajectories during model training. Additionally, when two discrete distributions are non-overlapping or contain zero-probability entries, the KL divergence fails to effectively quantify the magnitude of their discrepancy—often resulting in undefined or infinite values. Furthermore, its robustness against noise, interference, and distribution shifts remains insufficient, posing significant challenges in real-world scenarios characterized by inherent perturbations.

To address these limitations, potential future research directions are outlined as follows: First, developing symmetric variants of the KL divergence (e.g., symmetric KL divergence or extensions inspired by the Jensen-Shannon divergence) to mitigate optimization instability induced by asymmetry. Second, adopting a minimal value substitution strategy (e.g., replacing zero entries with an infinitesimally small positive value) to avoid undefined results when zero-probability events occur. Third, when WiFi data exhibits distribution shifts due to dynamic environmental changes (e.g., personnel movement, signal occlusion), the KL divergence struggles to distinguish between “true class differences” and “distribution differences caused by scene interference.” This limitation is particularly prominent in dynamic scenarios and can be addressed by integrating improved KL measurement methods with scenario features and data characteristics, which may serve as a viable solution. Fourth, future research could focus on enhancing resilience against noise and interference by targeting physical attack scenarios—such as defending against physical layer attacks (PhyFinAtt) and keystroke sniffing (KeystrokeSniffer) (Liu et al., 2025; Chai et al., 2025). PhyFinAtt is an undetectable attack framework specifically designed to undermine PHY layer fingerprint-based WiFi authentication. KeystrokeSniffer demonstrates how an off-the-shelf smartphone can eavesdrop on keyboard input from anywhere. By applying PCA/KL to stabilize CSI features, KLCFL may make PHY fingerprints more resistant to environmental manipulation. By Clustering techniques may identify when an environment is being perturbed to attack PHY fingerprints, implementing online PCA updates would allow the system to continuously adapt to changing environments. For mitigation of Keystroke Sniffing, PCA/KL could transform WiFi signals in a way that obscures keystroke-related patterns. Clustering techniques could distinguish between harmless environmental variations and suspicious keystroke-related signal patterns, etc.

The proposed KLCFL and its PCA/KL approach reveal a promising direction for addressing the challenges of heterogeneous WiFi sensing data in security applications. By systematically reducing noise, stabilizing features, and clustering similar patterns, KLCFL could significantly enhance defenses against both PhyFinAtt and keystroke sniffing attacks.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

ZT: Formal analysis, Methodology, Writing – original draft, Writing – review & editing, Conceptualization, Data curation, Resources, Software. JT: Formal analysis, Methodology, Writing – original draft, Writing – review & editing, Investigation, Validation, Visualization.

Funding

The author(s) declared that financial support was not received for this work and/or its publication.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that Generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Arivazhagan, M. G., Aggarwal, V., Singh, A. K., and Choudhary, S. (2019). Federated learning with personalization layers. arXiv, 108, 1–13. doi: 10.48550/arXiv.1912.00818 (1912.00818)

Crossref Full Text | Google Scholar

Aslam, Z. (2023). Data leakage analysis in wireless networks using supervised and unsupervised testing. Innov. Softw. 4, 52–62. doi: 10.48168/innosoft.s12.a108

Crossref Full Text | Google Scholar

Briggs, C., Fan, Z., and Andras, P. (2020). Federated learning with hierarchical clustering of local updates to improve training on non-IID data. 2020 international joint conference on neural networks (IJCNN). IEEE, pp.1–9.

Google Scholar

Chai, L., Xie, J., and Zhou, N. (2025). Prototype-based fine-tuning for mitigating data heterogeneity in federated learning. Futur. Gener. Comput. Syst. 170:107831. doi: 10.1016/j.future.2025.107831

Crossref Full Text | Google Scholar

Collins, L., Hassani, H., Mokhtari, A., and Shakkottai, S. (2021). Exploiting shared representations for personalized federated learning. International conference on machine learning. PMLR, Proc. ICML,virtual Event, 39:2089–2099.

Google Scholar

Deng, Y., Kamani, M. M., and Mahdavi, M., (2003). Adaptive personalized federated learning. arXiv, doi: 10.48550/arXiv.2003.13461,2020, pp.1–50preprint (arXiv:2003.13461).

Crossref Full Text | Google Scholar

Dinh, C. T., Tran, N., and Nguyen, J. (2020). Personalized federated learning with moreau envelopes. Adv. Neural Inf. Proces. Syst. 33, 21394–21405.

Google Scholar

Ghosh, A., Chung, J., Yin, D., and Ramachandran, K. (2020). An efficient framework for clustered federated learning. Adv. Neural Inf. Process. Syst. 33, 19586–19597. doi: 10.48550/arXiv.2006.04088

Crossref Full Text | Google Scholar

Huang, Y., Chu, L., Zhou, Z., Wang, L., Liu, J., Pei, J., et al. (2021). Personalized cross-silo federated learning on non-IID data. Vancouver, British Columbia, Canada: Proceedings of the AAAI conference on artificial intelligence, 35:7865–7873.

Google Scholar

Islam, M.S., Javaherian, S., Xu, F., Yuan, X., Chen, L., and Tzeng, N.F. (2024). FedClust: tackling data heterogeneity in federated learning through weight-driven client clustering. Proceedings of the 53rd international conference on parallel processing, pp.474–483.

Google Scholar

Li, T., Hu, S., Beirami, A., and Smith, V. (2021). Ditto: fair and robust federated learning through personalization. International conference on machine learning. PMLR. Available online at: https://arxiv.org/pdf/2012.04221.pdf, pp.6357–6368 (Accessed August 12-15, 2024).

Google Scholar

Liang, P. P., Liu, T., Ziyin, L., Allen, N. B., Auerbach, R. P., Brent, D., et al. (2022). Think locally, act globally: federated learning with local and global representations. Workshop on Federated Learning for Data Privacy and Confidentiality, NeurIPS 2019. 1–34. doi: 10.48550/arXiv.2001.01523

Crossref Full Text | Google Scholar

Liu, Y., Chang, S., Li, D., Shi, S., and Li, B. (2025). RoPe-door: toward robust and persistent backdoor data poisoning attacks in federated learning. IEEE Netw. 39, 302–310. doi: 10.1109/MNET.2024.3486228

Crossref Full Text | Google Scholar

McMahan, B., Moore, E., and Ramage, D. (2017) Communication-efficient learning of deep networks from decentralized data. Proceedings of the 20th international conference on artificial intelligence and statistics (AISTATS), Fort Lauderdale, Florida, USA, PMLR:W&CP, 54:1273–1282

Google Scholar

Pang, M., Wang, B., Ye, M., Cheung, Y. M., Zhou, Y., Huang, W., et al. (2025a). Heterogeneous prototype learning from contaminated faces across domains via disentangling latent factors. IEEE Trans. Neural Netw. Learn. Syst. 36, 7169–7183. doi: 10.1109/TNNLS.2024.3393072,

PubMed Abstract | Crossref Full Text | Google Scholar

Pang, M., Zhang, W., Lu, Y., Cheung, Y. –. M., and Zhou, N. (2025b). A unified multi-domain face normalization framework for cross-domain prototype learning and heterogeneous face recognition. IEEE Trans. Inf. Forensics Secur. 20, 5282–5295. doi: 10.1109/TIFS.2025.3570121

Crossref Full Text | Google Scholar

Sattler, F., Müller, K. R., and Samek, W. (2020). Clustered federated learning: model-agnostic distributed multitask optimization under privacy constraints. IEEE Trans. Neural Netw. Learn. Syst. 32, 3710–3722.

Google Scholar

Sattler, F., Wiedemann, S., Müller, K. -R., Samek, W., et al. (2019). Robust and communication-efficient federated learning from non-IID data. IEEE Trans. Neural Netw. Learn. Syst. 31, 3400–3413.

Google Scholar

Tan, A. Z., Yu, H., Cui, L., and Yang, Q. (2022). Towards personalized federated learning. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–17. doi: 10.1109/TNNLS.2022.3160699

Crossref Full Text | Google Scholar

Vahidian, S., Morafah, M., and Wang, W. (2023) Efficient distribution similarity identification in clustered federated learning via principal angles between client data subspaces. Walter E. eds., Proceedings of the AAAI conference on artificial intelligence, Washington, DC, USA: Washington Convention Center. pp.10043–10052. doi: 10.1609/aaai.v37i8.26197

Crossref Full Text | Google Scholar

Wang, F., Feng, J., Zhao, Y., Zhang, X., Zhang, S., and Han, J. (2019). Joint activity recognition and indoor localization with WiFi fingerprints. IEEE Access 7, 80058–80068. doi: 10.1109/access.2019.2923743

Crossref Full Text | Google Scholar

Zhang, M., Sapra, K., Fidler, S., Yeung, S., and Alvarez, J. M. (2020). Personalized federated learning with first order model optimization. arxiv, 1, 1–17. doi: 10.48550/arXiv.2012.08565 (preprint arxiv:2012.08565, ICLR)

Crossref Full Text | Google Scholar

Zhang, Y., Zheng, Y., Qian, K., et al. (2021). Widar3.0: zero-effort cross-domain gesture recognition with Wi-Fi. IEEE transactions on pattern analysis and machine intelligence, pp.1–18.

Google Scholar

Zhao, Y., Li, M., Suda, N., Civin, D., and Chandra, V. (2018). Federated learning with non-IID data. arXiv, abs/1806.00582, 1–12. doi: 10.48550/arXiv.1806.00582 (preprint, Computer Science, arXiv:1806.00582v2)

Crossref Full Text | Google Scholar

Keywords: data heterogeneity, federated learning algorithm, KL divergence, personalized federated learning algorithm, wireless sensing

Citation: Tian Z and Tian J (2026) A new clustered federated learning algorithm for heterogeneous data in high-precision wireless sensing. Front. Artif. Intell. 9:1718193. doi: 10.3389/frai.2026.1718193

Received: 03 October 2025; Revised: 27 November 2025; Accepted: 14 January 2026;
Published: 04 February 2026.

Edited by:

Jinjia Zhou, Hosei University, Japan

Reviewed by:

Nanrun Zhou, Shanghai University of Engineering Sciences, China
Shahzad Ashraf, Gachon University, Republic of Korea
Jinyang Huang, Hefei University of Technology, China
Prasan Yapa, Kyoto University of Advanced Science, Japan

Copyright © 2026 Tian and Tian. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jiasheng Tian, dGlhbmpzQGh1c3QuZWR1LmNu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.