ORIGINAL RESEARCH article

Front. Big Data, 09 February 2026

Sec. Data Mining and Management

Volume 9 - 2026 | https://doi.org/10.3389/fdata.2026.1782461

A genetic algorithm-based framework for online sparse feature selection in data streams

  • GL

    Guanyu Liu 1,2

  • JL

    Jinhang Liu 1

  • GH

    Guifan He 1

  • YL

    Yifan Liu 1

  • HB

    Huabo Bai 1

  • MZ

    Min Zhou 3*

  • 1. College of Computer and Information Science, Southwest University, Chongqing, China

  • 2. PetroChina Qinghai Oilfield Company, Qinghai, China

  • 3. Office of Informatization Construction, Southwest University, Chongqing, China

Article metrics

View details

425

Views

87

Downloads

Abstract

High-dimensional streaming data implementations commonly utilize online streaming feature selection (OSFS) techniques. In practice, however, incomplete data due to equipment failures and technical constraints often poses a significant challenge. Online Sparse Streaming Feature Selection (OS2FS) tackles this issue by performing missing data imputation via latent factor analysis. Nevertheless, existing OS2FS approaches exhibit considerable limitations in feature evaluation, resulting in degraded performance. To address these shortcomings, this paper introduces a novel genetic algorithm-based online sparse streaming feature selection (GA-OS2FS) in data streams, which integrates two key innovations: (1) imputation of missing values using a latent factor analysis model, and (2) application of genetic algorithm to assess feature importance. Comprehensive experiments conducted on six real-world datasets show that GA-OS2FS surpasses state-of-the-art OSFS and OS2FS methods, consistently attaining higher accuracy through the selection of optimal feature subsets.

1 Introduction

The rapid advancement of information technology has led to the widespread generation of high-dimensional data characterized by multiple levels, granularities, and modalities. This complexity poses significant challenges to foundational technologies in fields such as artificial intelligence, data management, communication, and storage (Yao et al., 2022; Ramírez-Gallego et al., 2018; Li Z. et al., 2025). To address the issues associated with high-dimensional data, feature selection has proven to be a highly effective technique (Chandrashekar and Sahin, 2014; Casmiry et al., 2025; Wang et al., 2025; Chen et al., 2025). In recent years, a diverse array of feature selection methodologies has emerged (Kundu and Mitra, 2017; Yang et al., 2018; Albattah and Khan, 2025), which can be broadly categorized into filter-based, wrapper-based (Xue et al., 2018), and embedded approaches (Xue et al., 2016). Furthermore, in the context of big data applications, the feature space frequently expands dynamically, potentially to an infinite scale (Ni et al., 2017; Ditzler et al., 2018). This reality has driven the development of Online Streaming Feature Selection (OSFS). For example, Wu et al. (2013) pioneered an OSFS framework utilizing online relevance and redundancy analysis. Their model classifies incoming features into strongly relevant, weakly relevant, and irrelevant groups, ultimately selecting features that are relevant (strongly or weakly) and non-redundant. Subsequently, Yu et al. (2016) introduced the SAOLA model, which extends this concept by evaluating the pairwise relationships between streaming features through a specific mechanism.

However, most existing Online Streaming Feature Selection (OSFS) models are formulated under the assumption of complete feature streams, where all incoming data points are fully observed without any missing values. In real-world scenarios, this assumption often fails to hold, as streaming features frequently contain substantial missing data due to a range of unforeseen factors. For instance, in single-cell sequencing, technological constraints make it challenging to profile every cell comprehensively, preventing reliable weight assignment for all measured entities (Badsha et al., 2020). Similarly, in clinical settings, complete patient data collection is often hindered by equipment failures or procedural inconsistencies (Idri et al., 2018). This prevalent issue gives rise to the challenge of Online Sparse Streaming Feature Selection (OS2FS), which addresses the critical question of how to reliably select features from a stream that is inherently sparse and contains significant missing entries.

In real-world recommendation systems, features—including user behavior logs and product attributes–are often received as continuous streams. Since users typically interact with only a fraction of available items, missing data is commonplace. These features are also highly interdependent, complicating the decision of which should be retained or removed to fully and precisely model user interests. Consequently, identifying the most representative feature subset is essential for delivering prompt and relevant recommendations. Traditional approaches to feature evaluation are largely designed for fully observed feature streams and tend to overlook the inaccuracies arising from the imputation of missing data. Such neglect is particularly consequential, as feature selection constitutes an NP-hard binary discrete optimization problem. Evolutionary computation (EC) techniques are notably effective in overcoming these challenges, providing robust solutions for problems of high combinatorial complexity. The core advantages of genetic algorithms lie in their global exploration capability, low dependency on problem assumptions, coding flexibility, and ease of parallelization, making them particularly suitable for complex optimization problems. Furthermore, its strong global search ability helps prevent convergence to local optima, increasing the likelihood of discovering feature subsets that optimize the trade-off between model accuracy and feature sparsity. These advantages have led to the widespread adoption of GA-based strategies in feature subset selection tasks. Therefore, this paper proposes a novel genetic algorithm-based online sparse streaming feature selection (GA-OS2FS) in data streams. In smart factories, GA-OS2FS processes incomplete, high-dimensional streaming data from sensors (e.g., vibration, temperature) by imputing missing values via latent factor analysis and dynamically selecting the most discriminative features (e.g., failure-indicative patterns) using a genetic algorithm. This enables real-time, accurate anomaly detection and predictive maintenance, minimizing unplanned downtime. For IoT-based energy management systems, GA-OS2FS handles sparse and irregular streaming data from distributed sensors (e.g., occupancy, temperature). It recovers missing values and employs genetic algorithm-based evaluation to identify and retain features most relevant to energy consumption. This results in an optimized feature subset for real-time control of HVAC and lighting systems, enhancing energy efficiency in smart buildings.

2 Related work

Online Streaming Feature Selection (OSFS) models, which process feature streams in real time, have garnered significant research interest. For instance, Perkins and Theiler (2003) introduced Grafting, a regularized online feature selection framework. However, it requires careful tuning of regularization parameters prior to feature selection, making it less adaptable to scenarios with an unknown or expanding feature space. Zhou et al. (2006) proposed the Alpha-investing strategy, capable of handling infinite feature streams, though it does not account for redundancy among the selected features. Wu et al. (2013) categorized incoming features into strongly relevant, weakly relevant, and irrelevant groups, developing two OSFS variants: OSFS and Fast-OSFS. The latter specifically addresses redundancy between newly arrived features and the already selected subset. Building on mutual information, Yu et al. (2016) presented the SAOLA model, which evaluates feature relevance based on pairwise interactions. To capture more complex dependencies, Zhou et al. (2021b) developed the OGSFS-FI model by examining interactions between feature groups. This was later extended to the SFS-FI model Zhou et al. (2021a), which can identify features involved in multi-way interactions, including two-way, three-way, and higher-order relationships. Furthermore, to better model dynamic decision-making, Zhou et al. (2022) applied the three-way decision (3WD) principle to propose the OSSFS-DD model, which computes partition thresholds according to 3WD theory to mitigate decision risk.

In parallel, rough set theory has proven to be a valuable framework for Online Streaming Feature Selection (OSFS). For example, Zhou et al. (2019b) introduced the OFS-A3M model, which employs a neighborhood rough set relation with adaptive neighbors to identify features that exhibit high relevance, strong dependency, and low redundancy. This work was later extended to the OFS-Density model (Zhou et al., 2019a), where a novel adaptive density-based neighborhood relation is used to analyze domain characteristics and configure model parameters. In a different approach, Luo et al. (2023) leveraged the concept of rough hypercuboids to develop the RHDOFS model. Similarly, Shu et al. (2024) proposed the ANOHFS model, which relies on an adaptive neighborhood mechanism to effectively identify closely related feature hierarchies within high-dimensional data. Zhuo et al. (2024) proposed an online feature selection method for dynamic feature spaces, with innovations in Gaussian Copula-based correlation modeling, real-time tree-ensemble selection, and geometric inference for unlabeled data. Qiu et al. (2025) proposed an online confidence learning algorithm for noisy labeled features. It tackles instance distribution shifts and label noise in data streams by employing online confidence inference and geometric structure learning. Although current OSFS models play a crucial role in dynamically selecting streaming features, to our knowledge, they still lack the ability to effectively handle sparse streaming features. Missing data tends to raise the computational cost of OSFS models and may also lead to the selection of less relevant or redundant features. Sparse streaming features often exhibit weak associations with other features or the target variable, complicating the reliable evaluation of their importance. Moreover, they can cause uneven data distributions, where certain sample values appear very rarely–a situation that may undermine the overall performance of OSFS. While these methods demonstrate considerable effectiveness in tackling conventional OSFS problems, they share a common limitation: all are designed under the assumption of complete feature streams and do not account for missing data, thus leaving the challenges of OS2FS scenarios unaddressed.

Latent factor analysis (LFA) has established itself as an effective approach for estimating missing data (Wu et al., 2022). The method operates by mapping the observed entries of a high-dimensional, incomplete matrix onto latent representations associated with its rows and columns (Zhang Z. et al., 2017; Zhang J. D. et al., 2017). A learning objective is formulated to quantify the discrepancy between the original observed values and their reconstructions (Luo et al., 2018; Gong et al., 2018). Subsequently, the model constructs a complete, low-rank approximation of the target incomplete matrix by minimizing this generalized error, as defined by the learning objective (Luo et al., 2021b; Wu et al., 2021).

3 Preliminaries

3.1 Online streaming feature selection

The Online Streaming Feature Selection (OSFS) model provides an effective approach for identifying the optimal subset of streaming features, which is accomplished through online relevance analysis and online redundancy analysis. Consider a streaming feature set F = {F1, F2, ..., FT} and a label set , where each feature contains M samples, with t ∈ 1, 2, ..., T.

Suppose that two features Fp and Fq, where pq, p, q ∈ 1, 2, ..., T, if P(Fp|Fq, X) = P(Fp, X) or P(Fp|Fq, X) = P(Fq, X), such that Fp and Fq are conditionally independent to the subset XF.

For a streaming feature Ft at the time stamp t,

  • a) if ∀ς ⊆ FFt s.t. P(C|ς, Ft) ≠ P(C|ς), then decide Ft is strong relevance;

  • b) if ∃ς ⊆ FFt s.t. P(C|ς, Ft) ≠ P(C|ς), then decide Ft is weak relevance;

  • c) if ∀ς ⊆ FFt s.t. P(C|ς, Ft) = P(C|ς), then decide Ft is irrelevance.

Given a relevant feature Ft(M(Ft) ∉ M(C)t), and the redundant set XF is denoted as follows:

where M(·) denotes Markov blanket.

3.2 Latent factor analysis

The Latent Factor Analysis (LFA) model plays a significant role in pre-estimating sparse matrices. This section begins by presenting the formal definition of the LFA model (Hancer et al., 2022; Wang et al., 2023).

Let RM × H be a sparse matrix, and an LFA model trains two latent factor matrices UM × L and VH × L via the know entries, which precisely represent the rank-L approximation of R, where is formulated as = UVT, L is the dimension of U and V, and L ≪ min|M|, |H| (Li et al., 2024; Hancer et al., 2025).

The error of the LFA model is then formulated as:

where Λ denotes the known data of R, e(·) calculates the error between rm, h and , rm, h is m-th row and h-th column of R, the is the predicted value for rm, h, stands for predictive function.

Incorporating regularization is essential for the LFA model to prevent over-fitting (Wu et al., 2023; Li et al., 2023). Thus, by integrating regularization into Equation 2, the following objective function is derived:

where |·|F computes the Frobenius norm, λ represents the regularization coefficient.

4 Proposed algorithm

4.1 Problem of GA-OS2FS

Consider a collection of sparse streaming features denoted by , which is postulated to possess a missing data rate of ρ. Here, ρ = 1 − |Λ|/M, with |·| representing the cardinality of a set. From time point t to t + H-1, sparse streaming features Ft, Ft + 1, ..., Ft + H − 1 are generated sequentially and collected into an RM × H buffer. This buffer, of size H, forms the sparse streaming feature matrix . Subsequently, a completed streaming feature matrix, expressed as , is estimated based on the observed known data.

The principal objective of the GA-OS2FS method is to identify the optimal feature subset. Consequently, the GA-OS2FS framework is designed to address the following optimization problem:

4.2 The framework of GA-OS2FS

The framework of GA-OS2FO consists of two steps: first, estimating missing values, and then assessing feature importance.

4.2.1 Estimate sparse streaming features in advance

In practical applications, data quality is often difficult to guarantee due to missing feature values, making it exceptionally challenging to screen high-quality features from feature streams. Taking a medical monitoring system as an example, if a sensor fails and causes data loss, traditional OSFS models may transmit erroneous signals to control devices, which could ultimately endanger patients' lives. Therefore, preprocessing the data before feature selection to impute missing entries is of crucial importance. The LFA model holds significant value for missing-data imputation, as it completes missing values by mapping the sparse matrix onto two latent factor matrices (Li J. et al., 2025; Chen J. et al., 2024; Yuan et al., 2025). Traditional methods—such as mean imputation and matrix factorization—typically fill missing values based on observed data and rely on assumptions such as linearity or local similarity, which limits their ability to capture complex non-linear relationships in high-dimensional or sparse streaming data (Chen et al., 2023; Xu et al., 2025b). In contrast, the LFA model can capture the underlying structure of the data through latent space modeling, thereby handling complex dependencies and non-linear patterns more effectively (Liao et al., 2025; Lin M. et al., 2025).

The complete latent features extracted from incomplete data can be used for missing-value imputation, classification, clustering, and other tasks (Wu et al., 2025b; Lyu et al., 2026). The extraction methods are mainly divided into linear and non-linear feature extraction. Linear feature extraction mostly employs LFA-based models that rely on matrix factorization. When dealing with sparse data, such models aim to construct a low-rank approximation of the high-dimensional incomplete matrix (Wu et al., 2024; Wu H. et al., 2025). They map the known entries of the target high-dimensional incomplete matrix to its row and column nodes, formulate a learning objective that measures the discrepancy between the actual data and the estimated data, and thereby generate a complete low-rank approximation matrix of the target high-dimensional incomplete matrix. An optimizer is then used to minimize linear error, achieving efficient representation (Wu et al., 2025a; Qin et al., 2024; Xu et al., 2025a; Chen M. et al., 2024).

The initialization procedure subjects both matrices U and V to small random values. These values, generated by scaling a random permutation to a vicinity close to zero, serve as the non-zero starting point for the iterative algorithm, for which initial conditions are crucial. The following illustration details the update method, taking matrix U as a representative case.

The LFA model constructs a low-rank approximation for R (Tang et al., 2024; Luo et al., 2021a). Typically, matrices U and V are derived from R by minimizing a loss function defined by the Euclidean distance between R and (Xu et al., 2023). Building upon Equations 2, 3, the complete streaming features are predicted using the known values according to:

Subsequently, the loss function for the m-th element fm, j is calculated as:

To solve this loss function, stochastic gradient descent (SGD) is employed (Ahmadian et al., 2025; Lin X. et al., 2025; Lei et al., 2024). The method computes the gradient of the loss function with respect to the combined parameters and updates them in a descending direction:

From Equations 7 and 8, the partial derivative of the loss is derived:

Here, η denotes the learning rate. U and V are optimized to minimize errors on the known values, yielding R = UVT. The error between the estimated and actual data is , i.e.,

4.2.2 Online feature evaluation

The principal advantage of GA-OS2FS lies in its independence from missing value completion via an LFA model. The method sustains a feature subset whose fitness is evaluated through real classification accuracy. Consequently, the search process gains the capacity to tolerate and sidestep local misleading associations stemming from potential completion errors. As a prominent and widely implemented evolutionary optimization method, the GA offers several compelling advantages. GA maintains a diverse population of candidate solutions, enabling simultaneous exploration of multiple regions in the solution space. Through mechanisms such as selection, crossover, and mutation, it effectively combines and propagates beneficial gene patterns while continually introducing new variations. This population-based strategy significantly mitigates the risk of premature convergence to local optima, making GA particularly robust in navigating complex, multimodal search landscapes commonly encountered in feature selection. GA operates solely on the evaluation of candidate fitness, requiring no derivative information of the objective function. This characteristic renders it highly suitable for optimizing non-differentiable, discontinuous, or noisy objective functions—frequently the case in feature selection where the fitness is often a classification error rate or another performance metric derived from a learning model. The evolutionary process inherently promotes solutions that achieve an optimal balance between multiple, often competing, objectives. In feature selection, fitter individuals naturally tend to be those that maximize classification performance while minimizing the number of selected features, without needing an explicitly tuned regularization parameter. This emergent trade-off helps in discovering compact, discriminative feature subsets. The evaluation of fitness for each individual in a population is independent of others, making this computational step “embarrassingly parallel.” This allows for efficient distribution across multiple processors or cores, drastically reducing wall-clock time and enhancing the scalability of GA for large-scale or high-dimensional problems. Collectively, these advantages establish Genetic Algorithms as a powerful, flexible, and efficient metaheuristic framework for tackling the inherently combinatorial and complex problem of feature selection.

Given a dataset with m samples and n features, where represents the feature vector and denotes the class label, the feature selection problem aims to identify an optimal subset of features that maximizes classification performance while minimizing dimensionality. This binary optimization problem can be formulated as:

where is a binary vector with bj = 1 if feature j is selected, and bj = 0 otherwise, denotes the ℓ0-norm counting selected features, X ∈ ℝm × n is the feature matrix with Xij = xi, j, represents the classification error function, α and β are weighting coefficients balancing classification accuracy and feature sparsity, ⊙ denotes element-wise multiplication, 1m is an m-dimensional vector of ones.

Each candidate solution (chromosome) is encoded as a binary vector b ∈ 0, 1n. The initial population of size N is generated randomly:

where p0 = 0.5 ensures unbiased initial exploration of the feature space.

The fitness evaluation metric is primarily assessed by measuring the classification error achieved using the selected features. As a wrapper-based method, this approach directly employs the performance of a target classifier—such as support vector machine (SVM)—to determine the quality of a candidate feature subset. The classification error serves as a direct and interpretable indicator of how well the selected features support the learning algorithm in discriminating between classes. Typically, to ensure robustness and prevent overfitting, the error is estimated via cross-validation or hold-out validation. This design aligns the feature selection process closely with the end classification task, thereby enhancing the relevance and discriminative power of the final feature subset.

To form the mating pool for generation t, individuals are selected probabilistically based on their fitness. The selection probability for chromosome bi is:

Here, ι > 0 is a small constant preventing division by zero. The cumulative distribution function is:

For each selection, a random number r ~ U(0, 1) is generated, and chromosome bk is selected where .

With probability pc, pairs of parent chromosomes undergo single-point crossover. For parents bp and bq, a crossover point c is randomly selected:

Two offspring and are generated as:

If crossover is not applied (with probability 1 − pc), the offspring are exact copies of the parents.

Each gene in the offspring undergoes mutation with probability pm. For gene bij:

This operator maintains population diversity and enables exploration of new regions in the search space.

To guarantee monotonic improvement across generations, the algorithm employs an elitism strategy. The best chromosome from generation t is preserved by replacing the worst chromosome in the offspring population :

This ensures that:

The genetic algorithm for OS2FS begins with inputs including the feature matrix X ∈ ℝm × n, label vector y ∈ ℝm, population size N, maximum iterations Tmax, crossover probability pc (default: 0.8), mutation probability pm (default: 0.05), mutation strength μ (default: 0.01), and hold-out folds k (default: 0). Initially, it sets t = 0 and generates the population via Equation 11, then evaluates each individual by computing for i = 1, …, N, and identifies the best individual as with fitness . While t < Tmax, the algorithm performs selection to form the mating pool using roulette wheel selection (Equations 12, 13), generates offspring via single-point crossover (Equations 14, 15) with probability pc, applies bit-flip mutation (Equation 16) to with probability pm, and evaluates the offspring by computing . It updates the best individual if , then applies elitism by replacing the worst individual in with bbest(t) (Equation 17), followed by updating , f(t + 1) = f′, and the convergence curve , and incrementing t = t + 1. After the loop, it extracts the results: , , , and returns as output. Redundancy analysis is first performed on individual features using Markov, and then on the selected features using Equation 1.

The algorithm's convergence is guaranteed by the elitism strategy, which ensures the best fitness value is non-increasing:

Theorem 1 (Monotonic Convergence). For the GA-OS2FS algorithm with elitism, the sequence of best fitness values is monotonically non-increasing.

Proof. By construction, the elitism strategy preserves the best solution from generation t in generation t + 1. Therefore, for all t.

The expected time complexity per iteration is , where is the cost of evaluating the fitness function for one chromosome. The overall complexity for Tmax iterations is .

The proposed GA-OS2FS algorithm provides an effective approach for streaming feature selection, combining the global search capability of genetic algorithms with direct performance evaluation using the target classifier.

5 Experiments

5.1 General settings

5.1.1 Datasets

This section presents the experimental evaluation conducted on six real-world datasets obtained from two key sources: DNA microarray repositories and the benchmark collection from the Neural Information Processing Systems (NIPS) 2003 conference. These datasets are widely recognized in the machine learning and bioinformatics communities for assessing feature selection and classification algorithms under high-dimensional, small-sample conditions. The inclusion of microarray data ensures the examination of genetic expression patterns, while the NIPS 2003 datasets provide a diverse range of problem domains and complexity levels, thereby enabling a comprehensive analysis of the proposed method's robustness and generalizability. A detailed summary of the datasets—including the number of features, samples, and classes—is provided in Table 1 for reference.

Table 1

MarkDatasetFeaturesInstancesClass
D1USPS2421,5002
D2Madelon5012,6006
D3COIL201,0251,44020
D4Colon2,001622
D5Lung3,313835
D6DriveFace6,4016063

Details of the datasets.

5.1.2 Baselines

To comprehensively evaluate the efficacy of the proposed model, a rigorous comparative analysis is conducted against four state-of-the-art online streaming feature selection (OS2FS) methods, which are recognized as established benchmarks in the field. The selected competitors include Fast-OSFS, SAOLA, and LOS-SA. This diverse set of algorithms encompasses various strategic approaches to handling feature streams—such as leveraging pairwise feature relations, redundancy analysis, and sparsity-aware selection—thereby ensuring a robust and multifaceted comparison. Furthermore, to objectively assess the quality of the feature subsets selected by each method, the evaluation employs three fundamental yet powerful classifiers: Support Vector Machine (SVM), k-Nearest Neighbors (KNN), and Random Forest (RF). These classifiers were chosen for their distinct learning mechanisms: SVM seeks optimal separating hyperplanes, KNN relies on local similarity, and RF utilizes ensemble decision-making. Their combined use helps verify whether the selected features generalize well across different inductive biases and are not tailored to a single classification model.

Detailed parameter configurations for all compared OS2FS algorithms and the three classifiers are systematically summarized in Tables 2, 3, respectively, to ensure full reproducibility of the experiments. All algorithms are implemented in MATLAB to maintain a consistent computational environment. And all experiments utilize five-fold cross-validation, meaning each dataset is randomly divided into an 80% training portion and a complementary 20% test portion. To account for randomness in data partitioning and algorithm initialization, each dataset is executed 10 times; the final reported result is the average predictive accuracy across these runs, along with its standard deviation where applicable.

Table 2

MarkAlgorithmParameter
M1GA-OS2FSZ test, Alpha is 0.05.
M2LOSSAZ test, Alpha is 0.05. (TSMC, 2022)
M3Fast-OSFSZ test, Alpha is 0.05. (TPAMI, 2013)
M4SAOLAZ test, Alpha is 0.05. (TKDD, 2016)
M5SFS-FIZ test, Alpha is 0.05. (TNNLS, 2021)

Algorithm parameters.

Table 3

ClassifierParameter
KNNThe number of neighbors was set to 3.
Random forest6 decision trees.
CARTPredefined parameter settings.

Details of the classifiers.

All trials were conducted on a standard personal computer equipped with an Intel Core i7 processor running at 2.40 GHz and 16 GB of RAM, ensuring that the computational demands of the online feature selection and classification processes were feasibly met within a common research setup. This controlled hardware environment also aids in the fair comparison of runtime and efficiency where relevant.

5.1.3 Experimental configuration

The efficacy of the GA-OS2FS model is rigorously assessed by benchmarking it against the aforementioned suite of advanced algorithms, specifically within the challenging context of sparse streaming features. This scenario is deliberately chosen to simulate real-world conditions where data incompleteness and sequential feature arrival are prevalent, thereby testing the models' robustness and adaptability. To ensure a fair and statistically grounded comparison of performance across all algorithms, a non-parametric Friedman test is conducted at a stringent 95% confidence level. This test is employed under the null hypothesis that all algorithms perform equivalently, providing a holistic view of performance rankings across multiple datasets.

Furthermore, to drill down into pairwise performance differences, a paired Wilcoxon signed-rank test is applied at a 0.1 significance level. This test is specifically designed to examine whether the observed performance differences between the GA-OS2FS model and each individual baseline algorithm are statistically significant, rather than attributable to random chance. The resulting p-values from this comprehensive statistical analysis are consistently below the significance threshold. This robust statistical evidence leads to the conclusive finding that the GA-OS2FS model significantly and consistently outperforms all competing algorithms in the evaluation, demonstrating its superior capability in selecting informative features from sparse, evolving data streams.

5.2 Accuracy comparison

5.2.1 Detailed analysis under 10% missing data rate

To investigate the impact of missing data on feature selection performance, a missing-at-random scenario with a 10% data loss rate is established as a representative and practically relevant case for detailed analysis. This specific rate is chosen to simulate a common yet challenging level of data incompleteness encountered in real-world streaming applications. As illustrated in Figure 1, the average number of features selected by each compared method under this sparse condition is presented. The bar chart reveals distinct strategies among the algorithms: some methods maintain a conservative, highly selective profile, while others retain a larger fraction of the feature stream, reflecting different trade-offs between redundancy elimination and information preservation.

Figure 1

Correspondingly, Table 4 documents the concrete predictive performance outcomes, quantified by classification accuracy, when applying three fundamentally different classifiers—K-Nearest Neighbors (KNN), Support Vector Machine (SVM), and Random Forest (RF)—to the feature subsets identified by each method. This multi-classifier evaluation is crucial, as it demonstrates whether the selected features provide robust discriminative power independent of a specific learning algorithm's bias. The results in the table allow for a direct, quantitative comparison of how the parsimony or comprehensiveness of a selected feature subset, as shown in Figure 1, ultimately translates into generalization accuracy across diverse classifiers. This integrated analysis of subset size (Figure 1) and classification efficacy (Table 4) provides a comprehensive view of each algorithm's effectiveness in balancing feature reduction with predictive performance under the specified missing data condition.

Table 4

M/DD1D2D3D4D5D6
M190.79 ± 0.8759.14 ± 1.3692.62 ± 1.1381.77 ± 3.4889.47 ± 1.3194.91 ± 0.50
M284.33 ± 0.6454.97 ± 0.6984.87 ± 2.6380.45 ± 2.5984.79 ± 2.7792.29 ± 0.58
M385.48 ± 0.6354.83 ± 0.9771.08 ± 2.2278.88 ± 2.5984.40 ± 2.4393.14 ± 0.67
M480.18 ± 0.5453.79 ± 0.8088.59 ± 0.4977.65 ± 3.1983.38 ± 2.3293.34 ± 0.71
M572.18 ± 0.5449.33 ± 0.5480.56 ± 3.4878.63 ± 2.6662.49 ± 2.9685.51 ± 1.02

The accuracy when the missing rate is 0.1.

5.2.1.1 Statistical significance (Friedman test)

To statistically validate the performance differences observed among the compared algorithms under the 10% missing data scenario, a non-parametric Friedman test was conducted across all datasets. The test returned a P-value of 0.0011, which is substantially below the commonly adopted significance threshold of 0.05. This very small P-value allows us to firmly reject the null hypothesis that all algorithms perform equally. Therefore, the results provide strong statistical evidence that there are significant differences in the overall performance ranks of the evaluated methods. More specifically, the outcome underscores that the proposed GA-OS2FS model achieves a distinctly superior ranking compared to the alternative algorithms, confirming its enhanced robustness and effectiveness when handling incomplete streaming features with 10% missing values. Such a statistically significant finding further reinforces the practical relevance and reliability of the GA-OS2FS approach in real-world sparse data environments.

5.2.1.2 Analysis of feature selection quantity

The GA-OS2FS model demonstrates stable feature selection across different sparse datasets. In contrast, algorithms like SAOLA show considerable variation in the number of features selected depending on the dataset. Key observations include:

  • An intriguing pattern observed in the experiments is that several compared algorithms tend to select a considerably larger set of features, yet consistently deliver lower classification accuracy compared to the GA-OS2FS model. This indicates that simply retaining more features does not guarantee better predictive performance, and often points to insufficient or less effective redundancy analysis in the feature selection process. When redundancy is not adequately assessed, many retained features may be non-informative, noisy, or highly correlated with one another, thereby adding little discriminative value while increasing model complexity and the risk of overfitting. In contrast, the GA-OS2FS model appears to implement a more refined mechanism for evaluating feature relevance and redundancy, enabling it to identify and retain a compact yet highly informative subset of features that better supports accurate classification.

  • Other algorithms, such as SFS-FI, occasionally select an extremely small number of features—in some cases as few as only one—on particular datasets. This behavior is likely attributable to their limited ability to comprehensively capture all essential features when processing incomplete data streams. Specifically, these methods may prematurely converge on the first few features that appear sufficiently relevant, while failing to adequately evaluate or retain subsequently arriving features that are equally or more informative. As a result, they miss critical feature interactions and discard valuable discriminative information, ultimately leading to suboptimal classification performance due to an oversimplified and incomplete feature subset.

  • The GA-OS2FOS model performs comprehensive relevance and redundancy analysis through a structured genetic optimization process. By leveraging GA-based feature evaluation, it dynamically assesses each feature's discriminative power and mutual dependencies within the evolving stream. This enables the model to systematically identify and retain truly informative features while filtering out redundant or noisy ones. Consequently, it avoids the premature discarding of important predictive information—a common pitfall in many streaming feature selection methods. As a result, the model consistently achieves high classification accuracy while maintaining a compact and efficient feature subset, effectively balancing model simplicity with representational completeness.

5.2.1.3 Classification performance

As shown in Table 4, the GA-OS2FS model exceeds the performance of its rivals on six datasets. Key observations include:

  • GA-OS2FS vs. Fast-OSFS: the experimental results demonstrate that the GA-OS2FOS model consistently delivers superior classification accuracy across a majority of the benchmark datasets. In contrast, the Fast-OSFS algorithm exhibits notable limitations. Its performance is constrained by a reliance on zero-imputation to handle incomplete data—a method that simply fills missing values with zeros. While straightforward, this approach fails to capture any underlying data structure or relationships, potentially distorting the feature space. Furthermore, Fast-OSFS employs a less comprehensive analysis of feature relevance and redundancy. This dual shortcoming—crude data imputation coupled with insufficient feature evaluation—often results in the misclassification of features during the streaming selection process. Informative features may be incorrectly discarded, while redundant or noisy ones might be retained. Consequently, these limitations fundamentally undermine the quality of the final selected feature subset, leading to its comparatively poorer predictive performance.

  • GA-OS2FS vs. SAOLA: the SAOLA algorithm operates primarily by assessing pairwise relationships between features, evaluating them in isolation or through limited local comparisons. While efficient, this approach may overlook more complex, higher-order interactions among feature subsets, and its incremental update mechanism can be sensitive to the arrival order of features in a stream. In contrast, the proposed GA-OS2FS model integrates and fully leverages the complementary strengths of the LFA model and the GA framework. LFA assists in capturing underlying low-rank structures and global correlations even under sparse or missing data conditions, while GA performs robust, population-based search to dynamically evaluate and retain the most discriminative feature combinations. This hybrid strategy enables GA-OS2FS to consistently identify critical features in real-time from evolving data streams, without being constrained by purely local or pairwise evaluations.

  • GA-OS2FS vs. SFS-FI: sparse streaming data often loses critical feature interactions, which severely challenges methods like SFS-FI that rely on detecting these dependencies. Unable to accurately assess feature relevance under sparsity, SFS-FI tends to select redundant or omit informative features, resulting in the lowest classification accuracy in evaluations. This underscores its limited robustness with incomplete data and highlights the advantage of GA-OS2FS's more resilient design.

  • Among the evaluated models, LOSSA achieves the second-highest classification accuracy after GA-OS2FS when processing completed sparse streaming features, demonstrating the benefit of using Latent Factor Analysis (LFA) for data completion. However, LOSSA relies on conventional relevance and redundancy analyses, which lack adaptability to capture complex feature interactions or evolving stream characteristics, limiting its average accuracy. In contrast, GA-OS2FS integrates a Genetic Algorithm strategy, performing a global, population-based search that evaluates multiple feature subsets and iteratively refines them using crossover, mutation, and fitness feedback. This enables GA-OS2FS to discover more discriminative feature combinations, leading to superior predictive performance and offering a more adaptive solution for accurate feature selection in sparse streaming environments.

5.2.1.4 The Wilcoxon signed-ranks test

To rigorously substantiate the statistically significant superiority of the proposed GA-OS2FS algorithm, a non-parametric Wilcoxon signed-rank test was employed. This test was specifically chosen for its appropriateness in comparing the performance of two related samples—in this case, the paired average classification accuracy values of the GA-OS2FS model against each of the benchmark methods across multiple datasets. The detailed outcomes of these pairwise comparisons, including the calculated test statistics and corresponding P-values, are comprehensively presented in Table 5.

Table 5

M1 vs. OthersR+aR-aP-valuesb
M22100.0156
M32100.0156
M42100.0156
M52100.0156

The rank sum of the Wilcoxon signed-ranks.

aA larger value denotes a higher accuracy.

bThere is no significant difference when P-values ∈ [0.1, 0.9] at the 0.1 significance level.

The statistical analysis yields a clear and robust conclusion: even at a missing data rate of 0.1—representing a modest yet realistic level of data incompleteness–the GA-OS2FS approach demonstrates a consistent and statistically significant performance advantage. It reliably outperforms the alternative algorithms on a substantial majority of the evaluated datasets. This early and significant lead established by GA-OS2FS under sparse conditions highlights its inherent robustness and effective design for handling incomplete data streams from the outset.

In summary, relative to traditional OS2FS models, completing sparse streaming features via the LFA model generally minimizes information loss and enhances overall results. Consequently, both GA-OS2FS and LOSSA deliver the strongest performance on sparse streaming data. Nevertheless, the feature subsets selected by GA-OS2FS yield higher classification accuracy than those from LOSSA, demonstrating that GA can improve the accuracy of feature selection.

5.2.2 Accuracy analysis with higher missing rates

This study evaluates the effectiveness of the 3WDO model by comparing it against six prominent OS2FS models—Fast-OSFS, SAOLA, SFS-FI, and LOSSA—across six datasets under missing data rates ranging from 0.5 to 0.9. While LOSSA is designed to handle missing values, the other three baseline algorithms are oriented toward complete feature streams. To adapt them for sparse data, zero-imputation is applied to fill missing entries for Fast-OSFS, SAOLA, and SFS-FI. Results are highlighted where any algorithm demonstrates superior performance compared to the others. Table 6 provides a pairwise comparison between GA-OS2FS and each baseline using the Wilcoxon signed-ranks test. The average accuracy trends of all models on datasets D1-D4 are visualized in Figure 2.

Table 6

ρM2M3M4M5
R+aR-aR+aR-aR+aR-aR+aR-a
0.5210210210210
0.9210160160201

The rank sum of the the Wilcoxon signed-rank test on OSFS and OS2FS models.

adenotes missing data rate.

Figure 2

5.2.2.1 Overall accuracy of the GA-OS2FS model

Across all benchmark datasets examined, the average classification accuracy of the GA-OS2FS model demonstrates a gradual yet consistent decline as the missing data rate increases. This overall trend aligns with expectations, as higher rates of missing entries inevitably compromise the informational integrity of the feature stream, making it more challenging to reliably identify and retain discriminative features. Notably, however, on several specific datasets, the model's accuracy exhibits only minor fluctuations—remaining relatively stable even as the missing rate rises. This suggests that the GA-OS2FS approach maintains a notable degree of robustness in certain data environments, likely due to its effective integration of latent factor completion and evolutionary search, which together help preserve critical predictive information under moderate to high sparsity conditions.

5.2.2.2 Wilcoxon signed-rank test results

Table 6 presents the Wilcoxon signed-rank test results comparing the average accuracy of GA-OS2FS against other methods. The findings indicate that as the missing data rate increases from 0.5 to 0.9, the proposed algorithm outperforms most baseline methods on the majority of datasets.

5.2.2.3 Performance on datasets

Observations from Figure 2 lead to the following conclusions:

  • For the majority of the evaluated algorithms, classification accuracy exhibits a progressive decline as the rate of missing data increases. This decline can be attributed to the growing incompleteness of the feature stream, which hinders the reliable assessment of feature relevance and redundancy. In contrast, the proposed GA-OS2FS model consistently achieves superior accuracy across most benchmark datasets, maintaining a clear performance advantage even as the missing data rate escalates from 0.5 to 0.9. This robustness stems from its integrated use of latent factor analysis (LFA) for structured data completion and genetic algorithm (GA)-guided feature optimization, which together preserve discriminative information and adaptively select informative features under sparse conditions. By comparison, conventional methods such as Fast-OSFS, SAOLA, and SFS-FI rely primarily on zero-filling (zero-imputation) to handle incomplete streaming features. While computationally simple, this approach substitutes missing entries with zeros—a strategy that distorts the original data distribution, disrupts inherent feature correlations, and often introduces artificial noise. Consequently, these methods are prone to selecting uninformative or redundant features, which undermines their classification performance and explains their significantly poorer results relative to the GA-OS2FS framework, especially under higher missing-rate scenarios.

  • For missing rates between 0.1 and 0.5, LOSSA generally achieves higher accuracy than baselines like Fast-OSFS, SAOLA, and SFS-FI, due to its LFA-based data completion providing better estimates than simple imputation (e.g., zero-filling). However, as missing data increases, limited known entries raise LFA's estimation error. This distorts the recovered feature space, causing relevant features to be misclassified as irrelevant and discarded, degrading selection quality. To address this, GA-OS2FS employs a genetic algorithm to partition and evaluate features more robustly. This enables a global, resilient importance assessment that is less sensitive to local completion errors. By reducing feature misclassification, it retains a more discriminative subset, yielding consistently higher accuracy than LOSSA and other baselines, especially as sparsity grows.

In summary, by pre-estimating missing data via the LFA and GA model, the GA-OS2FS model enhances the accuracy of traditional OS2FS approaches.

6 Conclusions

This study introduces GA-OS2FS, a novel uncertainty-aware framework for Online Sparse Streaming Feature Selection, designed to address critical shortcomings in conventional approaches. The framework innovatively integrates Genetic Algorithms (GA) to navigate the complex search space of dynamic feature subsets. GA-OS2FS operates through a synergistic two-component architecture: firstly, a Latent Factor Analysis (LFA) model that performs robust, dynamic imputation and reconstruction of inherently sparse and incomplete data matrices in real-time; secondly, a GA-based optimization mechanism that drives an intelligent, evolutionary search for discriminative features, effectively evaluating feature importance and interactions under uncertainty. Extensive empirical evaluation conducted across 10 diverse real-world datasets—spanning various domains and data characteristics—demonstrates that GA-OS2FS consistently surpasses state-of-the-art OSFS and OS2FS benchmarks. It achieves superior performance not only in selection accuracy and robustness but also in maintaining operational stability, all while ensuring computational efficiency. These results collectively underscore the framework's strong potential and adaptability for reliable, real-time feature selection in challenging high-dimensional streaming data environments.

Looking ahead, future research will concentrate on advancing the theory and practice of feature quality assessment within non-stationary streaming contexts. A primary direction involves refining and extending the evolutionary computation core, leveraging advanced Genetic Algorithm strategies and other meta-heuristics to develop more adaptive feature evaluation criteria and dynamic fitness functions. These innovations will be specifically tailored to track and respond to shifting data distributions. Furthermore, we will investigate efficient, dedicated techniques to manage concept drift, such as sophisticated incremental model update protocols and adaptive sliding window mechanisms. Additional promising avenues include exploring ensemble-based feature selection tactics that combine multiple selectors, and developing dynamic feature weighting schemes to continuously prioritize the most relevant features. The overarching goal of these endeavors is to significantly enhance the framework's responsiveness, resilience, and scalability when confronted with the evolving patterns of complex, real-world streaming applications.

Statements

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

GL: Methodology, Software, Writing – original draft. JL: Validation, Writing – review & editing. GH: Visualization, Writing – review & editing. YL: Investigation, Writing – review & editing. HB: Writing – review & editing, Data curation. MZ: Writing – review & editing, Resources.

Funding

The author(s) declared that financial support was received for this work and/or its publication. This work was supported by the Key Project of Chongqing Technology Innovation and Application Development (No. CSTB2023TIAD-KPX0037, No. CSTB2025TIAD-KPX0027).

Conflict of interest

GL was employed by PetroChina Qinghai Oilfield Company.

The remaining author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was used in the creation of this manuscript. AI-assisted tools were employed only for post-writing language polishing (grammar and style). The author(s) are solely responsible for the research content, accuracy, and integrity of this work.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1

    AhmadianS.BerahmandK.RostamiM.ForouzandehS.MoradiP.JaliliM. (2025). Recommender systems based on nonnegative matrix factorization: a survey. IEEE Trans. Artif. Intell. 6, 25542574. doi: 10.1109/TAI.2025.3559053

  • 2

    AlbattahW.KhanR. U. (2025). Impact of imbalanced features on large datasets. Front. Big Data8:1455442. doi: 10.3389/fdata.2025.1455442

  • 3

    BadshaM. B.LiR.LiuB. X.LiY. I.XianM.BanovichN. E.et al. (2020). Imputation of single-cell gene expression with an autoencoder neural network. Quant. Biol. 8, 7894. doi: 10.1007/s40484-019-0192-7

  • 4

    CasmiryE.MdumaN.SindeR. (2025). Enhanced SQL injection detection using chi-square feature selection and machine learning classifiers. Front. Big Data8:1686479. doi: 10.3389/fdata.2025.1686479

  • 5

    ChandrashekarG.SahinF. (2014). A survey on feature selection methods. Comput. Electr. Eng. 40, 1628. doi: 10.1016/j.compeleceng.2013.11.024

  • 6

    ChenJ.LiuK.LuoX.YuanY.SedraouiK.Al-TurkiY.et al. (2024). A state-migration particle swarm optimizer for adaptive latent factor analysis of high-dimensional and incomplete data. IEEE/CAA J. Autom. Sin. 11, 22202235. doi: 10.1109/JAS.2024.124575

  • 7

    ChenJ.WangR.WuD.LuoX. (2023). A differential evolution-enhanced position-transitional approach to latent factor analysis. IEEE Trans. Emerg. Top. Comput. Intell. 7, 389401. doi: 10.1109/TETCI.2022.3186673

  • 8

    ChenJ.ZhuoS.HeJ.QiuW.ZhangQ.XiongZ.et al. (2025). Federated graph learning via constructing and sharing feature spaces for cross-domain IOT. IEEE Internet Things J.12, 2620026214. doi: 10.1109/JIOT.2025.3560635

  • 9

    ChenM.WangR.QiaoY.LuoX. (2024). A generalized Nesterov's accelerated gradient-incorporated non-negative latent-factorization-of-tensors model for efficient representation to dynamic QoS data. IEEE Trans. Emerg. Top. Comput. Intell. 8, 23862400. doi: 10.1109/TETCI.2024.3360338

  • 10

    DitzlerG.LaBarckJ.RitchieJ.RosenG.PolikarR. (2018). Extensions to online feature selection using bagging and boosting. IEEE Trans. Neural Netw. Learn. Syst. 29, 45044509. doi: 10.1109/TNNLS.2017.2746107

  • 11

    GongM.JiangX.LiH.TanK. C. (2018). Multiobjective sparse non-negative matrix factorization. IEEE Trans. Cybern. 49, 42504264. doi: 10.1109/TCYB.2018.2834898

  • 12

    HancerE.XueB.ZhangM. (2022). Fuzzy filter cost-sensitive feature selection with differential evolution. Knowl.-Based Syst. 241:108259. doi: 10.1016/j.knosys.2022.108259

  • 13

    HancerE.XueB.ZhangM. (2025). A many-objective diversity-guided differential evolution algorithm for multi-label feature selection in high-dimensional datasets. IEEE Trans. Emerg. Top. Comput. Intell. 9, 12261237. doi: 10.1109/TETCI.2025.3529840

  • 14

    IdriA.BenharH. J.Fernández-AlemánL.KadiI. (2018). A systematic map of medical data preprocessing in knowledge discovery. Comput. Methods Programs Biomed. 162, 6985. doi: 10.1016/j.cmpb.2018.05.007

  • 15

    KunduP. P.MitraS. (2017). Feature selection through message passing. IEEE Trans. Cybern. 47, 43564366. doi: 10.1109/TCYB.2016.2609408

  • 16

    LeiY.LiH.LiG. (2024). “PRSAMF: Personalized recommendation based on sentiment analysis and matrix factorization,” in Proceedings of the 2024 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) (Lisbon: IEEE), 65536560. doi: 10.1109/BIBM62325.2024.10822471

  • 17

    LiC.CheH.LeungM. F.LiuC.YanZ. (2023). Robust multi-view non-negative matrix factorization with adaptive graph and diversity constraints. Inf. Sci. 634, 587607. doi: 10.1016/j.ins.2023.03.119

  • 18

    LiJ.YuanY.LuoX. (2025). Learning error refinement in stochastic gradient descent-based latent factor analysis via diversified PID controllers. IEEE Trans. Emerg. Top. Comput. Intell.9, 35823597. doi: 10.1109/TETCI.2025.3547854

  • 19

    LiT.QianY.LiF.LiangX.ZhanZ. H. (2024). Feature subspace learning-based binary differential evolution algorithm for unsupervised feature selection. IEEE Trans. Big Data11, 99114. doi: 10.1109/TBDATA.2024.3378090

  • 20

    LiZ.ZhuoS.HeJ.QiuW.ZhengZ.ChenM.et al. (2025). Behavior enhanced representation learning for user behavior analysis. IEEE Trans. Inf. Forensics Secur.20, 92759288. doi: 10.1109/TIFS.2025.3601358

  • 21

    LiaoX.WuH.HeT.LuoX. (2025). A proximal-ADMM-incorporated nonnegative latent-factorization-of-tensors model for representing dynamic cryptocurrency transaction network. IEEE Trans. Syst. Man Cybern. Syst. 55, 83878401. doi: 10.1109/TSMC.2025.3605054

  • 22

    LinM.LinX.XuX.XuZ.LuoX. (2025). Neural networks-incorporated latent factor analysis for high-dimensional and incomplete data. IEEE Trans. Syst. Man Cybernet. Syst.55, 73027314. doi: 10.1109/TSMC.2025.3583919

  • 23

    LinX.YuS.LinM.XuX.LinJ.XuZ. (2025). An incremental nonlinear co-latent factor analysis model for large-scale student performance prediction. IEEE Trans. Serv. Comput. 18, 34633476. doi: 10.1109/TSC.2025.3621687

  • 24

    LuoC.WangS.LiT.ChenH.LvJ.YiZ. (2023). RHDOFS: a distributed online algorithm towards scalable streaming feature selection. IEEE Trans. Parallel Distrib. Syst. 34, 18301847. doi: 10.1109/TPDS.2023.3265974

  • 25

    LuoX.LiuZ.LiS.ShangM.WangZ. (2018). A fast non-negative latent factor model based on generalized momentum method. IEEE Trans. Syst. Man Cybern. Syst. 50, 111. doi: 10.1109/TSMC.2018.2875452

  • 26

    LuoX.WangD.ZhouM.YuanH. (2021a). Latent factor-based recommenders relying on extended stochastic gradient descent algorithms. IEEE Trans. Syst. Man Cybern. Syst. 51, 916926. doi: 10.1109/TSMC.2018.2884191

  • 27

    LuoX.WangZ.ShangM. (2021b). An instance-frequency-weighted regularization scheme for non-negative latent factor analysis on high-dimensional and sparse data. IEEE Trans. Syst. Man Cybern. Syst. 51, 35223532. doi: 10.1109/TSMC.2019.2930525

  • 28

    LyuC.MaZ.LuoX.ShiY. (2026). Dynamic stochastic reorientation particle swarm optimization for adaptive latent factor analysis in high-dimensional sparse matrices. IEEE Trans. Knowl. Data Eng. 38, 222234. doi: 10.1109/TKDE.2025.3621469

  • 29

    NiJ.FeiH.FanW.ZhangX. (2017). “Automated medical diagnosis by ranking clusters across the symptom-disease network,” in Proceedings of the 2017 IEEE International Conference on Data Mining (IEEE: New Orleans, LA), 10091014. doi: 10.1109/ICDM.2017.130

  • 30

    PerkinsS.TheilerJ. (2003). “Online feature selection using grafting,” in Proceedings of the 20th International Conference on Machine Learning (Washington, DC: AAAI Press), 592599.

  • 31

    QinW.LuoX.LiS.ZhouM. (2024). Parallel adaptive stochastic gradient descent algorithms for latent factor analysis of high-dimensional and incomplete industrial data. IEEE Trans. Autom. Sci. Eng. 21, 27162729. doi: 10.1109/TASE.2023.3267609

  • 32

    QiuJ.ZhuoS.YuP. S.WangC.HuangS. (2025). Online learning for noisy labeled streams. ACM Trans. Knowl. Discov. Data. 19, 129. doi: 10.1145/3734875

  • 33

    Ramírez-GallegoS.Mouriño-TalínH.Martínez-RegoD.Bolón-CanedoV.BenítezJ. M.Alonso-BetanzosA.et al. (2018). An information theory-based feature selection framework for big data under apache spark. IEEE Trans. Syst. Man Cybern. Syst. 48, 14411453. doi: 10.1109/TSMC.2017.2670926

  • 34

    ShuT.LinY.GuoL. (2024). Online hierarchical streaming feature selection based on adaptive neighborhood rough set. Appl. Soft Comput. 152:111276. doi: 10.1016/j.asoc.2024.111276

  • 35

    TangP.RuanT.WuH.LuoX. (2024). Temporal pattern-aware qos prediction by biased non-negative tucker factorization of tensors. Neurocomputing582:127447. doi: 10.1016/j.neucom.2024.127447

  • 36

    WangF. L.ZainA. M.RenY.BahariM.SamahA. A.Ali ShahZ. B.et al. (2025). Navigating the microarray landscape: a comprehensive review of feature selection techniques and their applications. Front. Big Data8:1624507. doi: 10.3389/fdata.2025.1624507

  • 37

    WangP.XueB.LiangJ.ZhangM. (2023). Feature clustering-assisted feature selection with differential evolution. Pattern Recognit. 140:109523. doi: 10.1016/j.patcog.2023.109523

  • 38

    WuD.HeY.LuoX. (2023). A graph-incorporated latent factor analysis model for high-dimensional and sparse data. IEEE Trans. Emerg. Top. Comput. 11, 907917. doi: 10.1109/TETC.2023.3292866

  • 39

    WuD.HeY.LuoX.ZhouM. (2022). A latent factor analysis-based approach to online sparse streaming feature selection. IEEE Trans. Syst. Man Cybern. Syst. 52, 67446758. doi: 10.1109/TSMC.2021.3096065

  • 40

    WuD.HuY.LiuK.LiJ.WangX.DengS.ZhengN.LuoX. (2025a). An outlier-resilient autoencoder for representing high-dimensional and incomplete data. IEEE Trans. Emerg. Top. Comput. Intell. 9, 13791391. doi: 10.1109/TETCI.2024.3437370

  • 41

    WuD.LiZ.YuZ.HeY.LuoX. (2025b). Robust low-rank latent feature analysis for spatiotemporal signal recovery. IEEE Trans. Neural Netw. Learn. Syst. 36, 28292842. doi: 10.1109/TNNLS.2023.3339786

  • 42

    WuD.LuoX.ShangM.HeY.WangG.ZhouM. (2021). A deep latent factor model for high-dimensional and sparse matrices in recommender systems. IEEE Trans. Syst. Man Cybern. Syst. 51, 42854296. doi: 10.1109/TSMC.2019.2931393

  • 43

    WuD.ZhangP.HeY.LuoX. (2024). MMLF: Multi-metric latent feature analysis for high-dimensional and incomplete data. IEEE Trans. Serv. Comput. 17, 575588. doi: 10.1109/TSC.2023.3331570

  • 44

    WuH.WangQ.LuoX.WangZ. (2025). Learning accurate representation to nonstandard tensors via a mode-aware tucker network. IEEE Trans. Knowl. Data Eng. 37, 72727285. doi: 10.1109/TKDE.2025.3617894

  • 45

    WuX.YuK.DingW.WangH.ZhuX. (2013). Online feature selection with streaming features. IEEE Trans. Pattern Anal. Mach. Intell. 35, 11781192. doi: 10.1109/TPAMI.2012.197

  • 46

    XuR.WuD.LuoX. (2025a). Recursion-and-fuzziness reinforced online sparse streaming feature selection. IEEE Trans. Fuzzy Syst. 33, 25742586. doi: 10.1109/TFUZZ.2025.3569272

  • 47

    XuR.WuD.WangR.LuoX. (2025b). A highly-accurate three-way decision-incorporated online sparse streaming features selection model. IEEE Trans. Syst. Man Cybern. Syst. 55, 42584272. doi: 10.1109/TSMC.2025.3548648

  • 48

    XuX.LinM.LiW.ZhangJ.WuH. (2023). “Time-varying QoS estimation via non-negative latent factorization of tensors with extended linear biases,” in Proceedings of the 2023 IEEE International Conference on Big Data (BigData) (Sorrento: IEEE), 8695. doi: 10.1109/BigData59044.2023.10386709

  • 49

    XueB.ZhangM.BrowneW. N.YaoX. (2016). A survey on evolutionary computation approaches to feature selection. IEEE Trans. Evol. Comput. 20, 606626. doi: 10.1109/TEVC.2015.2504420

  • 50

    XueX.YaoM.WuZ. (2018). A novel ensemble-based wrapper method for feature selection using extreme learning machine and genetic algorithm. Inf. Syst. 57, 389412. doi: 10.1007/s10115-017-1131-4

  • 51

    YangY.ChenD.WangH.WangX. (2018). Incremental perspective for feature selection based on fuzzy rough sets. IEEE Trans. Fuzzy Syst. 26, 12571273. doi: 10.1109/TFUZZ.2017.2718492

  • 52

    YaoF.DingY. L.HongS. G.YangS. H.BenítezJ. M.Alonso-BetanzosA.HerreraF. (2022). A survey on evolved lora-based communication technologies for emerging internet of things applications. Int. J. Netw. Dyn. Intell. 1, 419. doi: 10.53941/ijndi0101002

  • 53

    YuK.WuX.DingW.PeiJ. (2016). Scalable and accurate online feature selection for big data. ACM Trans. Knowl. Discov. Data11:16. doi: 10.1145/2976744

  • 54

    YuanY.LuS.LuoX. (2025). A proportional integral controller-enhanced non-negative latent factor analysis model. IEEE/CAA J. Autom. Sin. 12, 12461259. doi: 10.1109/JAS.2024.125055

  • 55

    ZhangJ. D.ChowC. Y.XuJ. (2017). Enabling Kernel-based attribute-aware matrix factorization for rating prediction. IEEE Trans. Knowl. Data Eng. 29, 798812. doi: 10.1109/TKDE.2016.2641439

  • 56

    ZhangZ.JiangW.LiF.ZhaoM.LiB.ZhangL. (2017). Structured latent label consistent dictionary learning for salient machine faults representation-based robust classification. IEEE Trans. Ind. Inform. 13, 644656. doi: 10.1109/TII.2017.2653184

  • 57

    ZhouJ.FosterD. P.StineR. A.UngarL. H. (2006). Streamwise feature selection. J. Mach. Learn. Res. 7, 18611885.

  • 58

    ZhouP.HuX.LiP.WuX. (2019a). OFS-density: a novel online streaming feature selection method. Pattern Recognit., 86, 4861. doi: 10.1016/j.patcog.2018.08.009

  • 59

    ZhouP.HuX.LiP.WuX. (2019b). Online streaming feature selection using adapted neighborhood rough set. Inf. Sci. 481, 258279. doi: 10.1016/j.ins.2018.12.074

  • 60

    ZhouP.LiP. P.ZhaoS.WuX. D. (2021a). Feature interaction for streaming feature selection. IEEE Trans. Neural Netw. Learn. Systs32, 46914702. doi: 10.1109/TNNLS.2020.3025922

  • 61

    ZhouP.WangN.ZhaoS. (2021b). Online group streaming feature selection considering feature interaction. Knowl. Based Syst. 226:107157. doi: 10.1016/j.knosys.2021.107157

  • 62

    ZhouP.ZhaoS.YanY. T.WuX. D. (2022). Online scalable streaming feature selection via dynamic decision. ACM Trans. Knowl. Discov. Data16, 120. doi: 10.1145/3502737

  • 63

    ZhuoS.-D.QiuJ.-J.WangC.-D.HuangS.-Q. (2024). Online feature selection with varying feature spaces. IEEE Trans. Knowl. Data Eng. 36, 48064819. doi: 10.1109/TKDE.2024.3377243

Summary

Keywords

feature selection, genetic algorithm, latent factor analysis, missing data, online learning

Citation

Liu G, Liu J, He G, Liu Y, Bai H and Zhou M (2026) A genetic algorithm-based framework for online sparse feature selection in data streams. Front. Big Data 9:1782461. doi: 10.3389/fdata.2026.1782461

Received

07 January 2026

Revised

18 January 2026

Accepted

20 January 2026

Published

09 February 2026

Volume

9 - 2026

Edited by

Qingguo Lü, Chongqing University, China

Reviewed by

Peng Zhou, Anhui University, China

Shengda Zhuo, Jinan University, China

Updates

Copyright

*Correspondence: Min Zhou,

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics