- 1Faculty of Computing, Universiti Teknologi Malaysia, Skudai, Johor, Malaysia
- 2Hebei Institute of Mechanical and Electrical Technology, Xingtai, China
- 3Faculty of Management, Universiti Teknologi Malaysia, Skudai, Johor, Malaysia
- 4Faculty of Computer Science and Information Technology, Universiti Malaysia Sarawak, Kota Samarahan, Malaysia
- 5Department of Software Engineering, Faculty of Computer Science and Information Technology, Universiti Tun Hussein Onn Malaysia, Parit Raja, Johor, Malaysia
- 6Faculty of Computing, Universiti Malaysia Pahang Al-Sultan Abdullah, Kuantan, Pahang, Malaysia
- 7Faculty of Artificial Intelligence, Universiti Teknologi Malaysia, Skudai, Johor, Malaysia
This review systematically summarizes recent advances in microarray feature selection techniques and their applications in biomedical research. It addresses the challenges posed by the high dimensionality and noise of microarray data, aiming to integrate the strengths and limitations of various methods while exploring their applicability across different scenarios. By identifying gaps in current research, highlighting underexplored areas, and proposing clear directions for future studies, this review seeks to inspire academics to develop novel techniques and applications. Furthermore, it provides a comprehensive evaluation of feature selection methods, offering both a theoretical foundation and practical guidance to help researchers select the most suitable approaches for their specific research questions. Emphasizing the importance of interdisciplinary collaboration, the study underscores the potential of feature selection in transformative applications such as personalized medicine, cancer diagnosis, and drug discovery. Through this review, not only does it provide in-depth theoretical support for the academic community, but also practical guidance for the practical field, which significantly contributes to the overall improvement of microarray data analysis technology.
1 Introduction
The microarray is a powerful biotechnological tool that allows for the simultaneous evaluation of the expression levels of multiple genes (Joseph and Sandoval, 2023). This technique involves immobilizing numerous nucleic acid probes onto a solid surface, such as a glass slide or a silicon chip, which are designed to specifically interact with their corresponding RNA or DNA sequences (Wang et al., 2023a). Through the examination of probe-target interactions, scientists can determine the expression levels of each gene in the sample. Due to its versatility, microarray technology finds broad applications in the study of gene expression mechanisms, identification of biomarkers, disease diagnosis, and pharmaceutical development. On the other hand, the expression levels of thousands of genes can be studied simultaneously in microarray experiments, which are a crucial aspect of modern molecular biology (Maolmhuaidh et al., 2023). However, the resulting data can be challenging to analyze due to their high dimensionality and small sample size. This complexity often leads to inaccurate results and unreliable conclusions when traditional statistical methods and machine learning algorithms are applied directly (Prajapati et al., 2023b). To address these issues, microarray feature selection techniques are employed to identify the most informative gene features, thereby reducing the complexity of the data and improving its interpretability.
Despite the advantages of microarray datasets, excessively high dimensions can have several negative effects on model performance in microarray data analysis, including overfitting, increasing computational costs, and poor interpretability of results. To combat these issues, various methods are used when dealing with microarray datasets that contain too many dimensions. Commonly used methods include feature selection and feature extraction (Labory et al., 2024). Compared to feature extraction, feature selection retains biological significance and interpretability by filtering the most important original features, and usually has a lower computational overhead (Pudjihartono et al., 2022). Feature selection has significant advantages over feature extraction in the downscaling process of microarray data (Pirch et al., 2021). First, feature selection preserves the original gene characteristics and thus results are more interpretable, which enables researchers to directly correlate selected genes with specific biological processes or disease mechanisms, providing clear guidance for biological research and clinical applications. Second, the high biological relevance of feature selection helps identify potential biomarkers and provide insight into the molecular mechanisms of disease. In addition, feature selection methods are often computationally more efficient, especially when dealing with large-scale microarray data, and many filtering methods based on statistical tests can quickly and efficiently screen out important features. By reducing the number of features, feature selection also reduces the complexity of the model, thereby minimizing the risk of overfitting, which is particularly important for high-dimensional microarray data with a limited number of samples. Finally, since feature selection preserves the original feature structure, the model can be trained and predicted directly using these features, avoiding complex transformation or preprocessing steps. As a result, feature selection shows clear advantages in scenarios that require high interpretability, direct biological relevance, and computational efficiency.
Microarray feature selection is an essential step in the analysis of gene expression data. It helps streamline the data, making it more accessible for study and providing actionable insights for researchers. Focusing on the most informative features through feature selection can not only improve the quality and interpretability of the data, but also establish a foundation for the development of the precise predictive model. However, inadequate feature selection can lead to several challenges in the analysis of microarray data, such as increased risk of overfitting, inefficient use of computational resources, and reduced clarity of data interpretation. Overcoming these issues can enable researchers to gain a deeper understanding of data and advance biomedical research. In recent developments, many studies have identified problems in existing microarrays and proposed methods to solve these problems. For example, Fadhil and Abdulazeez (2024) summarized the application of deep learning methods to overcome the high-dimensionality problem of microarray datasets. They explored how deep learning methods can be applied in the complex research field of cancer classification. Osama et al. (2023) summarized preprocessing methods for microarray datasets and discussed different feature selection methods based on machine learning. In contrast, Hambali et al. (2020) failed to provide an application of feature selection in their summary of different feature selection techniques. Given these drawbacks, there has been a lack of comprehensive summaries that cover the entire process of microarray feature selection, this study covers a wider range of research areas than previous reviews, especially summarizing research results in recent years, allowing researchers to better understand research trends.
To bridge this gap, this paper aims to explore dataset-specific feature selection methods and summarize the advantages and disadvantages of each category of methods. Additionally, this article will discuss the various application areas of microarrays. In preparation for this paper, the keyword 'microarray feature selection' was used to search for articles published after 2019 on Google Scholar. This review will first provide an overview of the basic concepts of microarray technology. Next, various methods of microarray feature selection will be compared and analyzed, different application fields of microarray feature selection will be summarized, and technical challenges and potential future research directions in this field will be evaluated. The following chapters will cover key aspects of microarray feature selection. Section 2 will give a detailed overview of the relevant concepts of microarray feature selection. Section 3 will comprehensively review the existing literature and evaluate the advantages and disadvantages of various feature selection methods. Section 4 will focus on the practical applications of microarray feature selection in different fields. The final section will explore potential issues and predict future development directions, aiming to provide valuable guidance and insights for future research.
2 Microarray feature selection process
This section focuses on basic concepts and terminology related to microarray feature selection, providing an in-depth look at the complexity of microarray datasets, design principles, and the various advantages and disadvantages of different feature selection methods. Furthermore, it will be explored how to evaluate and compare the effectiveness of these methods and how to select the most appropriate subset of features to achieve accurate model results.
2.1 Introduction
Microarray technology dates back to the late 1990s and early 2000s and is designed to measure the expression levels of numerous genes simultaneously (Moses and Pachter, 2022). As microarray technology continues to mature, the fields of molecular biology, bioinformatics and statistical analysis have also made great progress (Singh et al., 2023). These advances ensure that microarray technology remains an indispensable tool for systems biologists and clinical researchers, driving discoveries and improving patient care (Vatansever et al., 2021). Figure 1 shows The development of microarrays in recent years. Researchers can use high-throughput microarray technology to simultaneously analyze the expression levels of thousands of genes or detect specific DNA sequences, which allows researchers to delve deeper into a gene's activity under specific circumstances. The technology can also be used for gene expression analysis, genotyping, drug discovery and disease diagnosis.
In the analysis of high-dimensional microarray data, the choice of feature selection methods is critical to control the risk of overfitting. Microarray data are usually of extremely high dimensionality but with a limited number of samples, resulting in a model that is prone to overfitting on the training data. This risk can be effectively reduced by choosing an appropriate feature selection method. First, feature selection removes redundant and noisy features and reduces model complexity, which is critical to minimizing overfitting. Controlling the number of features selected is equally critical; too many features may cause the model to capture random fluctuations instead of the true signal. In addition, a robust feature selection method improves the reliability of the selected features and avoids instability due to small variations in the data, which further reduces the likelihood of overfitting. Choosing a feature selection method that matches the complexity of the model ensures that the selected feature set best fits the current model. Finally, embedding cross-validation into the feature selection process can more accurately assess the contribution of features to model performance and avoid features that are only valid for the training data, thus effectively reducing the risk of overfitting.
The class imbalance problem in microarray datasets can be effectively addressed through feature selection, and the key is to employ multiple strategies to enhance the recognition of minority classes. First, prioritizing features that can significantly differentiate between minority and majority classes ensures that the model is more likely to capture signals from minority classes. Second, class weights are introduced into the feature selection process so that features of minority class samples are given higher importance in the selection. In addition, balancing the dataset before feature selection through undersampling or oversampling techniques prevents the majority class from dominating the feature selection process, resulting in a more representative feature set.
This paper will examine specific feature selection methods in more detail, discussing their theoretical bases, practical applications, and the challenges associated with implementing these methods in different research contexts. This discussion will provide a comprehensive understanding of how microarray feature selection is integral to refining data analysis and ensuring the reliability of research outcomes in the field of genomics.
Based on the overview provided previously, the main concepts of microarray feature selection can be divided into three key parts, as shown in Figure 2. This visual framework helps succinctly organize various aspects of feature selection into a coherent structure, thereby promoting deeper understanding. Next, we will delve into the related concepts of these three parts. Each component plays a unique role throughout the feature selection process, covering everything from initial data preparation to the final selection of features that best predict the outcome of interest.
2.2 Microarray dataset
Microarray technology, a pivotal tool in genomic research, enables the high-throughput analysis of gene expression across numerous conditions and diseases (Yang et al., 2020). Within the realm of binary classification, several classic datasets are frequently employed, each specific to particular types of cancer or disease states. For instance, the Colon Cancer dataset includes gene expression profiles from colon tissue and is used to study colorectal cancer. This dataset helps in identifying genes or patterns associated with different stages or types of colon cancer, thereby aiding in diagnostics and potential treatment strategies (Shafi et al., 2020). Similarly, the leukemia dataset provides gene expression data specifically related to leukemia, a type of blood cancer. Including samples from various subtypes of leukemia, this dataset allows researchers to delve into the molecular characteristics of the disease and identify potential biomarkers for diagnosis and treatment. Additionally, the prostate dataset focuses on prostate cancer, a prevalent condition among men. It contains gene expression profiles associated with prostate tissue or cells to identify markers that can differentiate benign from malignant prostate disease or enhance our understanding of disease progression. Another key dataset is the DLBCL dataset, which stands for diffuse large B-cell lymphoma, one of the most common non-Hodgkin lymphomas. It includes gene expression data from lymphoma tissues, helping researchers to identify genetic markers or patterns associated with different DLBCL subtypes and treatment responses (Shukla and Tripathi, 2020). What's more, the CNS dataset involves various molecular data related to diseases affecting the central nervous system. This dataset includes gene expression profiles from conditions such as brain tumors and neurological disorders, enabling researchers to understand the molecular signatures associated with CNS disorders (Sánchez-Maroño et al., 2019).
For multi-classification datasets, this study also summarizes commonly used datasets, which are equally important in genomic research. The SRBCT dataset involves gene expression profiling of small round blue cell tumors. The dataset includes four categories of tumors and is commonly used to distinguish them and identify specific genetic markers associated with each subtype (Sahu and Dash, 2023). Likewise, the Lung Cancer (Harvard) dataset focusing on lung cancer is another great resource (Karthika et al., 2023). Organized into five categories, the dataset helps identify genetic patterns that distinguish various subtypes or stages of lung cancer, thereby aiding diagnostic and treatment strategies. Additionally, the Leukemia2 dataset contains three categories that help researchers understand the molecular differences between leukemia subtypes and assist in identifying biomarkers for accurate diagnosis or targeted therapy (Rupapara et al., 2022). Additionally, the 9Tumor and Brain Tumor1 datasets provide valuable insights into molecular variations between different tumor types or subtypes in different tissues or organs, aiding in classification and providing potential therapeutic insights (Zhu et al., 2023).
In summary, these microarray datasets, encompassing both binary and multi-classification data, serve as invaluable resources for researchers across numerous fields. By exploring and analyzing these datasets, scientists can uncover crucial insights and advancements in areas such as cancer research and neurological disorders. This paper provides a comprehensive analysis of datasets used in various articles highlighted the most frequently employed datasets in both binary and multi-class classifications, as detailed in Tables 1, 2. This review not only underscores the importance of these datasets but also reflects ongoing efforts to address the challenges associated with microarray data analysis. Figure 3 provides a visual representation of the proportion of datasets used, further illustrating the critical role these datasets play in advancing our understanding of complex biological processes and diseases.
2.3 Feature selection method
Despite the power of microarray technology in analyzing gene expression and other biological processes, it still faces several challenges. These include the complexity of data analysis, high costs, sensitivity and dynamic range issues, and the need for high-quality biological samples. In addition, microarray data often exhibit characteristics such as high dimensionality and small sample sizes, which pose additional challenges such as noise and outlier issues (Hamraz et al., 2023). Feature selection is the main approach to this problem, with the goal of selecting a subset of the most important and useful features from a larger set of attributes or variables (Dhal and Azad, 2022). This process is particularly important in microarray data analysis because it identifies features that represent gene or protein expression levels and can better enhance data analysis. By identifying and retaining only the most important features, feature selection can greatly improve prediction accuracy and generalization capabilities, especially when dealing with limited sample data. In microarray analysis, this approach helps pinpoint genes associated with specific biological processes or disease states, providing valuable insights for interpretation and discovery of potential therapeutic targets.
Given the large number of genes typically present in microarray data, it is often the case that only a subset of these genes are relevant to the specific biological process or disease being studied (Jovic et al., 2022). The challenges of microarray data analysis are amplified by the presence of redundant features, which significantly increases computational complexity and the risk of overfitting. For example, for a dataset containing N features, the number of potential feature subsets is up to 2N (Singh and Singh, 2021). This high dimensionality increases the risk of overfitting and highlights the urgent need to select a high-quality feature subset. Without effective feature selection, analysis can produce inaccurate results and lead to unnecessarily complex models. Feature selection simplifies analysis by reducing data dimensionality, which not only makes analysis more effective and efficient, but also improves prediction accuracy, increases interpretability, reduces the risk of overfitting, and improves computational efficiency.
Therefore, feature selection is a critical pre-processing step before applying machine learning algorithms to simplify data by eliminating irrelevant or redundant features. This improves model accuracy, reduces computational load, and produces results that are easier to interpret. In microarray analysis, the dimensionality of the data is very high, so how to obtain the optimal feature subset is particularly important (Lee et al., 2021). A good feature subset will significantly affect the performance and interpretability of the model (Yun et al., 2023). Each step in the feature selection process is closely linked, so it is crucial to design and execute these steps carefully. The feature selection process in microarray analysis begins with an initial subset search to create an initial subset of features that is evaluated and compared to previously considered subsets. If a new subset is found to be more suitable under the given evaluation criteria, this subset is retained. This iterative process continues until a predefined stopping condition is met, marking the end of the feature selection process. The selected feature subset is then used to verify the effectiveness of the feature selection method. Among the most commonly used methods for feature selection in microarrays are filter, wrapper, embedded, and other methods.
The filter feature selection method is characterized by simplicity and effectiveness. It filters features to eliminate those features that have the least impact on the target variable. This is usually achieved by setting a threshold or selecting the top k features based on statistical significance, this method can minimize the computational overhead. Evaluate the importance of each feature by calculating indicators such as information gain (IG), mutual information (MI), chi-square test, correlation coefficient, minimum redundancy, maximum correlation or Fisher score of the feature (Gong et al., 2022), rank these features according to importance and select those with the highest importance.
The wrapper method is a feature selection technique that directly links the evaluation of feature subsets to the performance of a machine learning model (Effrosynidis and Arampatzis, 2021). Unlike filter methods that rely on general statistical measures, wrapper methods are inherently more complex, as they involve training the model multiple times with different subsets of features and determining the most effective combination of features that enhance model performance through an iterative process. Wrapper methods can be divided into three core steps: the first step involves generating various feature subsets; the second step is the evaluation phase, where each subset is used to train the model to assess its performance; the final step involves selecting the feature subset that meets the criteria best, thereby effectively optimizing the model's predictive accuracy.
In the literature, the application of wrapper methods is usually a combination of intelligent optimization algorithms to search as many possible feature subsets as possible, and classifiers to identify those features that maximize the performance of the classifier. Subset. The role of the classifier is crucial as it evaluates the quality of each feature subset in terms of prediction accuracy. Through this collaborative interaction, wrapper methods exploit iterative refinement of feature subsets, aiming to arrive at a near-optimal set. Specifically, the effectiveness of wrapper methods depends on their ability to fine-tune the feature selection process through continuous feedback between subset evaluation and model training stages. This feedback is crucial to obtain the most informative subset of genes, thereby ensuring that the final subset of features is selected to be the best suited for the specific prediction task of the model.
The embedded feature selection is an important component in machine learning and data analysis due to its integration in the model training process. Unlike other techniques that operate independently of the training process, embedded methods make feature selection an inherent part of model learning. This integration allows the method to evaluate feature importance directly through the learning algorithm itself. During the training phase, the embedded method automatically weighs the relevance of each feature, allowing the model to focus on those features that are most critical for the prediction task. This inherent capability makes the model more effective in reducing the risk of overfitting and improving overall prediction accuracy by focusing on the most relevant features. The efficiency of the embedded approach is further demonstrated by its simplified feature selection process, which is directly embedded into the model's training algorithm, which simplifies the entire process and helps develop more efficient and effective machine learning models.
Figure 4 illustrates the filter method, the wrapper method and the embedded method. The filter method in Figure 4a can ensure that the most influential features are retained, thus improving the effectiveness of the predictive model while meeting the challenges of high-dimensional datasets; the wrapper method in Figure 4b outlines the sequential steps of subset generation, evaluation, and final selection; and the embedded method in Figure 4c saves computational resources by combining feature selection with model training, and also enables feature selection to be more closely aligned with the specific goals of the model.

Figure 4. Process of different methods. (a) Process of filter method. (b) Process of wrapper method. (c) Process of embedded method.
In addition to the main feature selection methods such as filter, wrapper and embedded methods, there are “other methods,” the main ones being hybrid methods. Hybrid methods are particularly important in microarray feature selection because of their ability to combine multiple techniques to optimize performance and mitigate limitations inherent to individual methods. Since microarray data are often rich in features, hybrid methods are more suitable for such datasets, as these methods take advantage of various selection methods to obtain subsets of features. There are multiple strategies for implementing hybrid methods, each taking advantage of different feature selection techniques to obtain more robust results. One strategy is to integrate multiple feature selection algorithms. This approach may include combining filters, wrappers, and embedding methods, with the goal of leveraging the unique strengths of each method to achieve a more comprehensive and efficient feature selection process. Another popular implementation strategy is the multi-stage feature selection method. In this approach, the selection task is divided into multiple stages, with different techniques applied at each stage. For example, filter methods can be employed in the initial stage to quickly reduce the size of the feature set. This reduced set can then be refined into a highly correlated final subset using more computationally expensive wrapper or embedded methods.
Table 3 provides a summary of the limitations and application scope of each method. When choosing an appropriate feature selection method, it is crucial to understand the characteristics of the dataset, the requirements of the current problem, and the available computing resources. It is crucial to recognize the advantages and limitations of each method, as different methods may be more suitable for different scenarios. In practical applications, a comprehensive evaluation is required on a case-by-case basis to determine the most effective feature selection method. This decision-making process ensures that the chosen method is a good fit with the goals and limitations of the study, ultimately helping to obtain more precise and reliable results in microarray data analysis.
2.4 Subset evaluation criteria
In microarray data analysis, evaluating feature subsets effectively is crucial for building accurate and reliable predictive models. This evaluation is conducted using a set of standards and methods known collectively as Subset Evaluation Criteria. A feature subset in this context refers to a selection of features chosen from the original set based on their potential relevance to the analysis or predictive tasks at hand. Several key metrics are commonly used to assess the performance of these feature subsets. These criteria include accuracy, recall, precision, sensitivity, and the F1 score, each serving a specific purpose in measuring different aspects of model performance: the Accuracy criterion measures the overall correctness of the predictions made by the model. It is a general indicator of how well the model performs across all classes. Recall (Sensitivity) measures the model's ability to correctly identify all positive samples. It is crucial for scenarios where missing a positive instance could have serious consequences. Precision evaluates the proportion of identified positives that are correctly predicted. High precision indicates that a model does not label negative samples as positive. F1-score metric combines precision and recall into a single metric by calculating their harmonic mean. The F1 score is particularly useful when you need to balance precision and recall, which is often the case in studies where both false positives and false negatives carry significant costs. These criteria are fundamental in assessing the quality of feature subsets and optimizing the feature selection process. They help researchers understand not just the effectiveness of the feature selection but also the potential impact of selected features on the model's ability to make accurate predictions. Each of these criteria can be quantitatively assessed using specific equations, outlined from Equations 1–4, which detail how each metric is calculated based on the true positives, false positives, true negatives, and false negatives derived from the model output.
In addition to the general evaluation criteria, the analysis of high-dimensional microarray data also relies on several specific validation methods. These methods are particularly crucial due to the challenges posed by the large number of features relative to the number of samples, a common scenario in microarray data (Alhenawi et al., 2022). These validation methods often consider two critical factors, the final subset size and the time required for the selection process. The final subset size is an important metric because it directly affects both the complexity of the model and its generalizability. A smaller subset can lead to a simpler, more interpretable model that is less likely to overfit, whereas a larger subset might capture more complex patterns at the risk of overfitting. Balancing this size is crucial for building robust predictive models. The time required for the feature selection process is another vital consideration, especially in high-dimensional data scenarios (Chen et al., 2020). Feature selection in microarray data can be computationally intensive, and the time spent selecting features can significantly impact the overall efficiency of the data analysis pipeline. Faster methods that still maintain high accuracy are preferable in scenarios where time is a constraint or when dealing with very large datasets.
The proper application of these evaluation criteria allows for a comprehensive assessment of feature subsets, guiding researchers in refining their feature selection strategies to enhance model accuracy and reliability. By using these metrics, researchers can ensure that the chosen features contribute positively to the overall performance of their models, particularly in the predictive analysis of complex biological data such as that encountered in microarray studies.
Where TP is true positive, which denotes the number of positive categories predicted correctly, TN is true negative, which denotes the number of negative categories predicted correctly, FP is false positive, which denotes the number of negative categories misclassified as positive, and FN is false negative, which denotes the number of positive categories misclassified as negative.
2.4.1 Cross validation
Cross-validation is designed to assess the generalization ability of a model. By repeating training and evaluation on different training-validation set divisions, cross-validation helps us to reduce the model's dependence on specific data divisions and effectively prevent overfitting. In addition, cross-validation plays a key role in model selection by verifying the stability and accuracy of each configuration and selecting the optimal model.
When evaluating different models using cross-validation, the first step is to choose the appropriate method and set the relevant parameters. Taking k-fold cross-validation as an example, the dataset is usually divided into k subsets, and k experiments are repeated, each time one of the subsets is selected as the validation set, and the remaining k−1 subsets are used as the training set. This process will result in k validation results, and finally the mean and standard deviation of these k results are calculated as the overall performance metrics of the model on this dataset.
The robustness and consistency of the model can be understood by observing the average performance and standard deviation of the model across folds. Smaller standard deviations indicate that the model's performance is more stable across different data divisions; while higher average performance values indicate that the model has better generalization ability. Based on these results, the reliability of the model can be further determined and the best solution can be selected by comparing the cross-validation performance of different models or parameter combinations.
2.5 Summary
This section provides a summary of the basic principles of the microarray feature selection procedure. It covers important topics such as microarray data, the feature selection method, evaluation metrics for feature subsets, and the key components of the feature selection process. Understanding and implementing microarray feature selection based on these concepts is crucial for researchers to make well-informed decisions when working with microarray data.
3 Feature selection method on microarray
The process of feature selection is crucial in data analysis as it aims to identify the most relevant and informative features from a dataset, especially in high-dimensional data like microarray gene expression. By reducing the dimensionality of irrelevant data, feature selection enhances the efficiency and accuracy of subsequent analysis. This chapter offers a comprehensive overview of different methods of feature selection.
3.1 Filter feature selection method on microarray
In microarray data analysis, feature selection is a crucial step that helps identify and select genes most relevant to specific biological phenomena. The filter feature selection method is widely popular as a main strategy because it is highly efficient and easy to implement. This section explores the application of filter feature selection methods in microarray data analysis.
Nematzadeh et al. (2019) proposed a filter method employing the whale algorithm and Mutual Congestion to address this issue. They initially set the number of whales equal to the number of features and applied the whale algorithm to eliminate irrelevant features. They then ranked the remaining features using Mutual Congestion. While effective in selecting features with lower interference frequencies, the non-deterministic specification of the subset size using a threshold of 10 could benefit from an adaptive value. In another study, Li and Xu (2019) focused on obtaining effective gene expression data related to Hepatocellular Carcinoma (HCC). They utilized the Fisher score algorithm to identify characteristic HCC-related genes and performed various functional enrichment analyses. Additionally, they conducted a survival analysis to assess the relationship between selected central genes and patient survival. Addressing class imbalance, He et al. (2019) introduced the imRelief algorithm, demonstrating superior performance in handling minority sample dispersion across microarray datasets compared to various evaluation metrics.
Tavasoli et al. (2021) took measures to enhance classification accuracy. They employed data shuffling to prevent overfitting and utilized a soft-weighted ensemble mechanism with five criteria for feature selection. The study highlighted the effectiveness of combining improved algorithms and multi-mechanism soft weighting in mitigating overfitting and instability issues. However, its robustness was only tested on a limited number of benchmark datasets, requiring further verification. Furthermore, Lee et al. (2021) introduced the MB Ranking method, effectively addressing data type inconsistency in microarray datasets by leveraging the formal definition of Markov Blanket (MB) for multivariate feature ranking. This technique outperformed other ranking methods due to its inherent feature ranking advantages. In their pursuit of addressing the computational complexity linked to wrapper-based models in high-dimensional microarray datasets, Saberi-Movahed et al. (2022) introduced the Dual Regularized Unsupervised Feature Selection Based on Matrix Factorization and Minimum Redundancy (DR-FS-MFMR). This approach efficiently combines matrix factorization and subspace learning techniques to represent datasets through a matrix factorization form, enhancing the selection of more efficient features by capturing local and global correlations within the feature space. The proficiency of DR-FS-MFMR was demonstrated across nine gene expression datasets, and it was compared with nine methods using clustering accuracy and normalized mutual information. However, as feature selection was conducted via clustering, redundant features might exist within the final subset.
Overall, these studies offer a range of approaches to tackle specific challenges in feature selection in microarray datasets. Each study provides unique insights and methodologies to enhance accuracy and efficiency in selecting significant features. Filter method ranks features by calculating statistical metrics or scoring functions and does not rely on learning algorithms, the process involves calculating the statistical metrics or scoring functions for each feature, ranking the features, and selecting the top-ranked subset of features as the final result. The advantages of this method are high computational efficiency, not easy to overfitting and simplicity, which is suitable for preliminary feature screening. However, the disadvantages of the filter method are that it ignores the correlation between features, which may lead to the omission of important features, the selected subset of features may not be able to improve the model performance in some cases, and the filter method has a limited generalization ability when facing complex datasets.
3.2 Wrapper feature selection method on microarray
The wrapper feature selection method is a commonly used approach for finding the best feature subset using a specific algorithm. In this chapter, we will explore the use of wrapper feature selection methods in analyzing microarray data and their connection with optimization algorithms. Wrapper feature selection is closely tied to optimization algorithms. The objective of wrapper feature selection is to minimize or maximize a performance measure like classification accuracy or mean square error, thus treating it as an optimization problem. Optimization algorithms, such as genetic algorithms, simulated annealing, and particle swarm optimization, offer efficient means of optimizing wrapper feature selection methods.
In recent years, various optimization algorithms have been used for feature selection in microarrays. Almugren and Alshamlan (2019) introduced the innovative FireFly (FF) algorithm, while Chatterjee et al. (2020) improved the Social Ski Driver (SSD) algorithm by incorporating Late Acceptance Hill Climb (LAHC) to enhance its local search capabilities. They transformed the algorithm into a binary form using S-shaped and V-shaped transfer functions. To address the limited local search capabilities of the Whale Optimization Algorithm (WOA), Guha et al. (2020) introduced the embedded chaotic whale survival algorithm (ECWSA). This method introduced death and chaos mechanisms, improving the description of whale predation. Agrawal et al. (2020) proposed a feature selection method called QWOA, which modified the mutation and crossover operators applied to the quantum-inspired whale motion in WOA. Khamparia et al. (2020) developed a pioneering feature selection and classification method that utilized GA and a diverse ensemble of classifiers. They used the Bhattacharya coefficient and GA to remove noise features and derive the target feature set. Panda (2020) proposed an Elephant Search Algorithm (ESA) and Deep Learning (DL) based wrapper method for feature selection. Too and Mirjalili (2021) presented the Hyper Learning Binary Dragonfly algorithm (HLBDA) based on the Binary Dragonfly Algorithm (BDA). Abdel-Basset et al. (2021) combined the Harris Hawks Optimization algorithm (HHO) with simulated annealing (SA) to create a new feature selection approach. By using the HHO output as input for SA, they achieved a seamless integration of both algorithms. They also employed bitwise OR and bitwise AND operations to overcome limitations in population diversity that could affect HHO's performance. Das et al. (2022) introduced a novel feature selection method rooted in the Jaya optimization algorithm. By leveraging the Jaya algorithm's search technique, they streamlined the feature space by updating the weakest features. Hu et al. (2022) improved the slime mold algorithm (SMA) by employing V-shaped transfer functions to obtain binary BDFSMA. Ganesh et al. (2023) utilized the Weighted Superposition Attraction Optimization Algorithm (WSA) for microarray feature selection. There are also efficient Harmony search (HS) algorithms (Ye et al., 2023), the hybrid method developed by Bae et al. (2021) based on HS also achieved high accuracy in the colon cancer.
The wrapper method relies on a classifier to assess the performance of different subsets of features. By analyzing how well the classifier performs on a specific dataset, the wrapper method can choose the best subset of features to improve the model's performance. Some commonly used classifiers in this method are KNN, DT, RF, SVM, and others. KNN is especially popular among researchers. For example, Chatterjee et al. (2020), Guha et al. (2020), Too and Mirjalili (2021), Abdel-Basset et al. (2021), Hu et al. (2022), and Ganesh et al. (2023) used SSD, WOA, BDA, HHO, SMA, and WSA respectively in combination with the KNN classifier to achieve feature selection in microarray datasets. Other commonly used classifiers include SVM and deep learning (DL). Almugren and Alshamlan (2019) used a combination of FireFly and SVM, while Panda (2020) used ESA and DL for microarray dataset classification.
There are also approaches that involve multiple classifiers. For instance, Khamparia et al. (2020) developed a convolutional neural network with multiple classifiers to create a multi-level ensemble model for diagnosing neuromuscular samples. The ensemble method, based on deep convolutional neural networks, showed superior accuracy in disease diagnosis and prediction compared to other classifiers. In addition to combining multiple classifiers, there are cases where the same algorithm is used to test the classification effect of different classifiers. Agrawal et al. (2020) and Das et al. (2022) used multiple classifiers, such as KNN, LDC, SVM, C4.5, and RT, to compare their classification effects.
Table 4 provides a comprehensive summary of the methodologies, algorithms, and their performance in feature selection and classification across various studies.
Wrapper method microarray feature selection evaluates and selects a subset of features by using the performance of a learning algorithm, and the process involves starting with an initial set of features, gradually adding or removing features, and evaluating the effect of different subsets of features based on the performance metrics of the learning algorithm, and ultimately selecting the subset of features with the best performance as the result. The advantage of this approach lies in the direct optimization objective, which can better optimize the performance of the final model by directly using the performance of the learning algorithm to evaluate the feature subset. In addition, the wrapper method is flexible and can be combined with multiple learning algorithms to adapt to different data and tasks, and is usually capable of selecting a relatively small, but superior performance feature subset. However, the drawbacks of this method are the high computational cost and the need to train the learning algorithms multiple times to evaluate the performance of different feature subsets, which is computationally expensive. In addition, due to multiple evaluations on the training data, the wrapper method is susceptible to overfitting, which may reduce the generalization ability of the model on test data, and as the number of features increases, evaluating all the possible combinations of features becomes infeasible.
3.3 Embedded feature selection method on microarray
The essence of embedded feature selection lies in its integration with the model training process. This means that the selection of features is inherently tied to the learning algorithm. This approach allows for the concurrent optimization of both the model parameters and the feature subset, with the aim of enhancing the model's efficacy on both the training and validation datasets. In this section, we delve into contemporary embedded feature selection techniques. We examine their foundational principles, procedural frameworks, and their respective merits and limitations when applied to microarray data analysis. Furthermore, we showcase the practical utility of these methods through their application to real-world datasets. We analyze their performance across various contexts and highlight their comparative strengths.
Sun et al. (2019) addressed issues of data distribution in the error-correcting output coding (ECOC) algorithm by leveraging Data Complexity theory. Their algorithm optimized ECOC encoding matrices and consistently outperformed state-of-the-art algorithms across microarray datasets. Lopez-Rincon et al. (2019) proposed an integrated feature selection strategy that utilized multiple techniques and classifiers. Their approach aimed to discover robust miRNA signatures and demonstrated high classification accuracy across diverse datasets and platforms. Climente-González et al. (2019) developed Block HSIC Lasso, a feature selection method adept at handling ultra-high-dimensional data. This method showcased enhanced performance with larger datasets and required fewer features to achieve comparable classification accuracy to other methods. Tang et al. (2019) tackled non-IID features through latent representation learning and graph-based manifold regularization (LRLMR). Despite not excelling in one dataset, this innovative unsupervised feature selection method exhibited robust intrinsic data structure characterization in microarray datasets. Kang et al. (2019) proposed rL-GenSVM for high-dimensional tumor datasets. This method combined Relaxed Lasso for feature selection with GenSVM as the classifier. The approach effectively selected and classified features in tumor datasets. Jeon and Oh (2020) introduced the Hybrid-RFE ensemble algorithm, which amalgamated SVM-RFE, RF-RFE, and GBM-RFE methods. This method, validated on UCI and NCBI gene expression datasets, showcased improved performance over single RFE methods. This improvement was mainly due to weight summation, which significantly reduced the number of features while enhancing accuracy. In their pursuit of minimal yet informative gene combinations, Peng et al. (2021) introduced the multi-layer iterative feature selection method, MGREF. Their GA-REF algorithm, a fusion of Genetic Algorithm (GA) and Recursive Feature Elimination (REF), operated in a multi-layer fashion, dividing datasets and proceeding through three distinct stages. While effectively retaining optimal features, this method preserved a slightly larger feature set than existing selection methods. The method proposed by Hamla and Ghanem (2024) selects the top ranked features obtained from the Fisher score to provide a candidate subset for the embedding stage. Then Support Vector Machine Recursive Feature Elimination is utilized and applied to the candidate subset to find the best subset. To achieve better classification accuracy of Lasso in DNA microarray data classification, Vatankhah and Momenzadeh (2024) used a method to automatically find the optimal regularization parameters. Results on four commonly used datasets demonstrate the effectiveness of the method.
Embedded feature selection methods can tightly integrate feature selection and model parameter optimization with the model training process, thereby improving model performance on training and validation datasets. This approach allows automatic selection of the most relevant feature subsets within the framework of a learning algorithm, avoiding a separate feature selection step. The advantage is that feature selection is embedded in model training, which can process data efficiently. It can also optimize model parameters and feature subsets at the same time, improving the overall performance of the model. Feature selection and model training are performed simultaneously, which reduces the process of manual intervention and improves the overall performance. However, this approach relies on specific learning algorithms that increase the complexity and training time of the model, and some embedded methods may only be applicable to specific types of data or tasks and may not be as effective as specialized feature selection methods in some cases.
From the distribution of publication years, it is evident that most articles concerning embedded feature selection are concentrated in the year 2019. This trend may be correlated with the robustness of embedding methods and the advancement of alternative techniques. The robustness of embedding methods hinges upon the chosen machine learning models, the selection of an inappropriate model for a specific dataset or problem may result in unstable feature selection outcomes. Among alternative methods, hybrid methods are predominantly utilized. These methods effectively enhance model generalization by amalgamating the outcomes of various feature selection techniques. By integrating multiple approaches, hybrid methods better capture genuine patterns within the data and mitigate the risk of overfitting, thereby enhancing the predictive capacity of models on novel samples.
The studies examined various embedded feature selection methods, each offering unique strategies to optimize feature subsets within datasets. Table 5 is a comprehensive summary table that encapsulates the key methodologies, algorithms, and their performance in feature selection and classification across various studies. These embedded feature selection methodologies catered to diverse dataset complexities. They offered strategies to optimize feature subsets efficiently while addressing specific challenges in data distribution and dimensionality.
3.4 Hybrid feature selection method on microarray
In addition to classic methods, a variety of alternative strategies have emerged in the field of feature selection. Among these, hybrid methods have become one of the most popular methods, commonly involving a combination of filter and wrapper methods. When filter and wrapper methods are used together for microarray feature selection, the filter method initially eliminates irrelevant features quickly based on specific criteria or statistical metrics. Subsequently, the wrapper method selects features that significantly impact prediction accuracy under the guidance of model performance. Intelligent optimization algorithms are often employed for efficient subset search, and classifiers are typically used for evaluation. For instance, Shukla et al. (2019b) introduced the TLBOSA method, combining Teaching Learning-based Optimization and Simulated Annealing algorithms, utilizing SVM as a fitness function. Alanni et al. (2019) employed Information Gain (IG) and Gene Expression Programming for initial feature selection, followed by SVM-based fitness function for further refinement. Loey et al. (2020) proposed an intelligent decision support system utilizing IG for initial gene selection and Gray Wolf Optimization algorithm (GWO) for feature reduction, coupled with SVM for classification. Moreover, Alomari et al. (2021) introduced rMRMR-MGWO, combining mRMR and GWO methods, leveraging SVM for classification. Houssein et al. (2021) utilized IG in conjunction with Barnacles Mating Optimizer Algorithm (BMO) and SVM for feature selection. Mahesh et al. (2024) developed a new method for predicting leukemia microarray gene data based on a new technique of hybrid ant lion mutant colony optimization as well as PSO. Dabba et al. (2021b) proposed MIM-mMFA, employing MIN-MAX, Maximum Mutual Information (MIM), and a modified Moth Flame Algorithm for feature selection alongside SVM. Additionally, Dabba et al. (2021a) introduced another approach where mRMR is used in the first stage, and in the second stage, a quantum moth flame optimization algorithm (QMFOA) and SVM are employed to achieve similar effects.
Random Forest (RF), k-Nearest Neighbors (KNN), and Naive Bayes (NB) are also widely used. For example, Shukla et al. (2019c) proposed a feature selection framework, and the specific implementation steps are called Filter-Wrapper Feature Subset Selection (FWFSS). This hybrid method uses a conditional mutual information maximization-based filter method and GA algorithm-based wrapper method to enhance the overall classification performance, using the NB classifier as the fitness function during the wrapper method. This hybrid method outperforms the compared many existing filter algorithms in both classification accuracy and optimal number of features. Ali and Saeed (2023) also developed a hybrid method based on GA. Pashaei and Pashaei (2019) incorporated RF into their approach. Initially, they employed RF ranking to remove noise and redundant features. Subsequently, they applied the Intelligent Dynamic Genetic Algorithm (IDGA) and a RF-based wrapper method for Microarray feature selection. Tatwani and Kumar (2019) introduced a method termed Master-slave Genetic Algorithms (GAs) for feature selection. Their approach begins with an initial preprocessing stage utilizing IG to eliminate redundant features. Subsequently, employing the Master-slave GA and RF for feature selection. Additionally, it needs more comparison with other algorithms, necessitating further research to ascertain its effectiveness comprehensively. Alhenawi et al. (2023) developed a hybrid method based on improved intelligent water drop algorithm and filter method. Sahu and Dash (2024) developed a method based on Jaya algorithm and IG. Sucharita et al. (2024) applied moth-flame optimization and extreme learning machine for Microarray feature selection. Dash et al. (2022) employed statistical measures to select the top 100 features. They improved the Shuffled Frog Leaping Algorithm (SFLA) by adjusting the frog jumping step size and combined it with KNN for microarray feature selection. Experimental comparison results on binary classification datasets indicate certain advantages of this method, demonstrating its effectiveness.
There are also some studies that use different classifiers for comparison. Gangavarapu and Patil (2019) proposed a hybrid greedy ensemble approach optimized using the GA to reduce the dimensionality of high-dimensional biomedical datasets. This method uses different information measures in the filter stage and compares the efficiency of KNN, DT and RF classifiers in the wrapper stage. Shukla et al. (2019a) introduced various methods of methodology. Initially, they utilized Conditional Mutual Information Maximization (CMIM) for the primary feature selection stage. Subsequently, the Binary Genetic Algorithm (BGA) served as the fitness evaluator for the features. Furthermore, classifiers such as KNN, SVM, DT and RF were employed to compute the subset's fitness value. Shukla et al. (2020) take advantage of the advantages of teaching learning-based algorithm (TLBO) and gravitational search algorithm (GSA) algorithms to develop a new high-search efficiency algorithm, TLBOGSA, and introduce a new encoding strategy to convert its continuous search space into a binary search space. Before using TLBOGSA for feature selection, mRMR is first used to select a feature subset, and then the wrapper method based on TLBOGSA is used for feature selection, they compared the effects of four classifiers, SVM, KNN, DT and NB, and finally confirmed that NB classifier is the most effective.
Some other microarray feature selection methods are hybrids of the two methods. To capture the interaction of features and solve the classification problem of data imbalance, Hosseini and Moattar (2019) proposed a hybrid feature selection method called mutual information and Monte Carlo-based feature selection (MIMCFS). The technique is divided into two stages: mutual information to select main features and the Monte Carlo tree search technique to eliminate redundant features. However, in this method, some parameters are set based on an empirical basis. Finding a better method for setting these parameters may lead to better experimental results. Kilicarslan et al. (2020) employed the ReliefF and Stacked AutoEncoder (SAE) methods for dimensionality reduction. Subsequently, they utilized SVM and Convolutional Neural Networks (CNN) for classification. The dimensionality reduction and classification techniques were combined pairwise to validate the accuracy of feature selection. Jain and Singh (2021) proposed a fast, general-purposed, influential hybrid feature selection approach with an adaptive classification method for chronic disease datasets that can enhance the classifier's efficiency and decrease computation cost and time. This approach outperforms the traditional SVM classifier regarding all significant performance measures and shows outstanding results. The critical aspect of the approach is the selection of an appropriate threshold for selecting relevant features from the dataset. Dash (2021) combined the Harmony Search and Pareto Optimization methods to develop a new hybrid MA feature selection method. The first 100 features are generated using the adaptive harmony search based gene selection (AHSGS) method in the first stage. In the second stage, a bi-objective Pareto optimization method was employed to reduce the gene subset further through evaluation using four different classifiers, including KNN, NB, ANN, and SVM. It was found that when paired with the SVM classifier, it outperforms other classifiers. Zare et al. (2023) achieved the maximum relevance criterion by integrating a supervised Laplace eigenmap and a matrix, and then minimized the redundancy between the selected features by applying a Pearson correlation coefficient.
Furthermore, some studies combine multiple methods for feature selection. Prabhakar and Lee (2020) proposed a tri-level feature selection method to boost prostate cancer classification accuracy. Initially, discrete wavelet transformation reduces feature count. Subsequent steps involve employing various selection methods on the simplified feature set. Experimentation highlighted the best accuracy achieved by combining the MA feature selection method, Signal Noise Ratio (SNR), and Whale Optimization Algorithm (WOA), utilizing an Artificial Neural Network (ANN) as the classifier. Mandal et al. (2021) introduced a Tri-Stage Wrapper-Filter Feature Selection Framework for Disease Classification. In the initial stage, multiple filter methods (MI, CS, RFF, XV) and classification algorithms (KNN, SVM, NB) are combined to ensure high accuracy for each feature regardless of the filter method used. In the second stage, correlation analysis (PCC) removes highly correlated features from the top k features obtained in the first stage, aiming for a maximally informative yet minimally redundant subset. Following these stages, XGBoost further refines the feature set. Lastly, a WOA-based wrapper approach finalizes the optimal feature subset. This innovative framework effectively merges wrapper and filter methods, enhancing classification accuracy while reducing computational complexity. Overall, this approach provides a novel method for disease classification, potentially improving diagnostic and therapeutic outcomes.
Each study presented in this collection highlights innovative strategies that incorporate a combination of feature selection techniques. This underscores the substantial importance of employing hybrid methods to improve accuracy and efficiency in microarray data analysis. In the landscape of microarray data analysis, the evolution of hybrid feature selection methods has proven instrumental in surmounting challenges inherent to high-dimensional datasets. Through a fusion of filter, wrapper, and ensemble techniques, these methodologies have navigated the complexities of feature selection, attaining heightened accuracy, reduced redundancy, and improved computational efficiency. While each approach brings unique insights and strengths, their convergence into hybrid methodologies reflects a pivotal stride in advancing the accuracy and applicability of microarray data analysis.
Hybrid method microarray feature selection combines multiple feature selection techniques and classifiers to improve the effectiveness of feature selection and model performance by combining the advantages of different methods. The strength of this method lies in its versatility and robustness. By combining multiple methods, it is possible to capture data features more comprehensively, improve the robustness of feature selection, and typically achieve higher classification accuracy than a single method. In addition, hybrid method feature selection has the flexibility to adapt to specific problems and data characteristics by flexibly adjusting the combined methods. However, its drawbacks include increased complexity, the need to evaluate multiple combinations, high computational cost and time-consuming, and the difficulty of optimization, which requires careful adjustment and optimization of the combinations of individual methods and classifiers, which is more difficult. For a comprehensive overview of Hybrid methods articles, refer to Table 6.
3.5 Other feature selection method on microarray
In addition to these methods, Multi-objective algorithms also play an important role in feature selection, especially when competing objectives need to be balanced. The prediction accuracy of the model and the size of the feature subset are two key objectives in the feature selection task. Traditional single-objective optimization methods usually focus on a single objective, such as maximizing the accuracy of the model, which may result in selecting too large a subset of features, increasing the computational cost and complexity of the model. The other extreme is to oversimplify the feature subset, which reduces the computational cost but may also impair the predictive performance of the model. Multi-objective algorithms are able to generate a set of Pareto-optimal solutions by simultaneously optimizing multiple objectives, each of which represents the equilibrium point where one objective cannot be further improved without degrading the other. In the process of feature selection, accuracy and feature subset size are often the two most critical and competing objectives. Multi-objective algorithms are able to consider these two factors simultaneously, providing researchers with a set of different solutions. By analyzing the Pareto frontier, researchers can achieve a better balance by choosing the most suitable feature subset among these solutions based on specific application scenarios and requirements. For instance, Cao et al. (2019) proposed a feature selection method that considers classification error, number of features, and redundancy among features based on the Distributed Parallel Collaborative Coevolutionary Multi-Objective Large-Scale Evolutionary Algorithm. To reduce calculation time, they introduced feature number constraints respectively to reduce feature input. A distributed parallel strategy is adopted to parallelize the evolution process. Adopt sample-level parallelism strategies to parallelize the testing process. Qiu (2019) developed an innovative feature selection method, MSPSO, utilizing a multi-swarm PSO algorithm. This approach subdivided the population into sub-swarms to maintain diversity, with an elite learning strategy facilitating information exchange among these sub-swarms. The experiments highlighted MSPSO's superiority over traditional PSO-based methods and popular filters in feature subset size and classification accuracy. Zhang et al. (2020) proposed a multi-objective feature selection algorithm based on binary differential evolution incorporating self-learning strategies. This algorithm embedded novel operators like binary mutation and One-bit Purifying Search to balance local exploitation and global exploration, showcasing improved performance in reducing initial feature sets' complexity. Baliarsingh et al. (2020) presented a framework called C-HMOSHSSA for gene selection in cancer classification using multi-objective meta-heuristic and machine learning methods. The proposed framework utilizes the multi-objective spotted hyena optimizer and slap swarm algorithm for gene selection, with the goal of finding a minimum subset of genes while maximizing classification accuracy. The authors conducted experiments using seven different microarray datasets to evaluate the performance of the proposed technique and compared it with existing state-of-the-art techniques. Aljarah et al. (2020) used two operators, a dynamic time-varying strategy and local fittest solutions, to improve the performance of multi-objective SSA for feature selection and used the S-shaped function to convert the improved SSA into MODSSA-Ibest, which can achieve feature selection. It can achieve faster convergence speed while avoiding local optimal solutions. Judging from the performance on both microarray datasets, features were reduced by more than 40%, and significant results were also achieved in terms of average error rate and g-mean. Dhal and Azad (2021) present a multi-objective hybrid binary version of the FS approach based on two evolutionary approaches, PSO and GWO. The approach can efficiently learn from a smaller number of samples and high-dimensional data and simultaneously considers two objectives: classification error rate and the number of features. The paper introduces a novel concept, population factor, for generating the population and a modified version of the velocity update equation based on Newton's second law of motion. The search space is divided into two phases, global and local search, and the efficacy of the method is evaluated using benchmark high-dimensional datasets. Han et al. (2021) proposed a new feature selection method based on an adaptive strategy multi-objective particle swarm optimization algorithm called MOPSO-ASFS. MOPSO-ASFS uses the PBI decomposition method to adaptively provide different penalty values for each weight vector so that more optimal solutions are retained on the Pareto front. Zhou et al. (2021) proposed an evolutionary multi-objective optimization framework of discretization-based feature selection for classification. Many heuristic search methods can be used in this framework; they take PSO, for example, as the search method; to obtain the Pareto solutions, a flexible cut-point PSO is introduced to help better explore relevant subsets of features. Moslemi and Ahmadian (2023) developed a new feature selection method based on rank constrained and dual regularized nonnegative matrix factorization, which outperforms the latest unsupervised feature selection techniques in multiple mediums in terms of clustering accuracy and normalized mutual information. Analogously, Samareh-Jahani et al. (2024) developed a low-redundancy unsupervised feature selection method based on data structure learning and feature orthogonalization, which first uses QR decomposition to obtain an orthogonal representation of the feature space, and then determines the distance between the feature set and the orthogonal set obtained from the original features based on a matrix decomposition. Also, Saberi-Movahed et al. (2024) proposed a deep non-negative matrix factorization method by combining global and local structures that preserves both global and local structures in the data space. Furthermore, regularization terms that promote sparsity by exploiting the notion of inner product are applied to represent matrices of lower dimensions as a way to preserve the underlying data structure while discarding less important features. Sheikhpour et al. (2025) proposed a feature selection method expressed in the form of trace ratios, which utilizes the discriminative information of labeled data to maximize class separability, as well as the hypergraph Laplace operator to capture geometric structure and higher-order relationships in labeled and unlabeled data. Lv et al. (2021) proposed an innovative framework, SFAM, that combines adaptive global structure learning and stream shape learning with the aim of improving the effectiveness of semi-supervised multi-label feature selection. The framework overcomes the limitations of existing methods in dealing with label correlation by utilizing both local and global data structures. The authors also develop an efficient iterative optimization algorithm to address the non-smooth objective function of the model.
The multi-objective method to feature selection finds the best subset of features by simultaneously optimizing multiple objectives (for example, classification accuracy and number of features). It is characterized by multi-objective optimization, which provides more comprehensive feature selection results and efficiently handles conflicts and trade-offs between different objectives through the optimization algorithm. This approach is widely applicable and can be applied to a variety of data types and tasks with strong adaptability. The advantages of the multi-objective approach include the ability to optimize on multiple performance metrics at the same time, providing more comprehensive and effective results than the single-objective approach. It is flexible and applicable to complex tasks that need to balance multiple performance requirements, and can adjust the optimization objectives according to different needs. In addition, through search methods such as evolutionary algorithms, the multi-objective approach may find the global optimal or near-optimal subset of features. However, the computational complexity of this approach is high, especially when dealing with high-dimensional data, which requires larger computational resources and time. Implementation complexity is also a major challenge, requiring a deep understanding of the principles and methods of multi-objective optimization. In addition, multi-objective optimization produces a potentially large set of solutions (Pareto front), and selecting the best solution and interpreting its significance may be more difficult.
Besides, there are also some less commonly used methods. Zhang et al. (2021) proposed a feature selection method based on information-theoretic lower bounds of feature inner correlations for high-dimensional data. The authors introduce two lower bounds for feature redundancy and complementarity, which have simple forms and are closer to the optima than existing lower bounds used by some state-of-the-art information-theoretic methods. They then propose a simple and effective feature selection method based on these lower bounds and verify its effectiveness with a wide range of real-world datasets. Xie et al. (2024) proposed a graph neural network-based feature selection algorithm with a classification model to achieve feature selection. They use a multidimensional graph to represent interactions between genes, utilize link prediction techniques to enrich existing graph structure relationships, and use a multidimensional node evaluator and a spectral clustering-based supernode discovery algorithm to achieve initial screening of nodes. Subsequently, we further screen the nodes using downsampling-based hierarchical graph pooling techniques to achieve feature selection and build classification models.
3.6 Summary of different methods
This section will present some summaries on microarray feature selection, including the number of articles on different methods in recent years, the classifiers used by different methods and the classification accuracies of various methods.
3.6.1 Comparison of different methods
Filter method is a model-independent feature selection technique that performs feature selection by calculating the correlation or amount of information between the features and the target variable. The main advantage of this method is that it is computationally efficient and suitable for large-scale datasets because it does not require model training for each feature combination. Meanwhile, the filter method is highly interpretable, and the feature selection process is intuitive and easy to understand and implement. However, the disadvantage of the filtering method is that it tends to ignore the interrelationships between features and relies only on the correlation of individual features with the target variable, which may leave out certain important features. This neglect may result in compromising the performance of the model in cases where there are important interactions between features, thus limiting its predictive power in practical applications.
Wrapper method is better able to capture the complex relationships between features by training the model while selecting features and considering the interactions between features. This method typically provides superior feature selection results because it evaluates the effectiveness of feature combinations based on the performance of the model. However, the computational overhead of the wrapper method is high because the model needs to be trained for each feature selection, especially on large datasets, which can significantly increase the consumption of computational resources. In addition, repeatedly evaluating the model performance may lead to overfitting. The reason is that it evaluates the performance of feature combinations by repeatedly training the model. During the feature selection process, the wrapper method constantly adjusts the feature subset based on the performance of the training set, and this high-frequency model evaluation may result in a model that overfits the noise and features of the training data, thus performing well on the training set but having reduced generalization ability on new, unseen data. Especially in the case of a small sample size of the training set, the model's over-reliance on a specific combination of features may make the selected features not representative, thus decreasing the prediction accuracy in real-world applications.
Embedded method combines the advantages of the filter and wrapper methods by automatically selecting features during the model training process. This method takes into account the interactions between features and is relatively efficient because feature selection is synchronized with model training. The embedded method is able to reduce the computational complexity and usually yields better feature selection results. However, its drawback is that it strongly depends on the selected model, this is because it integrates the feature selection process directly into the training of the model, making the assessment of the importance of features dependent on the model algorithm used. Different models, use different criteria to assess the importance of features, which means that the feature selection results can vary from model to model. In addition, the effectiveness of the embedded method is closely linked to the generalization ability of the model, and if the selected model performs well on a specific dataset but poorly on other datasets, the results of feature selection may also lack generalization. This dependency not only affects feature selection, but also requires that when faced with a new problem or dataset, feature selection may need to be redone to accommodate the new model configuration. As well, embedded methods usually require tuning the hyperparameters of the model to optimize performance, and different hyperparameter settings can also lead to variations in the feature selection results. Therefore, when using embedding methods, researchers and practitioners need to have a deep understanding of the characteristics and behaviors of the models used to ensure that the selected features can effectively support model learning and prediction, and to avoid performance degradation or improper selection due to model dependency.
Hybrid method combines the advantages of the filter and wrapper methods by first performing initial feature screening through the filter method to quickly exclude irrelevant features, and then performing refined selection through the wrapper method. The advantage of this approach is that it increases both the efficiency of feature selection and the accuracy of the final feature set. However, the hybrid approach is more complex to implement and requires coordinating the implementation of the two methods, which can lead to misconfigurations. In addition, the use of wrapper methods may still consume significant computational resources on large datasets, despite the fact that the initial screening reduces the number of features. The complex implementation process may lead to irrational feature selection, which affects the performance of the model, especially when the features are poorly selected, which may result in the model not being able to learn the structure of the data efficiently.
3.6.2 Method selection guidelines
In real-world biomedical applications, the choice of an appropriate feature selection method strongly depends on the dataset characteristics and practical constraints. For example, filter-based methods are usually preferred when dealing with high and small sample datasets due to their simplicity and scalability. Wrapper methods, while usually yielding higher accuracy, may not be suitable for large datasets due to their high computational cost. Embedded methods provide a compromise by integrating model training and feature selection, making them more popular in scenarios where classifier performance is critical. Hybrid methods are especially valuable when both selection quality and computational feasibility are required. These practical considerations are crucial when applying feature selection to tasks such as personalized medicine or early cancer diagnosis.
When selecting an appropriate feature selection method for microarray datasets, which typically exhibit high dimensionality, low sample size, and class imbalance, it is essential to make targeted decisions based on the specific characteristics of the data. For datasets with extremely small sample sizes and extremely high feature dimensions, such as Leukemia, Colon, and Prostate, the Filter method is recommended. This method is computationally efficient, relatively robust to small samples, and can quickly eliminate a large number of redundant features. Additionally, the Filter method does not depend on specific learners and is suitable as a pre-screening step in the first phase to reduce the difficulty of subsequent modeling.
In datasets with significant class imbalance, such as Leukemia1, MLL, and CNS, Hybrid methods or Wrapper methods with class-aware mechanisms perform more stably. Especially under feature score bias caused by class imbalance, Hybrid methods can effectively mitigate bias by combining independent scoring with model feedback. For multi-class datasets, such as SRBCT, Lymphoma and Leukemia2, it is important to consider the method's support for multi-class discrimination capabilities. In such tasks, embedded methods like LASSO and tree models are more suitable. These methods can dynamically adjust feature importance during training based on the objective function and effectively account for inter-class differences, adapting to multi-class structures.
In summary, different feature selection methods have their own advantages on different types of microarray datasets. Filter methods are suitable for datasets with high feature redundancy and severe small sample problems; Hybrid and Wrapper methods demonstrate high adaptability when dealing with class imbalance, and Embedded methods are suitable for multi-class classification or tasks with high requirements for feature interdependencies. By reasonably combining different methods, more stable and interpretable results can be achieved in various data scenarios.
3.6.3 Number of papers with different methods
The number of papers for every year can be found in Figure 5. The research on feature selection for microarray data has experienced a trend of first decline and then rise in the number of articles, mainly due to the fact that initially researchers focused on simple feature selection methods, while with the maturity of the technology and the rise of deep learning and integration methods, the research has gradually shifted to more complex techniques, which led to a decline in the number of studies on traditional methods. However, in recent years, with the improvement of computational power and the increase of data complexity, the research on feature selection has become active again, and the number of related literatures has risen rapidly, especially driven by new technologies. This change is attributed to technology iteration, data complexity, and cross-fertilization between fields such as bioinformatics and computer science. Studying feature selection can not only significantly improve model performance, reduce computational resource consumption and risk of overfitting, but also help to extract biologically significant and important features, facilitate the understanding of disease mechanisms, and promote the development of new algorithms and models. Therefore, the study of feature selection is of great significance for the progress and innovation in the field of microarray data analysis.
In conclusion, while hybrid methods dominate the landscape of microarray feature selection, challenges remain in achieving comprehensive and interpretable results. Continued research efforts focused on improving classification accuracy, reducing dimensionality, and enhancing interpretability are essential to unlock the full potential of microarray feature selection in biomedical applications.
3.6.4 Classification results
Classifiers play a pivotal role in feature selection by serving as the core component for evaluating and selecting feature subsets. They aid in identifying which features are most conducive to predicting the target variable by training on the training set and assessing their performance. The performance of classifiers frequently serves as a criterion for selecting feature subsets, and they are also utilized to guide the optimization of feature subsets. This ensures that the chosen feature subset enhances performance of the model and generalization capabilities. We have summarized the frequency with which different methods employ various classifiers, as depicted in Figure 6. The figure illustrates that KNN and SVM are frequently employed as classifiers, likely due to their robustness and generalization capabilities, making them effective in handling high-dimensional data. The KNN classifier is known for its simplicity and ease of implementation. It operates by identifying the K instances in the training dataset that are closest to the new sample and making predictions based on their majority class. This method is particularly suitable for processing nonlinear data and scenarios with numerous outliers. On the other hand, SVM stands out as a powerful supervised learning algorithm that excels in separating different categories of data by identifying a hyperplane that maximizes the classification margin. It is adept at handling both linearly separable and inseparable problems, and can be extended to address nonlinear challenges through kernel techniques. Given these attributes, KNN and SVM have become staples in microarray feature selection. They assist researchers in extracting valuable insights from intricate gene expression data, consequently enhancing prediction accuracy and the biological interpretability of models. Their versatility and effectiveness make them indispensable tools in the pursuit of understanding and leveraging genetic information for various applications in biomedicine and beyond.
The classification results obtained from microarray datasets hold significant implications for disease diagnosis, biomarker discovery, drug development, understanding disease mechanisms, and advancing personalized medicine. Accurate classification of microarray data unveils the relationship between gene expression patterns and biological states, furnishing a scientific foundation for medical decision-making, fostering precision medicine development, and facilitating profound biomedical research endeavors. Thus, ensuring the precision of classification outcomes for microarray datasets is imperative for research and clinical applications in related domains. This study compiled data on the classification accuracy of various methods applied to microarray feature selection, utilizing commonly employed datasets. The statistical classification accuracy of microarray feature selection methods is delineated in Tables 7, 8. In both tables, “–” means that the feature selection method is not tested on this dataset. Furthermore, in addition to assessing classification accuracy, some methods use the dimensionality of the selected feature subset as an evaluation criterion for the feature selection process. The entries highlighted in bold in the tables represent the methods that produce the highest performance on the corresponding dataset. These tables are valuable resources for researchers and practitioners to gain insight into the efficacy of different feature selection methods in accurately classifying microarray data. By utilizing this information, informed decisions can be made regarding the selection and implementation of feature selection techniques to advance the field of microarray data analysis and its diverse applications in biomedical and other fields.
This chapter explores different feature selection methods for microarray data analysis, including filter, wrapper, embedded, hybrid and other methods, evaluating their strengths and weaknesses. Hybrid methods, combining various techniques, have emerged as powerful tools, enhancing model performance and interpretability. With technology advancements and increasing data volume, hybrid methods are expected to play a key role in future research, offering adaptable solutions to address evolving challenges in microarray data analysis and biomedical research.
4 Microarray feature selection application
Microarray feature selection technology holds a central position in disease classification and diagnosis. By scrutinizing patient gene expression data through feature selection, it markedly enhances early diagnostic precision, identifies disease subtypes, discovers biomarkers, and predicts drug sensitivity. The successful application of this high-throughput analysis method, notably in cancer research, has significantly influenced both medical research and clinical practice.
In real-world biomedical research, feature selection methods are often chosen with more than just classification accuracy concerns. Researchers must also balance the computational efficiency of the algorithm, the cost of implementation, and the interpretability of the results. For example, in time-sensitive clinical environments such as cancer subtype prediction or diagnostic screening, the ability to obtain results quickly may be more important than small improvements in accuracy. Therefore, an algorithm that can provide acceptable accuracy in a shorter period of time may be preferred to some complex but computationally expensive methods.
Data distribution is also a key consideration in real-world applications, especially in scenarios with small sample sizes or severe category imbalances. For example, in disease prediction tasks, the number of positive cases is usually small. In such cases, distance-based feature selection methods may perform poorly due to neighborhood structure bias. In contrast, some hybrid or embedded methods that incorporate category prior information or regularization strategies tend to have better stability. Therefore, in specific applications, in addition to the method category, its robustness to sample bias is also an important criterion for method selection.
Therefore, the adoption of feature selection methods in the real application often requires a trade-off between algorithm complexity, selection accuracy and scalability. This trade-off is especially critical when translating computational research results into clinical practice, where clinical environments often have practical constraints on time, interpretability, and compatibility with downstream analysis tools. Taking these factors into account can help in choosing feature selection methods that have both theoretical strengths and practical needs.
Several advanced methodologies have been proposed for microarray feature selection and classification. Rochayani et al. (2020) introduced a two-stage method employing the Lasso regularization method followed by Classification and Regression Trees for further refinement and classification. Xie et al. (2022b) proposed a feature selection algorithm and classification model grounded in graph neural networks, overcoming existing method limitations by enriching graph structural relationships via link prediction techniques. Wu et al. (2022) utilized XGBoost followed by the gray wolf algorithm to pinpoint the optimal gene subset for cancer classification. Wang et al. (2020) devised a novel feature selection approach within the ensemble learning framework, corroborating its robustness through multiple aggregation methods. Zare et al. (2023) advocated a supervised feature selection approach based on manifold learning, integrating Supervised Laplacian eigenmaps and matrices for comprehensive feature selection. Prajapati et al. (2023a) employed ant colony optimization in tandem with logistic regression, decision tree, and random forest for exhaustive feature selection and classification accuracy comparison. Prajapati et al. (2023c) used a genetic algorithm for feature selection in combination with classification algorithms such as logistic regression, decision tree, and random forest to detect cancer, tumors and various other diseases. Sahu and Dash (2024) proposed a hybrid FS model based on the Jaya optimization algorithm and information gain, which verified the effectiveness of IG technology in feature selection. Additionally, there are many cases where microarray feature selection has been used in disease subtype diagnosis. Mehrabani et al. (2022) used microarray gene expression data from 72 patients with acute myeloid leukemia (AML) and lymphoblastic leukemia (ALL), and the RF and SVM classifiers correctly classified all AML and ALL samples.
Disease subtypes play a significant role in disease classification and treatment choice. Understanding disease subtypes supports personalized medicine and tailored treatment strategies. Maulik et al. (2013) demonstrated the effectiveness of feature selection and transductive SVM in predicting cancer subtypes. Roberts et al. (2018) distinguished clinically relevant cancer subtypes using a differential variance classifier, with combined methods yielding superior results. Wang et al. (2023b) validated the efficacy of feature selection and Bayesian networks in identifying protein biomarkers for cancer subtypes.
Biomarkers serve as crucial indicators in disease diagnosis, monitoring, and assessing treatment effectiveness. Trevizan and Recamonde-Mendoza (2021) proposed Ensemble Feature Selection for identifying potential breast cancer biomarkers. Colombelli et al. (2022) developed a hybrid ensemble feature selection design to enhance the reproducibility of genomic biomarker discovery. Xie et al. (2022a) introduced a novel biomarker selection method, demonstrating its effectiveness in feature reduction and classification accuracy improvement. Alzubaidi et al. (2022) addressed challenges in breast cancer staging by developing a deep learning-based feature extraction module for identifying robust biomarkers. Ge (2023) proposed FSRL for identifying potential biomarkers for various high-mortality cancers, demonstrating superior classification accuracy and computational efficiency.
Drug sensitivity prediction employs various methods and techniques to anticipate an individual's response to specific medications, leveraging their biological characteristics, genomic information, or other biomarkers. This predictive approach serves the goal of personalized medicine in the medical field, striving to maximize drug treatment efficacy while minimizing adverse reactions. Microarray technology plays a crucial role in this endeavor by collecting gene expression data from individual samples and scrutinizing the correlation between this data and drug responses. For instance, Chen and Sun (2017) devised a novel method for high-dimensional dual-layer feature selection, utilizing a set of response variables that share a standard set of predictive variables. Simulation results indicate heightened sensitivity and specificity compared to existing methods. Meanwhile, Ahmed et al. (2020) described a network-based approach for identifying features in drug response prediction. They employed a gene co-expression network to pinpoint representative features and proposed a graph neural network model integrating gene network information for outcome prediction. Koras et al. (2020) introduced a prior-knowledge-driven feature selection method grounded in drug targets, target pathways, and gene expression features. Validation underscored the importance of selecting appropriate feature selection strategies, particularly for drugs targeting specific genes, pathways, or affecting general mechanisms such as immune response and DNA replication. These models show promise in guiding treatment design. Ataei et al. (2021) initially employed gene fuzzy score and principal component analysis to reduce data dimensions, followed by SVM classification of sensitive and resistant data samples. Subsequent Wilcoxon Rank Sum tests determined differentially expressed genes, contributing to the understanding of drug sensitivity mechanisms. Yang et al. (2022) proposed a cancer drug sensitivity prediction model based on multi-omics data constructed using stacked ensemble learning methods. Through functional annotation and enrichment analysis of feature genes, they elucidated potential resistance mechanisms of tumors to sorafenib, substantiating the model's interpretability from a biological perspective. This model holds promise in guiding clinical drug usage.
Given the diverse application scenarios outlined above, selecting an appropriate feature selection strategy must be tailored to the specific goals and data characteristics of each task. In microarray data analysis, the objective of feature selection is not singular but closely tied to the problem being addressed. For disease subtype classification, classification accuracy and generalization capability are critical, especially as the model needs to capture subtle yet important expression differences. Hence, wrapper or embedded methods are commonly adopted, as they can adaptively optimize feature subsets based on model performance feedback. These approaches are better suited for capturing nonlinear relationships and dealing with the challenges of high dimensionality and limited sample sizes. Some studies have further enhanced performance by incorporating transfer learning and graph-based structures to model complex biological dependencies.
In contrast, biomarker discovery emphasizes the stability, reproducibility, and biological interpretability of selected features. In such cases, filter methods are often favored due to their reduced dependency on specific classifiers and increased robustness across datasets. Recently, ensemble strategies and multi-criteria fusion techniques have gained popularity. These combine multiple scoring metrics or selection algorithms to ensure that the resulting biomarkers are both statistically significant and biologically meaningful.
For drug sensitivity prediction, the goals extend beyond classification accuracy to include interpretability and generalizability across diverse biological conditions. Since drug responses often involve intricate molecular mechanisms and multi-omics interactions, this domain frequently employs network-based analysis, embedding methods, and automated feature engineering techniques. These are often combined with ensemble learning or multi-layer integration models to enhance predictive performance. Furthermore, dimensionality reduction techniques such as PCA are commonly used during preprocessing, followed by supervised feature evaluation, to maintain both model robustness and biological interpretability.
In conclusion, microarray feature selection is integral to disease diagnosis, biomarker discovery, and drug sensitivity prediction. Its application in disease subtype diagnosis, biomarker discovery, and drug sensitivity prediction underscores its significance in advancing personalized medicine and improving treatment outcomes. Figure 7 illustrates a schematic diagram of a word cloud generated by the application of microarray feature selection in various fields. The prominent keywords include “gene,” “feature selection,” “identify,” “disease,” and “biology.” This visualization underscores the significance of microarray feature selection in bioinformatics research. It plays a pivotal role in selecting valuable information from complex gene expression data, thereby advancing biomedical research and contributing to the understanding and treatment of diseases.
The application trends in microarray feature selection are primarily characterized by the integration of deep learning technologies, the incorporation of multi-modal data integration, standardization and annotation, the pursuit of interpretability and biological relevance, the expansion into clinical applications, and the emphasis on privacy protection and security. These trends highlight the rapid evolution and shifting research focus within the field of microarray feature selection, foreshadowing both opportunities and challenges ahead. As technology advances and data volumes grow, these trends are expected to continue evolving, bringing forth more innovations and breakthroughs in the field of microarray data analysis.
5 Summary and future work
This study systematically reviews the literature on microarray feature selection and explore its significance in both academic and practical fields. By integrating existing studies, we aim to help readers gain a comprehensive understanding of the overall development of microarray feature selection, including the strengths and weaknesses of different approaches and their applicability scenarios. We identify gaps in current research and point out areas that have not yet been fully explored to provide clear directions for subsequent research, thus stimulating the academic community to explore new techniques and applications.
In addition, this review provides a comprehensive assessment of various feature selection methods, aiming to provide researchers with theoretical foundations and practical guidance in selecting methods suitable for their specific research questions, in order to promote the optimization of existing techniques and the development of new methods. We emphasize the importance of fostering communication and collaboration between multiple fields, including bioinformatics, computer science, and statistics, to help researchers draw on best practices from other disciplines to further advance microarray analysis techniques.
Finally, by emphasizing the importance of feature selection in real-world applications such as personalized medicine, cancer diagnosis, and drug discovery, we hope to enhance the understanding of the value of these techniques among industry practitioners, and thus promote their implementation in practice. In summary, this review not only provides theoretical support for the academic community, but also provides practical guidance for the practical field, significantly contributing to the overall improvement of microarray data analysis techniques.
With the development of technology, the field of microarray feature selection is facing unprecedented innovation opportunities, and its future development will focus on the dual breakthroughs of technological innovation and practical applications. At the technical level, the deep integration of deep learning and feature selection will become an important breakthrough. By building a deep feature selection framework with adaptive capability, researchers can automatically capture high-order nonlinear feature interactions in the data and significantly improve the accuracy and efficiency of feature selection. Meanwhile, the optimization innovation of integrated learning methods will promote the establishment of multi-algorithm collaborative selection mechanism, which will realize the synergistic enhancement of the stability and generalization of the feature selection results through intelligent weighted fusion and dynamic voting strategies. It should be noted that, while performing performance breakthroughs, interpretability has become a key bottleneck in the development of this field. The “black-box” nature of the current deep feature selection model severely restricts its application in clinical practice, so there is an urgent need to develop new algorithms with both high performance and interpretability, as well as a framework for evaluating the importance of features by integrating causal reasoning, to provide a transparent and traceable scientific basis for biomedical decision-making.
In application expansion, microarray feature selection will play a greater role in the future in the fields of precision medicine, drug discovery and multi-omics analysis. As research shifts from static classification to dynamic prediction and mechanism exploration, feature selection will help model disease processes and develop individualized treatment strategies. The fusion of multi-omics data will promote the systematic understanding of complex disease mechanisms, while in drug development, feature selection will accelerate target identification and drug response prediction. In addition, the technology will be expanded to systems biology, environmental monitoring and other emerging fields to support the in-depth analysis of complex biological systems and ecological factors. In the future, microarray feature selection is expected to become an important tool for data-driven knowledge discovery, promoting the transformation of life science research into intelligent and systematic.
It is worth noting that in recent years, with the rapid development of reinforcement learning (RL) and large language models (LLMs), their potential in microarray feature selection has gradually attracted attention. Reinforcement learning models feature selection as a sequential decision-making process, where an agent dynamically adjusts the selected feature subset based on feedback. This not only improves selection efficiency and has strong generalization capabilities. For example, the reinforcement learning-based automated feature selection framework proposed by Liu et al. (2021) demonstrates better robustness and selection stability on multiple high-dimensional datasets. Fan et al. (2020)'s AutoFS design integrates diversity reward mechanisms and interactive reinforcement learning strategies, enhancing interpretability while maintaining performance. Additionally, multi-agent collaborative selection methods have also achieved outstanding results in feature subspace exploration (Liu et al., 2019).
On the other hand, the combination of large language models and structured data analysis also shows great potential. The CAAFE framework proposed by Hollmann et al. (2023) integrates LLM with tabular predictors to achieve an integrated process of feature construction, model guidance, and interpretation. This framework can combine biological background knowledge with natural language instructions in practical applications, assisting researchers in efficiently identifying biologically meaningful features from complex gene expression data. The integration of these technologies not only provides smarter and more automated tools for feature selection, which also expands new possibilities for cross-modal data analysis and human-machine collaborative modeling.
While this review systematically compares and synthesizes experimental results reported in previous literature, we acknowledge that this study has not yet conducted new empirical benchmarking experiments. As this paper is a literature-focused review, its scope and focus dictate that we primarily rely on existing experimental results to draw comparative conclusions. However, we recognize the importance of standardized, unified benchmarking across different datasets and methods and plan to incorporate such comparative assessments in future research. By conducting research under consistent experimental conditions, we aim to provide a more objective assessment of the strengths and weaknesses of each method, thereby offering deeper empirical insights into microarray feature selection techniques.
In the future, microarray feature selection technology will continue to develop under the dual-wheel drive of algorithm innovation and application expansion. On the one hand, with the introduction of cutting-edge technologies such as interpretable AI and causal inference, the feature selection process will be more transparent and reliable; on the other hand, its in-depth application in the fields of precision medicine and drug discovery will continue to promote the transformation of biomedical research into a new paradigm of data-driven and knowledge discovery. These advances will significantly enhance the analytical value of microarray data, and will revolutionize human health research and clinical practice.
Author contributions
FW: Data curation, Methodology, Writing – review & editing, Writing – original draft. AZ: Supervision, Writing – review & editing, Project administration. YR: Writing – original draft. MB: Writing – review & editing. AS: Writing – original draft. ZA: Writing – review & editing. NY: Writing – original draft, Data curation. RJ: Funding acquisition, Writing – original draft. AM: Resources, Formal analysis, Writing – review & editing. NA: Writing – review & editing, Funding acquisition.
Funding
The author(s) declare that no financial support was received for the research and/or publication of this article.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Gen AI was used in the creation of this manuscript.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Abdel-Basset, M., Ding, W., and El-Shahat, D. (2021). A hybrid Harris Hawks optimization algorithm with simulated annealing for feature selection. Artif. Intell. Rev. 54, 593–637. doi: 10.1007/s10462-020-09860-3
Agrawal, R. K., Kaur, B., and Sharma, S. (2020). Quantum based whale optimization algorithm for wrapper feature selection. Appl. Soft Comput. 89:106092. doi: 10.1016/j.asoc.2020.106092
Ahmed, K. T., Park, S., Jiang, Q., Yeu, Y., Hwang, T., Zhang, W., et al. (2020). Network-based drug sensitivity prediction. BMC Med. Genomics 13:193. doi: 10.1186/s12920-020-00829-3
Alanni, R., Hou, J., Azzawi, H., and Xiang, Y. (2019). A novel gene selection algorithm for cancer classification using microarray datasets. BMC Med. Genomics 12:10. doi: 10.1186/s12920-018-0447-6
Alhenawi, E., Al-Sayyed, R., Hudaib, A., and Mirjalili, S. (2022). Feature selection methods on gene expression microarray data for cancer classification: a systematic review. Comput. Biol. Med. 140:105051. doi: 10.1016/j.compbiomed.2021.105051
Alhenawi, E., Al-Sayyed, R., Hudaib, A., and Mirjalili, S. (2023). Improved intelligent water drop-based hybrid feature selection method for microarray data processing. Comput. Biol. Chem. 103:107809. doi: 10.1016/j.compbiolchem.2022.107809
Ali, W., and Saeed, F. (2023). Hybrid filter and genetic algorithm-based feature selection for improving cancer classification in high-dimensional microarray data. Processes 11:562. doi: 10.3390/pr11020562
Aljarah, I., Habib, M., Faris, H., Al-Madi, N., Heidari, A. A., Mafarja, M., et al. (2020). A dynamic locality multi-objective salp swarm algorithm for feature selection. Comput. Ind. Eng. 147:106628. doi: 10.1016/j.cie.2020.106628
Almugren, N., and Alshamlan, H. (2019). “FF-SVM: new firefly-based gene selection algorithm for microarray cancer classification,” in 2019 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB) (Siena: IEEE), 1–6. doi: 10.1109/CIBCB.2019.8791236
Alomari, O. A., Makhadmeh, S. N., Al-Betar, M. A., Alyasseri, Z. A. A., Doush, I. A., Abasi, A. K., et al. (2021). Gene selection for microarray data classification based on Gray Wolf Optimizer enhanced with TRIZ-inspired operators. Knowl.-Based Syst. 223:107034. doi: 10.1016/j.knosys.2021.107034
Alzubaidi, A., Tepper, J., Inden, B., and Lotfi, A. (2022). mRNA biomarkers for invasive breast cancer based on a deep feature selection approach. J. Biomed. Res. Environ. Sci. 3, 1163–1176. doi: 10.37871/jbres1572
Ataei, A., Arab, S. S., Zahiri, J., Rajabpour, A., Kletenkov, K., Rizvanov, A., et al. (2021). Filtering of the gene signature as the predictors of cisplatin-resistance in ovarian cancer. Iran. J. Biotechnol. 19, 48–59. doi: 10.30498/ijb.2021.209370.2643
Bae, J. H., Kim, M., Lim, J., and Geem, Z. W. (2021). Feature selection for colon cancer detection using k-means clustering and modified harmony search algorithm. Mathematics 9:570. doi: 10.3390/math9050570
Baliarsingh, S. K., Vipsita, S., Gandomi, A. H., Panda, A., Bakshi, S., Ramasubbareddy, S., et al. (2020). Analysis of high-dimensional genomic data using MapReduce based probabilistic neural network. Comput. Methods Programs Biomed. 195:105625. doi: 10.1016/j.cmpb.2020.105625
Cao, B., Zhao, J., Yang, P., Yang, P., Liu, X., Qi, J., et al. (2019). Multiobjective feature selection for microarray data via distributed parallel algorithms. Future Gener. Comput. Syst. 100, 952–981. doi: 10.1016/j.future.2019.02.030
Chatterjee, B., Bhattacharyya, T., Ghosh, K. K., Singh, P. K., Geem, Z. W., Sarkar, R., et al. (2020). Late acceptance hill climbing based social SKI driver algorithm for feature selection. IEEE Access 8, 75393–75408. doi: 10.1109/ACCESS.2020.2988157
Chen, C.-W., Tsai, Y.-H., Chang, F.-R., and Lin, W.-C. (2020). Ensemble feature selection in medical datasets: combining filter, wrapper, and embedded feature selection results. Expert Syst. 37:e12553. doi: 10.1111/exsy.12553
Chen, T. H., and Sun, W. (2017). Prediction of cancer drug sensitivity using high-dimensional omic features. Biostatistics 18, 1–14. doi: 10.1093/biostatistics/kxw022
Climente-González, H., Azencott, C.-A., Kaski, S., and Yamada, M. (2019). Block HSIC Lasso: model-free biomarker detection for ultra-high dimensional data. Bioinformatics 35, i427–i435. doi: 10.1093/bioinformatics/btz333
Colombelli, F., Kowalski, T. W., and Recamonde-Mendoza, M. (2022). A hybrid ensemble feature selection design for candidate biomarkers discovery from transcriptome profiles. Knowl.-Based Syst. 254:109655. doi: 10.1016/j.knosys.2022.109655
Dabba, A., Tari, A., and Meftali, S. (2021a). Hybridization of Moth flame optimization algorithm and quantum computing for gene selection in microarray data. J. Ambient Intell. Humaniz. Comput. 12, 2731–2750. doi: 10.1007/s12652-020-02434-9
Dabba, A., Tari, A., Meftali, S., and Mokhtari, R. (2021b). Gene selection and classification of microarray data method based on mutual information and moth flame algorithm. Expert Syst. Appl. 166:114012. doi: 10.1016/j.eswa.2020.114012
Das, H., Naik, B., and Behera, H. S. (2022). A Jaya algorithm based wrapper method for optimal feature selection in supervised classification. J. King Saud Univ. Comput. Inf. Sci. 34(6, Part B), 3851–63. doi: 10.1016/j.jksuci.2020.05.002
Dash, R. (2021). An adaptive harmony search approach for gene selection and classification of high dimensional medical data. J. King Saud Univ. Comput. Inf. Sci. 33, 195–207. doi: 10.1016/j.jksuci.2018.02.013
Dash, R., Dash, R., and Rautray, R. (2022). An evolutionary framework based microarray gene selection and classification approach using binary shuffled frog leaping algorithm. J. King Saud Univ. Comput. Inf. Sci. 34, 880–891. doi: 10.1016/j.jksuci.2019.04.002
Dhal, P., and Azad, C. (2021). A multi-objective feature selection method using Newton's law based PSO with GWO. Appl. Soft Comput. 107:107394. doi: 10.1016/j.asoc.2021.107394
Dhal, P., and Azad, C. (2022). A comprehensive survey on feature selection in the various fields of machine learning. Appl. Intell. 52, 4543–4581. doi: 10.1007/s10489-021-02550-9
Effrosynidis, D., and Arampatzis, A. (2021). An evaluation of feature selection methods for environmental data. Ecol. Inform. 61:101224. doi: 10.1016/j.ecoinf.2021.101224
Fadhil, J., and Abdulazeez, A. M. (2024). Classification of cancer microarray data based on deep learning: a review. Indones. J. Comput. Sci. 13. doi: 10.33022/ijcs.v13i1.3711
Fan, W., Liu, K., Liu, H., Wang, P., Ge, Y., Fu, Y., et al. (2020). “Autofs: automated feature selection via diversity-aware interactive reinforcement learning,” in 2020 IEEE International Conference on Data Mining (ICDM) (Sorrento: IEEE), 1008–1013. doi: 10.1109/ICDM50108.2020.00117
Ganesh, N., Shankar, R., Čep, R., Chakraborty, S., and Kalita, K. (2023). Efficient feature selection using weighted superposition attraction optimization algorithm. Appl. Sci. 13:3223. doi: 10.3390/app13053223
Gangavarapu, T., and Patil, N. (2019). A novel filter-wrapper hybrid greedy ensemble approach optimized using the genetic algorithm to reduce the dimensionality of high-dimensional biomedical datasets. Appl. Soft Comput. 81:105538. doi: 10.1016/j.asoc.2019.105538
Ge, C. (2023). “Selection of Potential Cancer Biomarkers Based on Feature Selection Method,” in Third International Conference on Intelligent Computing and Human-Computer Interaction (ICHCI 2022). SPIE, 2023, Vol. 12509 (Guangzhou), 227–236.
Gong, L., Xie, S., Zhang, Y., Wang, M., and Wang, X. (2022). Hybrid feature selection method based on feature subset and factor analysis. IEEE Access 10, 120792–120803. doi: 10.1109/ACCESS.2022.3222812
Guha, R., Ghosh, M., Mutsuddi, S., Sarkar, R., and Mirjalili, S. (2020). Embedded chaotic whale survival algorithm for filter-wrapper feature selection. Soft Comput. 24, 12821–12843. doi: 10.1007/s00500-020-05183-1
Hambali, M. A., Oladele, T. O., and Adewole, K. S. (2020). Microarray cancer feature selection: review, challenges and research directions. Int. J. Cogn. Comput. Eng. 1, 78–97. doi: 10.1016/j.ijcce.2020.11.001
Hamla, H., and Ghanem, K. (2024). A hybrid feature selection based on fisher score and SVM-RFE for microarray data. Informatica 48:4759. doi: 10.31449/inf.v48i1.4759
Hamraz, M., Ali, A., Mashwani, W. K., Aldahmani, S., and Khan, Z. (2023). Feature selection for high dimensional microarray gene expression data via weighted signal to noise ratio. PLoS ONE 18:e0284619. doi: 10.1371/journal.pone.0284619
Han, F., Chen, W.-T., Ling, Q.-H., and Han, H. (2021). Multi-objective particle swarm optimization with adaptive strategies for feature selection. Swarm Evol. Comput. 62:100847. doi: 10.1016/j.swevo.2021.100847
He, Y., Zhou, J., Lin, Y., and Zhu, T. (2019). A class imbalance-aware relief algorithm for the classification of tumors using microarray gene expression data. Comput. Biol. Chem. 80, 121–127. doi: 10.1016/j.compbiolchem.2019.03.017
Hollmann, N., Müller, S., and Hutter, F. (2023). “Caafe: combining large language models with tabular predictors for semi-automated data science,” in 1st Workshop on the Synergy of Scientific and Machine Learning Modeling@ ICML2023.
Hosseini, E. S., and Moattar, M. H. (2019). Evolutionary feature subsets selection based on interaction information for high dimensional imbalanced data classification. Appl. Soft Comput. 82:105581. doi: 10.1016/j.asoc.2019.105581
Houssein, E. H., Abdelminaam, D. S., Hassan, H. N., Al-Sayed, M. M., and Nabil, E. (2021). A hybrid barnacles mating optimizer algorithm with support vector machines for gene selection of microarray cancer classification. IEEE Access 9, 64895–64905. doi: 10.1109/ACCESS.2021.3075942
Hu, J., Gui, W., Heidari, A. A., Cai, Z., Liang, G., Chen, H., et al. (2022). Dispersed foraging slime mould algorithm: continuous and binary variants for global optimization and wrapper-based feature selection. Knowl.-Based Syst. 237:107761. doi: 10.1016/j.knosys.2021.107761
Jain, D., and Singh, V. (2021). A two-phase hybrid approach using feature selection and adaptive SVM for chronic disease classification. Int. J. Comput. Appl. 43, 524–536. doi: 10.1080/1206212X.2019.1577534
Jeon, H., and Oh, S. (2020). Hybrid-recursive feature elimination for efficient feature selection. Appl. Sci. 10. doi: 10.3390/app10093211
Joseph, R. C., and Sandoval, N. R. (2023). Single and multiplexed gene repression in solventogenic clostridium via cas12a-based crispr interference. Synth. Syst. Biotechnol. 8, 148–156. doi: 10.1016/j.synbio.2022.12.005
Jovic, D., Liang, X., Zeng, H., Lin, L., Xu, F., Luo, Y., et al. (2022). Single-cell rna sequencing technologies and applications: a brief overview. Clin. Transl. Med. 12:e694. doi: 10.1002/ctm2.694
Kang, C., Huo, Y., Xin, L., Tian, B., and Yu, B. (2019). Feature selection and tumor classification for microarray data using relaxed Lasso and generalized multi-class support vector machine. J. Theor. Biol. 463, 77–91. doi: 10.1016/j.jtbi.2018.12.010
Karthika, M. S., Rajaguru, H., and Nair, A. R. (2023). Evaluation and exploration of machine learning and convolutional neural network classifiers in detection of lung cancer from microarray gene–a paradigm shift. Bioengineering 10:933. doi: 10.3390/bioengineering10080933
Khamparia, A., Singh, A., Anand, D., Gupta, D., Khanna, A., Arun Kumar, N., et al. (2020). A novel deep learning-based multi-model ensemble method for the prediction of neuromuscular disorders. Neural Comput. Appl. 32, 11083–11095. doi: 10.1007/s00521-018-3896-0
Kilicarslan, S., Adem, K., and Celik, M. (2020). Diagnosis and classification of cancer using hybrid model based on ReliefF and convolutional neural network. Med. Hypotheses 137:109577. doi: 10.1016/j.mehy.2020.109577
Koras, K., Juraeva, D., Kreis, J., Mazur, J., Staub, E., Szczurek, E., et al. (2020). Feature selection strategies for drug sensitivity prediction. Sci. Rep. 10:9377. doi: 10.1038/s41598-020-65927-9
Labory, J., Njomgue-Fotso, E., and Bottini, S. (2024). Benchmarking feature selection and feature extraction methods to improve the performances of machine-learning algorithms for patient classification using metabolomics biomedical data. Comput. Struct. Biotechnol. J. 23, 1274–1287. doi: 10.1016/j.csbj.2024.03.016
Lee, J., Choi, I. Y., and Jun, C.-H. (2021). An efficient multivariate feature ranking method for gene selection in high-dimensional microarray data. Expert Syst. Appl. 166:113971. doi: 10.1016/j.eswa.2020.113971
Li, C., and Xu, J. (2019). Feature selection with the Fisher score followed by the maximal clique centrality algorithm can accurately identify the hub genes of hepatocellular carcinoma. Sci. Rep. 9:17283. doi: 10.1038/s41598-019-53471-0
Liu, K., Fu, Y., Wang, P., Wu, L., Bo, R., Li, X., et al. (2019). “Automating feature subspace exploration via multi-agent reinforcement learning,” in Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (New York, NY: ACMM), 207–215. doi: 10.1145/3292500.3330868
Liu, K., Fu, Y., Wu, L., Li, X., Aggarwal, C., Xiong, H., et al. (2021). Automated feature selection: a reinforcement learning perspective. IEEE Trans. Knowl. Data Eng. 35, 2272–2284. doi: 10.1109/TKDE.2021.3115477
Loey, M., Jasim, M. W., EL-Bakry, H. M., Taha, M. H. N., and Khalifa, N. E. M. (2020). Breast and Colon Cancer Classification from Gene Expression Profiles Using Data Mining Techniques. Basel: Symmetry, MDPI. doi: 10.20944/preprints202002.0324.v1
Lopez-Rincon, A., Martinez-Archundia, M., Martinez-Ruiz, G. U., Schoenhuth, A., and Tonda, A. (2019). Automatic discovery of 100-miRNA signature for cancer classification using ensemble feature selection. BMC Bioinformatics 20:480. doi: 10.1186/s12859-019-3050-8
Lv, S., Shi, S., Wang, H., and Li, F. (2021). Semi-supervised multi-label feature selection with adaptive structure learning and manifold learning. Knowl.-Based Syst. 214:106757. doi: 10.1016/j.knosys.2021.106757
Mahesh, T., Santhakumar, D., Balajee, A., Shreenidhi, H., Kumar, V. V., Annand, J. R., et al. (2024). Hybrid ant lion mutated ant colony optimizer technique with particle swarm optimization for leukemia prediction using microarray gene data. IEEE Access 12:10910. doi: 10.1109/ACCESS.2024.3351871
Mandal, M., Singh, P. K., Ijaz, M. F., Shafi, J., and Sarkar, R. (2021). A tri-stage wrapper-filter feature selection framework for disease classification. Sensors 21:5571. doi: 10.3390/s21165571
Maolmhuaidh, F. Ó., Thantilage, R. D., Burtenshaw, D., Raja, K., Le-Khac, N.-A., Cahill, P., et al. (2023). Clinical data warehouses for disease diagnosis: a focus on exosomal microrna biomarkers. Inform. Med. Unlocked 43:101390. doi: 10.1016/j.imu.2023.101390
Maulik, U., Mukhopadhyay, A., and Chakraborty, D. (2013). Gene-expression-based cancer subtypes prediction through feature selection and transductive SVM. IEEE Trans. Biomed. Eng. 60, 1111–1117. doi: 10.1109/TBME.2012.2225622
Mehrabani, S., Soroush, M. Z., Kheiri, N., Sheikhpour, R., and Bahrami, M. (2022). Prediction of blood cancer using leukemia gene expression data and sparsity-based gene selection methods. Iran. J. Pediatr. Hematol. Oncol. 13, 13–21. doi: 10.18502/ijpho.v13i1.11629
Moses, L., and Pachter, L. (2022). Museum of spatial transcriptomics. Nat. Methods 19, 534–546. doi: 10.1038/s41592-022-01409-2
Moslemi, A., and Ahmadian, A. (2023). Dual regularized subspace learning using adaptive graph learning and rank constraint: unsupervised feature selection on gene expression microarray datasets. Comput. Biol. Med. 167:107659. doi: 10.1016/j.compbiomed.2023.107659
Nematzadeh, H., Enayatifar, R., Mahmud, M., and Akbari, E. (2019). Frequency based feature selection method using whale algorithm. Genomics 111, 1946–1955. doi: 10.1016/j.ygeno.2019.01.006
Osama, S., Shaban, H., and Ali, A. A. (2023). Gene reduction and machine learning algorithms for cancer classification based on microarray gene expression data: a comprehensive review. Expert Syst. Appl. 213:118946. doi: 10.1016/j.eswa.2022.118946
Panda, M. (2020). Elephant search optimization combined with deep neural network for microarray data analysis. J. King Saud Univ. Comput. Inf. Sci. 32, 940–948. doi: 10.1016/j.jksuci.2017.12.002
Pashaei, E., and Pashaei, E. (2019). “Gene selection using intelligent dynamic genetic algorithm and random forest,” in 2019 11th International Conference on Electrical and Electronics Engineering (ELECO) (Bursa: IEEE), 470–474. doi: 10.23919/ELECO47770.2019.8990557
Peng, C., Wu, X., Yuan, W., Zhang, X., Zhang, Y., Li, Y., et al. (2021). MGRFE: Multilayer Recursive Feature Elimination Based on an Embedded Genetic Algorithm for Cancer Classification. IEEE/ACM Transactions on Computational Biology and Bioinformatics, 18, 621–632. doi: 10.1109/TCBB.2019.2921961
Pirch, S., Müller, F., Iofinova, E., Pazmandi, J., Hütter, C. V., Chiettini, M., et al. (2021). The vrnetzer platform enables interactive network analysis in virtual reality. Nat. Commun. 12:2432. doi: 10.1038/s41467-021-22570-w
Pirgazi, J., Alimoradi, M., Esmaeili Abharian, T., and Olyaee, M. H. (2019). An efficient hybrid filter-wrapper metaheuristic-based gene selection method for high dimensional datasets. Sci. Rep. 9:18580. doi: 10.1038/s41598-019-54987-1
Prabhakar, S. K., and Lee, S.-W. (2020). Transformation based tri-level feature selection approach using wavelets and swarm computing for prostate cancer classification. IEEE Access 8, 127462–127476. doi: 10.1109/ACCESS.2020.3006197
Prajapati, S., Das, H., and Gourisaria, M. K. (2023a). “Feature selection using ant colony optimization for microarray data classification,” in 2023 6th International Conference on Information Systems and Computer Networks (ISCON) (Mathura: IEEE), 1–6. doi: 10.1109/ISCON57294.2023.10112054
Prajapati, S., Das, H., and Gourisaria, M. K. (2023b). Feature selection using differential evolution for microarray data classification. Discov. Int. Things 3:12. doi: 10.1007/s43926-023-00042-5
Prajapati, S., Das, H., and Gourisaria, M. K. (2023c). “Feature selection using genetic algorithm for microarray data classification,” in 2022 OPJU International Technology Conference on Emerging Technologies for Sustainable Development (OTCON) (Raigarh: IEEE), 1–6. doi: 10.1109/OTCON56053.2023.10113937
Pudjihartono, N., Fadason, T., Kempa-Liehr, A. W., and O'Sullivan, J. M. (2022). A review of feature selection methods for machine learning-based disease risk prediction. Front. Bioinformatics 2:927312. doi: 10.3389/fbinf.2022.927312
Qiu, C. (2019). A novel multi-swarm particle swarm optimization for feature selection. Genet. Program. Evolvable Mach. 20, 503–529. doi: 10.1007/s10710-019-09358-0
Roberts, A. G. K., Catchpoole, D. R., and Kennedy, P. J. (2018). “Variance-based feature selection for classification of cancer subtypes using gene expression data,” in 2018 International Joint Conference on Neural Networks (IJCNN) (Rio de Janeiro: IEEE), 1–8. doi: 10.1109/IJCNN.2018.8489279
Rochayani, M. Y. Sa'adah, U., and Astuti, A. B. (2020). Two-stage gene selection and classification for a high-dimensional microarray data. J. Online Inform. 5, 9–18. doi: 10.15575/join.v5i1.569
Rupapara, V., Rustam, F., Aljedaani, W., Shahzad, H. F., Lee, E., Ashraf, I., et al. (2022). Blood cancer prediction using leukemia microarray gene data and hybrid logistic vector trees model. Sci. Rep. 12:1000. doi: 10.1038/s41598-022-04835-6
Saberi-Movahed, F., Biswas, B., Tiwari, P., Lehmann, J., and Vahdati, S. (2024). Deep nonnegative matrix factorization with joint global and local structure preservation. Expert Syst. Appl. 249:123645. doi: 10.1016/j.eswa.2024.123645
Saberi-Movahed, F., Rostami, M., Berahmand, K., Karami, S., Tiwari, P., Oussalah, M., et al. (2022). Dual regularized unsupervised feature selection based on matrix factorization and minimum redundancy with application in gene selection. Knowl.-Based Syst. 256:109884. doi: 10.1016/j.knosys.2022.109884
Sahu, B., and Dash, S. (2023). “Hybrid multifilter ensemble based feature selection model from microarray cancer datasets using gwo with deep learning,” in 2023 3rd International Conference on Intelligent Technologies (CONIT) (Hubli: IEEE), 1–6. doi: 10.1109/CONIT59222.2023.10205668
Sahu, B., and Dash, S. (2024). Optimal feature selection from high-dimensional microarray dataset employing hybrid ig-jaya model. Curr. Mater. Sci. 17, 21–43. doi: 10.2174/2666145416666230124143912
Samareh-Jahani, M., Saberi-Movahed, F., Eftekhari, M., Aghamollaei, G., and Tiwari, P. (2024). Low-redundant unsupervised feature selection based on data structure learning and feature orthogonalization. Expert Syst. Appl. 240:122556. doi: 10.1016/j.eswa.2023.122556
Sánchez-Maroño, N., Fontenla-Romero, O., and Pérez-Sánchez, B. (2019). Classification of microarray data. Microarray Bioinformatics 1986, 185–205. doi: 10.1007/978-1-4939-9442-7_8
Shafi, A., Molla, M. I., Jui, J. J., and Rahman, M. M. (2020). Detection of colon cancer based on microarray dataset using machine learning as a feature selection and classification techniques. SN Appl. Sci. 2, 1–8. doi: 10.1007/s42452-020-3051-2
Sheikhpour, R., Berahmand, K., Mohammadi, M., and Khosravi, H. (2025). Sparse feature selection using hypergraph laplacian-based semi-supervised discriminant analysis. Pattern Recognit. 157:110882. doi: 10.1016/j.patcog.2024.110882
Shukla, A. K., Singh, P., and Vardhan, M. (2019a). A new hybrid feature subset selection framework based on binary genetic algorithm and information theory. Int. J. Comput. Intell. Appl. 18:1950020. doi: 10.1142/S1469026819500202
Shukla, A. K., Singh, P., and Vardhan, M. (2019b). A new hybrid wrapper TLBO and SA with SVM approach for gene expression data. Inf. Sci. 503, 238–254. doi: 10.1016/j.ins.2019.06.063
Shukla, A. K., Singh, P., and Vardhan, M. (2020). Gene selection for cancer types classification using novel hybrid metaheuristics approach. Swarm Evol. Comput. 54:100661. doi: 10.1016/j.swevo.2020.100661
Shukla, A. K., Singh, P., Vardhan, M., Thampi, S. M., and El-Alfy, E.-S. M. (2019c). A hybrid framework for optimal feature subset selection. J. Intell. Fuzzy Syst. Appl. Eng. Technol. 36, 2247–2259. doi: 10.3233/JIFS-169936
Shukla, A. K., and Tripathi, D. (2020). Detecting biomarkers from microarray data using distributed correlation based gene selection. Genes Genomics 42, 449–465. doi: 10.1007/s13258-020-00916-w
Singh, A. V., Chandrasekar, V., Paudel, N., Laux, P., Luch, A., Gemmati, D., et al. (2023). Integrative toxicogenomics: advancing precision medicine and toxicology through artificial intelligence and omics technology. Biomed. Pharmacother. 163:114784. doi: 10.1016/j.biopha.2023.114784
Singh, N., and Singh, P. (2021). A hybrid ensemble-filter wrapper feature selection approach for medical data classification. Chemom. Intell. Lab. Syst. 217:104396. doi: 10.1016/j.chemolab.2021.104396
Sucharita, S., Sahu, B., Swarnkar, T., and Meher, S. K. (2024). Classification of cancer microarray data using a two-step feature selection framework with moth-flame optimization and extreme learning machine. Multimed. Tools Appl. 83, 21319–21346. doi: 10.1007/s11042-023-16353-2
Sun, M., Liu, K., Wu, Q., Hong, Q., Wang, B., Zhang, H., et al. (2019). A novel ECOC algorithm for multiclass microarray data classification based on data complexity analysis. Pattern Recognit. 90, 346–362. doi: 10.1016/j.patcog.2019.01.047
Tang, C., Bian, M., Liu, X., Li, M., Zhou, H., Wang, P., et al. (2019). Unsupervised feature selection via latent representation learning and manifold regularization. Neural Netw. 117, 163–178. doi: 10.1016/j.neunet.2019.04.015
Tatwani, S., and Kumar, E. (2019). A master slave parallel genetic algorithm for feature selection in high dimensional datasets. Int. J. Recent Technol. Eng. 8, 379–384. doi: 10.35940/ijrte.C4184.098319
Tavasoli, N., Rezaee, K., Momenzadeh, M., and Sehhati, M. (2021). An ensemble soft weighted gene selection-based approach and cancer classification using modified metaheuristic learning. J. Comput. Des. Eng. 8, 1172–1189. doi: 10.1093/jcde/qwab039
Too, J., and Mirjalili, S. (2021). A hyper learning binary dragonfly algorithm for feature selection: a COVID-19 case study. Knowl.-Based Syst. 212:106553. doi: 10.1016/j.knosys.2020.106553
Trevizan, B., and Recamonde-Mendoza, M. (2021). “Ensemble feature selection compares to meta-analysis for breast cancer biomarker identification from microarray data,” in Computational Science and Its Applications-ICCSA 2021, Lecture Notes in Computer Science, eds. O. Gervasi, B. Murgante, S. Misra, C. Garau, I. Blečić, D. Taniar, et al. (Cham: Springer International Publishing), 162–178. doi: 10.1007/978-3-030-86653-2_12
Vatankhah, M., and Momenzadeh, M. (2024). Self-regularized lasso for selection of most informative features in microarray cancer classification. Multimed. Tools Appl. 83, 5955–5970. doi: 10.1007/s11042-023-15207-1
Vatansever, S., Schlessinger, A., Wacker, D., Kaniskan, H.Ü., Jin, J., Zhou, M.-M., et al. (2021). Artificial intelligence and machine learning-aided drug discovery in central nervous system diseases: state-of-the-arts and future directions. Med. Res. Rev. 41, 1427–1473. doi: 10.1002/med.21764
Wang, A., Liu, H., Liu, J., Ding, H., Yang, J., and Chen, G. (2020). “Stable and accurate feature selection from microarray data with ensembled fast correlation based filter,”' in 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) (IEEE), 2996–2998.
Wang, X., Dai, C., Wu, Y., Liu, Y., and Wei, D. (2023a). Molecular-electromechanical system for unamplified detection of trace analytes in biofluids. Nat. Protoc. 18, 2313–2348. doi: 10.1038/s41596-023-00830-x
Wang, Y., Gao, X., Ru, X., Sun, P., and Wang, J. (2023b). Using feature selection and Bayesian network identify cancer subtypes based on proteomic data. J. Proteomics 280:104895. doi: 10.1016/j.jprot.2023.104895
Wu, B., Li, M., Lou, M., and Deng, S. (2022). “Gene selection based on XGBoost and GWO for microarray data classification,” in 2022 2nd International Conference on Big Data, Artificial Intelligence and Risk Management (ICBAR) (Xi'an: IEEE), 123–129. doi: 10.1109/ICBAR58199.2022.00031
Xie, W., Li, W., Zhang, S., Wang, L., Yang, J., Zhao, D., et al. (2022a). A novel biomarker selection method combining graph neural network and gene relationships applied to microarray data. BMC Bioinformatics 23, 1–18. doi: 10.1186/s12859-022-04848-y
Xie, W., Zhang, S., Wang, L., Yu, K., and Li, W. (2022b). Feature Selection of Microarray Data Using Multidimensional Graph Neural Network and Supernode Hierarchical Clusterin. Dordrecht: Artificial Intelligence Review, Springer.
Xie, W., Zhang, S., Wang, L., Yu, K., and Li, W. (2024). Feature selection of microarray data using multidimensional graph neural network and supernode hierarchical clustering. Artif. Intell. Rev. 57:63. doi: 10.1007/s10462-023-10700-3
Yang, C., Liu, Z., Dai, P., Zhang, Y., Huang, P., Lin, Y., et al. (2022). Predicting tumor drug sensitivity with multi-omics data. Sheng Wu Gong Cheng Xue Bao 38, 2201–2212. doi: 10.13345/j.cjb.210676
Yang, X., Kui, L., Tang, M., Li, D., Wei, K., Chen, W., et al. (2020). High-throughput transcriptome profiling in drug and biomarker discovery. Front. Genet. 11:19. doi: 10.3389/fgene.2020.00019
Ye, S., Zhou, K., Zain, A. M., Wang, F., and Yusoff, Y. (2023). A modified harmony search algorithm and its applications in weighted fuzzy production rule extraction. Front. Inf. Technol. Electron. Eng. 24, 1574–1590. doi: 10.1631/FITEE.2200334
Yun, K. K., Yoon, S. W., and Won, D. (2023). Interpretable stock price forecasting model using genetic algorithm-machine learning regressions and best feature subset selection. Expert Syst. Appl. 213:118803. doi: 10.1016/j.eswa.2022.118803
Zare, M., Azizizadeh, N., and Kazemipour, A. (2023). Supervised feature selection on gene expression microarray datasets using manifold learning. Chemometr. Intell. Lab. Syst. 237:104828. doi: 10.1016/j.chemolab.2023.104828
Zhang, Y., Gong, D.-w., Gao, X.-z., Tian, T., and Sun, X.-y. (2020). Binary differential evolution with self-learning for multi-objective feature selection. Inf. Sci. 507, 67–85. doi: 10.1016/j.ins.2019.08.040
Zhang, Y., Zhu, R., Chen, Z., Gao, J., and Xia, D. (2021). Evaluating and selecting features via information theoretic lower bounds of feature inner correlations for high-dimensional data. Eur. J. Oper. Res. 290, 235–247. doi: 10.1016/j.ejor.2020.09.028
Zhou, Y., Kang, J., Kwong, S., Wang, X., and Zhang, Q. (2021). An evolutionary multi-objective optimization framework of discretization-based feature selection for classification. Swarm Evol. Comput. 60:100770. doi: 10.1016/j.swevo.2020.100770
Keywords: cancer classification, feature selection, microarray data, machine learning, gene expression analysis
Citation: Wang F, Zain AM, Ren Y, Bahari M, Samah AA, Ali Shah ZB, Yusup NB, Jalil RA, Mohamad A and Azmi NFM (2025) Navigating the microarray landscape: a comprehensive review of feature selection techniques and their applications. Front. Big Data 8:1624507. doi: 10.3389/fdata.2025.1624507
Received: 08 May 2025; Accepted: 16 June 2025;
Published: 10 July 2025.
Edited by:
Kunpeng Liu, Portland State University, United StatesReviewed by:
Xinhao Zhang, Portland State University, United StatesYeyang Zhou, Uber, United States
Copyright © 2025 Wang, Zain, Ren, Bahari, Samah, Ali Shah, Yusup, Jalil, Mohamad and Azmi. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Azlan Mohd Zain, YXpsYW5tekB1dG0ubXk=