ORIGINAL RESEARCH article

Front. Pharmacol., 28 April 2025

Sec. Ethnopharmacology

Volume 16 - 2025 | https://doi.org/10.3389/fphar.2025.1543966

AMFGNN: an adaptive multi-view fusion graph neural network model for drug prediction

  • 1. Faculty of Pediatrics, The Chinese PLA General Hospital, Beijing, China

  • 2. Department of Child Growth and Development Clinic, The Seventh Medical Center of PLA General Hospital, Beijing, China

  • 3. National Engineering Laboratory for Birth Defects Prevention and Control of Key Technology, Beijing, China

  • 4. Beijing Key Laboratory of Pediatric Organ Failure, Beijing, China

  • 5. Department of Pediatric Surgery, The Seventh Medical Center of PLA General Hospital, Beijing, China

Article metrics

View details

1,6k

Views

492

Downloads

Abstract

Introduction:

Drug development is a complex and lengthy process, and drug-disease association prediction aims to significantly improve research efficiency and success rates by precisely identifying potential associations. However, existing methods for drug-disease association prediction still face limitations in feature representation, feature integration, and generalization capabilities.

Methods:

To address these challenges, we propose a novel model named AMFGNN (Adaptive Multi-View Fusion Graph Neural Network). This model leverages an adaptive graph neural network and a graph attention network to extract drug features and disease features, respectively. These features are then used as the initial representations of nodes in the drug-disease association network to enable efficient information fusion. Additionally, the model incorporates a contrastive learning mechanism, which enhances the similarity and differentiation between drugs and diseases through cross-view contrastive learning, thereby improving the accuracy of association prediction. Furthermore, a Kolmogorov-Arnold network is employed to perform weighted fusion of various final features, optimizing prediction performance.

Results:

AMFGNN demonstrates a significant advantage in predictive performance, achieving an average AUC value of 0.9453, which reflects the model‘s high accuracy in prediction.

Discussion:

Cross-validation results across multiple datasets indicate that AMFGNN outperforms seven advanced drug-disease association prediction methods. Additionally, case studies on Hepatoblastoma, asthma and Alzheimer‘s disease further confirm the model‘s effectiveness and potential value in real-world applications.

1 Introduction

Since the outbreak of the COVID-19 pandemic, the global healthcare system has faced unprecedented challenges (Vaz et al., 2023; Meng et al., 2024), making the need for safe and effective treatment strategies more urgent than ever. Drug repositioning has attracted widespread attention because of its ability to rapidly identify new therapeutic options, effectively reducing both the cost and timeline of new drug development (Pushpakom et al., 2019; Tang et al., 2023). Recent breakthroughs in high-throughput screening technologies and continuous improvements in computational methods have significantly enhanced the efficiency and accuracy of computer-aided drug repositioning in identifying potential drug-disease associations (Singh et al., 2024; Zeng et al., 2024).

In the field of drug-disease association (DDA) prediction, research methods have gradually evolved from traditional machine learning models to deep learning techniques (Lavecchia, 2015). Traditional machine learning-based DDA prediction identifies potential drug-disease associations through data modeling and feature extraction, and the application of diverse algorithms has further expanded the depth of research in this area. For example, Gao et al. (2022a) combined similarity fusion technology with Laplacian regularization algorithms to accurately predict new indications for drugs and diseases; Yang et al. (2021) developed the MKDGRLS model, employing a multi-kernel approach and Laplacian regularization to handle complex interactions and optimizing model parameters through alternating least squares; additionally, Zhang et al. (2020) proposed the Bayesian inductive matrix completion (DRIMC) method, which integrates features from multiple data sources for analysis in latent space, effectively predicting new drug applications; Niu et al. (2024) introduced the SRR-DDI model, which utilizes a self-attention mechanism to represent drug substructures finely and incorporates drug similarity features, significantly enhancing the stability and performance of drug interaction predictions.

Compared to traditional methods, drug-disease association (DDA) prediction methods based on graph neural networks (GNNs) have made significant breakthroughs in recent years (Zhang et al., 2021; Meng et al., 2024; Zeng et al., 2024). Their unique advantage lies in the ability of GNNs to effectively process multimodal data and model complex network structures, a characteristic that has also gained attention in other fields (Liu et al., 2023). For example, Gao et al. (2022b) proposed the CTST model, which constructs a heterogeneous network of drugs and diseases, using graph convolutional autoencoders to encode shared and unique features of nodes. The model also integrates features through an attention mechanism, significantly improving prediction accuracy. Zhao et al. (2022a) developed the HINGRL model, which leverages a heterogeneous information network encompassing drug-disease and protein-protein interactions, enhances feature recognition through graph representation learning, and combines a random forest algorithm for precise drug indication prediction. Yang et al. (2024) designed the GCNGAT model, which integrates graph convolutional networks with graph attention networks, particularly suitable for drug repositioning. This model analyzes drug-disease associations by constructing heterogeneous graphs and extracts key interaction features in multi-disease contexts. Additionally, Liu et al. (2024) proposed the AMDGT framework, which uses a dual-graph transformer technique to integrate similarity data and complex biochemical information, deeply merging drug and disease features to efficiently predict potential drug associations. Wang et al. (2025) proposed an automatic collaborative learning framework that integrates neighbor interaction metrics with the message-passing mechanism of Graph Neural Networks to enhance prediction accuracy.

Although machine learning and deep learning methods have made significant progress in drug-disease association (DDA) prediction, existing approaches still face challenges in handling complex, multidimensional data and efficiently integrating information from multiple perspectives. To address this issue, we propose an Adaptive Multi-view Fusion Graph Neural Network (AMFGNN) model. The model first constructs drug-drug similarity networks and disease-disease similarity networks, using graph attention networks to extract drug and disease features, which are then used as initial features for the downstream drug-disease association network nodes. To further improve the accuracy of drug-disease association prediction, the model incorporates a contrastive learning mechanism that enhances the similarity and dissimilarity between drugs and diseases through cross-view contrastive learning. The final features are predicted using a Kolmogorov-Arnold Networks (KAN), further improving the prediction accuracy of drug-disease associations. By integrating multi-source heterogeneous information and achieving adaptive weighted fusion, the model flexibly handles complex drug and disease data, dynamically adjusting the weights of different information sources, thereby improving the accuracy, stability, and generalization of predictions. This approach establishes deeper connections between multidimensional data and multi-view information, providing stronger support for drug discovery and personalized medicine.

2 Materials and methods

In this section, we first describe the benchmark datasets used in the proposed model. Next, we introduce the AMFGNN model framework, which consists of three main components. As shown in Figure 1, the framework includes: (i) construct similarity network, (ii) feature extraction and fusion module, and (iii) prediction module.

FIGURE 1

2.1 Dataset

To comprehensively evaluate the performance of the proposed model, we adopted three benchmark datasets: Gottlieb et al. (2011), Luo et al. (2016), and Liang et al. (2017), which are widely used in drug repositioning research. Fdataset is a validated dataset containing 1,632 known drug-disease associations, involving 468 drugs and 298 diseases, providing a reliable reference standard for drug repositioning studies. Cdataset includes 663 drugs, 409 diseases, and 2,352 drug-disease interaction pairs, with the data first appearing in Luo et al.’s study. LRSSL consists of 3,051 validated drug-disease associations, involving 763 drugs and 681 diseases, and has been widely applied in drug repositioning research.

2.2 Graph attention networks

In this study, Graph Attention Network (GAT) (Veličković et al., 2017) is introduced for drug-disease prediction tasks. Graph Neural Networks (GNNs) (Wu et al., 2020) learn interactions between nodes and edges in a graph to perform tasks such as node classification, graph classification, and link prediction. GAT extends this by assigning different weights to each node and its neighbors using an attention mechanism. These weighted features are aggregated to learn the node’s embedding representation. We use a three-layer twin GAT network to extract features for both drugs and diseases. Taking the drug similarity network as an example, the drug similarity graph consists of a feature matrix and an adjacency matrix . To construct the adjacency matrix , we use a k-nearest neighbors (KNN) algorithm to identify the most similar drugs to a drug , and establish connections. For drugs that do not belong to the nearest neighbors, no connection is made.

In GAT, the attention mechanism computes the importance of a drug for its neighboring drug .Specifically, the attention coefficient between drug and its neighbor is calculated as follows Equation 1:

Here, is the learnable weight matrix, and and represent the feature vectors of drugs and , respectively. The LeakyReLU function serves as the activation function, and the attention coefficient represents the importance of neighboring drug node to the central drug node . A higher value of indicates a greater contribution from the features of node to updating the representation of node . Through this mechanism, GAT effectively identifies and emphasizes connections in the drug similarity network that carry higher predictive significance, thus enhancing the quality of drug embedding representations and ultimately improving the accuracy of drug-disease association predictions.

After computing the attention coefficients, the feature vector of drug is updated by aggregating the features of all its neighboring drugs weighted by the attention coefficients. The updated feature vector is calculated as follows Equation 2:where is the activation function, and LeakyReLU is used. To capture the complex relationships between nodes, GAT introduces a multi-head attention mechanism that learns multiple sets of attention weights. The multi-head attention calculation is expressed as Equation 3:Here, is the number of attention heads, and represents the weight of the -th attention head between drug and drug is the weight matrix for the -th attention head.

The multi-head attention mechanism allows GAT to capture diverse relationships between drugs from multiple perspectives, alleviating information bottlenecks and enhancing the model’s generalization ability. To integrate information from multiple heads, GAT averages the outputs of the different heads in the final layer, providing a more comprehensive and stable embedding representation Equation 4:Through this multi-head attention mechanism, GAT can more effectively capture the complex dependencies between drugs and generate accurate drug embeddings. In this study, both the drug similarity and disease similarity networks employ a three-layer GAT structure to improve the accuracy of drug-disease prediction and the model’s expressive power.

2.3 Adaptive fusion

The core idea of the Graph Attention Network (GAT) is to update node features based on the importance weights of neighboring nodes. However, as the depth of the GAT network increases, an issue of over-smoothing may arise, where the features of all nodes in the graph become too similar, severely affecting the model’s prediction accuracy. To address this issue, we introduce residual connections to ensure that GAT can adaptively retain the original features while updating node features. The specific operation is as follows Equations 58:

Here, R denotes the residual connection operation, denotes the initial feature representation of the drug similarity network (i.e., the pre-processed embedding derived from the original drug feature matrix ), while represents the initial feature representation of the disease similarity network (i.e., the pre-processed embedding derived from the original disease feature matrix ). By introducing residual connections, the model adaptively preserves initial feature information, effectively addressing the issue of over-smoothing that typically occurs as the depth of the network increases. corresponds to the feature of the drug-disease association view. The weight matrix is represented by , where is the bias term, and denotes the residual result. The activation function used is the Elu function.

We fuse the residual connections of different views with the features from the previous layer, as shown in the following Equations 911:

Here, denotes the feature vector of the -th view of the drug, is the feature representation of the drug similarity modality in the final layer of GAT, and is an adaptive variable learned during training, used to control the fusion weight between residual and original features.

To effectively integrate multi-view features of drugs and diseases, we adopt an adaptive feature fusion strategy Equations 12, 13:

Here, is the drug embedding after multi-view fusion, is the disease embedding after multi-modal fusion, and is the adaptive variable learned during model training. Finally, we input the fused drug embedding and disease embedding into KAN to compute the predicted potential connection Equation 14.

During training, we use the cross-entropy loss function to minimize the error between the model’s predictions and the true labels Equation 15:

Here, and represent the positive and negative sample sets in the dataset, respectively, and is the true association score between drug and disease . To further improve the model’s prediction accuracy and constrain the parameter updates across modalities, we introduce multi-view contrastive learning as a regularization term. This method reduces the distance between features of the same sample from different views, while increasing the distance between features of different samples from different views, thereby enhancing the model’s feature representation ability. The objective function is defined as follows Equation 16:

For each drug sample , the contrastive learning loss function is defined as Equation 17:

Here, is the total number of drug samples, represents the embedding of drug sample in the -th modality, and d (u, v) is the distance function calculated using cosine similarity Equation 18:

Similarly, for each disease sample , the loss function is defined as Equation 19:

Finally, the overall loss function of the model is defined as Equation 20:

2.4 Kolmogorov-Arnold Networks

In the drug repositioning task, to improve the parameter efficiency of the model, we modified the traditional multilayer perceptron (MLP) structure by replacing the final MLP module with Kolmogorov-Arnold Networks (KAN). KAN introduces learnable activation functions, replacing traditional linear weight matrices, which significantly enhance the network’s expressive power while maintaining or even improving model performance. The traditional MLP captures complex mappings through linear transformations and fixed nonlinear activation functions, which often limits model flexibility and increases parameter redundancy. Mathematically, an MLP is expressed as Equation 21:Where is the input vector, is the weight matrix, and is the activation function. Although effective in learning complex functions, the fixed linear transformations in MLPs may restrict the adaptability of the network. KAN, in contrast, utilizes learnable nonlinear activation functions instead of fixed linear weight matrices, providing greater flexibility in capturing complex relationships between input features. Specifically, each connection in KAN is modeled by a combination of a parametric basis function and B-spline functions Equation 22:

Each layer’s consists of a set of learnable activation functions, represented as Equation 23:

We assume that a KAN can be expressed as represents the number of neurons in the -th layer. We use to denote the -th neuron in the -th layer and to represent the activation value of the neuron . Between the -th and -th layers, there are activation functions: the activation function connecting and is represented as Equation 24:

The pre-activation value of is , and the post-activation value of is represented as Equation 25:

The activation value of neuron is the sum of all incoming post-activation values Equation 26:

The activation function is composed of a weighted sum of a basis function and a B-spline function Equations 2729:

Here, and represent the weights of the basis function and B-spline function, respectively. are the trainable parameters in the B-spline function, and is the B-spline basis function defined on a grid. By using adaptive activation functions, KAN significantly enhances the representational power of neural networks, allowing the model to learn smoother and more complex transformations without increasing model complexity excessively. This flexible representation is particularly beneficial for integrating diverse features from multi-view data, such as drug similarity, disease similarity, and their interactions, leading to improved predictive performance for drug-disease associations.

3 Results and discussion

3.1 Parameter settings

We perform 10-fold cross-validation to evaluate the performance of AMFGNN. In the 10-fold cross-validation, all known and unknown drug-disease associations are randomly divided into 10 subsets of approximately equal size. Each subset is used as the test set in turn, while the remaining nine subsets serve as the training set.

We set the feature embedding size to 128 to achieve the best prediction performance in drug-drug similarity, disease-disease similarity, and drug-disease association graphs, and we set the dropout rate to 0.2 to optimize the training process of the network layers. For the graph model selection, we choose to use a 3-layer GAT instead of GCN (Graph Convolution Network) because GAT outperforms GCN in terms of AUC. The 3-layer structure helps prevent both information redundancy and over-smoothing issues. For optimization, we employed the Adam optimizer with a learning rate of 0.001 and weight decay of 0.001, training the model for 300 epochs.

3.2 Model evaluation and cross-validation

To rigorously evaluate the performance of our proposed model, we employed a standard 10-fold cross-validation approach. Specifically, drug-disease associations from each benchmark dataset were randomly divided into ten subsets of approximately equal size. In each fold, nine subsets were combined to form the training set, while the remaining subset served as the test set for evaluating the model’s performance. This procedure was repeated ten times, with each subset serving as the test set exactly once. We reported the model’s performance using the average and standard deviation across these ten evaluations. Furthermore, this cross-validation procedure was independently conducted on all three benchmark datasets (Fdataset, Cdataset, and LRSSL) to ensure the reliability and generalizability of our results.

3.3 Baseline methods

To evaluate the performance of AMFGNN, we performed 10-fold cross-validation on three public datasets: Fdataset, Cdataset, and LRSSL. The models compared in this study include LBMFF, SCPMFDD, SCPMF, MKGCN, and MNGACDA.

  • • LBMFF is a model for drug-disease relationship prediction that combines latent bilinear matrix factorization and focal loss. The model captures latent associations between drugs and diseases through matrix factorization techniques and introduces focal loss to address the class imbalance problem, enhancing the model’s ability to handle sparse data and hard-to-predict instances (Kang et al., 2023).

  • • SCPMFDD is a semi-supervised learning model for drug-disease prediction. The model combines collaborative projection matrix factorization and semi-supervised learning strategies to enhance prediction performance by leveraging known drug-disease relationships and unlabeled data (Li et al., 2022).

  • • SCPMF is a semi-supervised learning model for drug-disease prediction. It learns the latent relationships between drugs and diseases through collaborative matrix factorization, while also utilizing semi-supervised learning to enhance the model’s learning capability by incorporating unlabeled data. (Meng et al., 2021).

  • • MKGCN is a model for complex drug-disease prediction tasks. By introducing multiple kernel functions (Multi-Kernel), it integrates different types of graph structure features. MKGCN uses Graph Convolutional Networks (GCN) to process drug-disease graph data and applies kernel functions to weight different features, thereby more accurately capturing the complex relationships between drugs and diseases and improving the model’s ability to model and predict (Cui et al., 2023).

  • • MNGACDA is a graph neural network model for drug-disease prediction. The model combines multi-node graph attention mechanisms and dual attention mechanisms, effectively processing the drug and disease relationship graph through Graph Convolutional Networks (Yang and Chen, 2023).

  • • DDAGDL (Zhao et al., 2022b) is a model for drug–disease prediction that applies geometric deep learning over heterogeneous information networks. It integrates biological information into the network structure and uses an attention mechanism to learn effective representations of drugs and diseases, enabling improved performance on non-Euclidean biomedical data.

  • • RGLDR (Zhao et al., 2025) combines regulation-aware graph representation learning with meta-path-based connectivity patterns to capture diverse regulatory mechanisms in heterogeneous biological networks. It enhances drug and disease embeddings using a multi-view attention mechanism and predicts drug-disease associations with an XGBoost classifier. Experimental results demonstrate its superior performance over state-of-the-art methods on benchmark datasets.

According to the results shown in Table 1, AMFGNN achieves the highest AUC in all three datasets. AMFGNN demonstrates outstanding performance across different datasets. On the F dataset, the AUC value of AMFGNN is 0.9328, significantly higher than other models such as LBMFF (0.7953), SCPMFDD (0.7740), and others. Similarly, on the C dataset and LRSSL dataset, AMFGNN also shows higher prediction accuracy, with AUC values of 0.9443 and 0.9588, respectively.

TABLE 1

DatasetsLBMFFSCPMFDDSCPMFMKGCNMNGACDADDAGDLRGLDRAMFGNN
Fdataset0.7953 0.0350.7740 0.0010.8957 0.0010.8870 0.0010.8179 0.0050.9266 0.0010.9311 0.0050.9328 0.014
Cdataset0.9069 0.0010.7937 0.0010.9117 0.0020.9109 0.0010.8406 0.0050.9256 0.0010.9359 0.0030.9443 0.012
LRSSL0.9139 0.0020.7668 0.0010.8977 0.0010.8596 0.0010.7936 0.0020.8990 0.0010.8032 0.0050.9588 0.006
Avg0.87200.77820.90170.88580.81730.91710.89010.9453

Performance comparison of different methods across datasets using AUC.

The results indicate that the AUC (Area Under the Curve) of the AMFGNN model outperform those of other models. This demonstrates that the AMFGNN model effectively improves the accuracy and stability of drug-disease association prediction by integrating multi-source heterogeneous information and dynamically adjusting the weights of different information sources. In addition, based on the results of Recall (Table 2) and F1-score (Table 3), AMFGNN also performs excellently in these metrics. In terms of Recall, AMFGNN effectively captures positive samples and reduces false negatives, indicating its high recall ability. A high Recall value means the model can identify more positive samples, which is especially important for drug-disease association prediction, as missing positive samples could lead to the omission of crucial information. Regarding F1-score, AMFGNN demonstrates a good balance between Precision and Recall, indicating its advantages in both accuracy and recall ability. A high F1-score means the model reduces false positives while effectively capturing more positive samples, avoiding the performance imbalance that may arise from optimizing a single metric. Overall, AMFGNN shows outstanding performance in improving prediction accuracy, stability, and comprehensiveness, further confirming its effectiveness in drug-disease association prediction, especially on complex and imbalanced datasets.

TABLE 2

DatasetsLBMFFSCPMFDDSCPMFMKGCNMNGACDADDAGDLRGLDRAMFGNN
Fdataset0.73280.02550.41280.00980.08010.45600.88140.9152
Cdataset0.70060.02170.47670.06440.16950.48300.87240.9443
LRSSL0.71650.01510.38120.02330.07370.43280.69750.9371
Avg0.71660.02080.42360.03250.10780.45730.81710.9322

Recall comparison of various models across multiple datasets.

TABLE 3 F1

DatasetsLBMFFSCPMFDDSCPMFMKGCNMNGACDADDAGDLRGLDRAMFGNN
Fdataset0.05770.11740.40240.01950.14750.60070.86150.8699
Cdataset0.23430.08570.45560.12100.28750.60130.88060.8890
LRSSL0.23040.02490.35480.04520.13520.55380.71940.9040
Avg0.17410.07600.40430.06190.19010.58530.82050.8876

-Score performance of different approaches across datasets.

As shown in Figure 2, the AUC (Area Under the Curve) values of the AMFGNN model outperform those of other models. This demonstrates that the AMFGNN model effectively improves the accuracy and stability of drug-disease association prediction by integrating multi-source heterogeneous information and dynamically adjusting the weights of different information sources.

FIGURE 2

3.4 Ablation studies

To thoroughly investigate the effectiveness of different components in our proposed AMFGNN model, we conduct comprehensive ablation studies. Specifically, we evaluate four variants of our model:

  • • Full Model: The complete AMFGNN architecture with all components, including the Kolmogorov-Arnold Networks and contrastive learning loss.

  • • AMFGNN w/o KAN: A variant without the Kolmogorov-Arnold Network, while maintaining the contrastive learning mechanism.

  • • AMFGNN w/o CL: A variant that removes the contrastive learning loss while retaining the Kolmogorov-Arnold Networks, utilizing only the main task loss function for optimization.

  • • AMFGNN w/o KAN & CL: The baseline variant that removes both the Kolmogorov-Arnold Network and contrastive learning loss, maintaining only the basic MLP and main task loss.

As shown in Figure 3,the experimental results demonstrate several key findings: The full AMFGNN model achieves the best performance across all metrics, validating the effectiveness of our proposed architecture. Removing the Kolmogorov-Arnold Networks (w/o KAN) leads to a performance decrease of AUC and AUPR, highlighting the importance of KAN layer in our model. The absence of contrastive learning (w/o CL) also results in a drop, indicating that the contrastive learning mechanism plays a crucial role in learning more discriminative feature representations. The baseline variant (w/o KAN & CL) shows the most significant performance degradation, confirming that both components contribute substantially to the model’s effectiveness.

FIGURE 3

These ablation studies clearly demonstrate that each component in our proposed AMFGNN makes meaningful contributions to the overall performance, and their combination leads to the optimal results.

3.5 Case study

To evaluate the practical application value of the model, we conducted a case study. Specifically, the model was trained on the Fdataset to predict potential drugs associated with asthma and hepatoblastoma. The predicted drugs were ranked in descending order based on their probability scores, and the top ten candidates were selected for further analysis. To assess the reliability of the model’s predictions, comprehensive validation was performed using multiple authoritative data sources, including DrugCentral, Avram et al. (2021) CTD, and ClinicalTrials databases.

Hepatoblastoma is a malignant liver tumor primarily occurring in children, typically diagnosed during infancy or early childhood (Sharma et al., 2017). Table 4 highlights five potential therapeutic agents for hepatoblastoma predicted by the AMFGNN model, four of which have been validated by authoritative databases. Furosemide, widely used in both adults and children to manage hypertension and edema caused by liver dysfunction, was identified by the AMFGNN model as a promising candidate for hepatoblastoma treatment. This prediction has been corroborated by the DrugCentral database. Furthermore, the model predicted citalopram as another potential therapeutic agent for hepatoblastoma, with this conclusion supported by both the DrugCentral and CTD databases. These findings provide new perspectives and directions for drug development targeting hepatoblastoma.

TABLE 4 Top

RankDrugBank IDsCandidate drugsEvidences
1DB00397PhenylpropanolamineDrugCentral
2DB00313Valproic acidDrugCentral, CTD
3DB00215CitalopramDrugCentral, CTD
4DB00448LansoprazoleDrugCentral, CTD
5DB00167IsoleucineUnconfirmed

5 candidate drugs for Hepatoblastoma predicted by AMFGNN.

Asthma is a chronic inflammatory disease influenced by both genetic and environmental factors (Toskala and Kennedy, 2015), making it a complex hereditary condition. Table 5 lists the top five potential asthma treatments predicted based on the F dataset, of which four are verified through reliable databases or clinical trials, further supporting the accuracy and practicality of the model’s predictions. AMFGNN predicts flunisolide as a potential drug for treating asthma, a conclusion supported by both DrugBank and ClinicalTrials.gov. Additionally, studies show that uncontrolled asthma is often associated with gastroesophageal reflux disease (GERD) (Harding, 2003). As a proton pump inhibitor, esomeprazole is widely used to treat GERD, and AMFGNN also predicts that esomeprazole might have therapeutic effects on asthma, a prediction that is verified by ClinicalTrials.gov.

TABLE 5 Top

RankDrugBank IDsCandidate drugsEvidences
1DB00180FlunisolideDB, ClinicalTrials.gov
2DB00736EsomeprazoleClinicalTrials.gov
3DB00182AmphetamineDrugCentral, ClinicalTrials.gov
4DB00181BaclofenUnconfirmed
5DB00695FurosemideClinicalTrials.gov

5 candidate drugs for asthma predicted by AMFGNN.

Alzheimer’s Disease (AD) is a neurodegenerative disorder characterized primarily by progressive cognitive decline and memory impairmentAbubakar et al. (2022). Table 6 lists five potential drug candidates for the treatment of Alzheimer’s Disease as predicted by the AMFGNN model, four of which have already been validated by authoritative pharmaceutical databases. Memantine, an approved N-methyl-D-aspartate (NMDA) receptor antagonist, was identified by AMFGNN as an effective therapeutic agent for Alzheimer’s Disease. Additionally, Methylphenidate and Levothyroxine were also predicted as promising therapeutic candidates, with supporting evidence from clinical trials documented in the ClinicalTrials.gov database, suggesting their potential clinical application in the future.

TABLE 6

RankDrugBank IDsCandidate drugsEvidences
1DB01043MemantineClinicalTrials.gov, CTD, DrugCentral
2DB00642MethylphenidateClinicalTrials.gov
3DB00636ClofibrateUnconfirmed
4DB00413PramipexoleClinicalTrials.gov, DrugCentral
5DB00337LevothyroxineClinicalTrials.gov

TOP5 candidate drugs for Alzheimer’s Disease predicted by AMFGNN.

Moreover, we selected five asthma-related target proteins and conducted molecular docking simulations to assess their binding abilities with five candidate drugs using AutoDock Vina (Trott and Olson, 2010). The interactions between the ligands and target proteins were further analyzed using Discovery Studio (DS) visualization software. Regarding the relevance of baclofen to asthma, we chose acidic mammalian chitinase (AMCase, PDB code: 3FY1) as the target protein and found that baclofen has a binding energy of −6.6 kcal/mol with AMCase (Table 7). Figure 4 shows the van der Waals interactions between baclofen and several specific amino acid residues (ALA:183, MET:385, MET:210, TYR:267, GLU:140, GLY:98, PHE:58). In addition, other types of molecular interactions are observed. For example, the oxygen atom forms conventional hydrogen bonds with residues TYR:212, ASP:213, TRP:99, and NA:1, while the nitrogen atom exhibits carbon-hydrogen bond interactions with residue ASP:138. Furthermore, interactions are observed between the small molecule and residues TYR:27 and TRP:360, and covalent bonds form between the oxygen atom and the functional group of residue NA:1.

TABLE 7

Drug1KTJ3FY14P0I5BOW7XXW
Flunisolide−7.4−9.6−9.6−7.5−6.8
Esomeprazole−8.0−8.5−6.6−7.0−5.9
Amphetamine−5.7−5.5−4.8−5.1−4.4
Baclofen−6.0−6.6−5.7−6.4−4.3
Furosemide−6.2−7.4−6.2−6.8−5.9

Molecular binding energies (kcal/mol) between the top 5 candidate drugs for asthma predicted by AMFGNN and 5 target proteins.

FIGURE 4

4 Conclusion

This paper presents the Adaptive Multi-View Fusion Graph Neural Network (AMFGNN), a novel model designed for drug-disease association prediction. AMFGNN integrates multiple data sources, including drug similarity, disease similarity, and drug-disease interactions, using an adaptive multi-view feature fusion strategy. The model combines Graph Attention Networks (GAT) with contrastive learning to improve the accuracy of drug-disease predictions by effectively capturing relationships between nodes in the graph. Additionally, replacing traditional multilayer perceptron (MLP) layers with Kolmogorov-Arnold Networks (KAN) enhances the model’s flexibility, expressive capability, and overall predictive performance.

Through 10-fold cross-validation on three benchmark datasets (F dataset, C dataset, and LRSSL), AMFGNN outperforms existing models, achieving high area under the curve (AUC) scores of 0.9328, 0.9443, and 0.9588, respectively. These results demonstrate that AMFGNN significantly improves drug-disease prediction accuracy, making it a valuable tool in drug repositioning and personalized medicine.

Statements

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding authors.

Author contributions

FH: Writing – original draft, Writing – review and editing. LD: Writing – original draft, Writing – review and editing. GX: Writing – original draft, Writing – review and editing. XC: Investigation, Writing – original draft. HZ: Supervision, Writing – review and editing. MY: Supervision, Writing – review and editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by the Innovation Talent Fund of Senior Department of Pediatrics, The Seventh Medical Center of PLA General Hospital (Grant No. QZX-04-EKCLJH-5).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1

    AbubakarM. B.SanusiK. O.UgusmanA.MohamedW.KamalH.IbrahimN. H.et al (2022). Alzheimer’s disease: an update and insights into pathophysiology. Front. aging Neurosci.14, 742408. 10.3389/fnagi.2022.742408

  • 2

    AvramS.BologaC. G.HolmesJ.BocciG.WilsonT. B.NguyenD.-T.et al (2021). Drugcentral 2021 supports drug discovery and repositioning. Nucleic acids Res.49, D1160D1169. 10.1093/nar/gkaa997

  • 3

    CuiX.QuX.LiD.YangY.LiY.ZhangX. (2023). Mkgcn: multi-modal knowledge graph convolutional network for music recommender systems. Electronics12, 2688. 10.3390/electronics12122688

  • 4

    GaoC.-Q.ZhouY.-K.XinX.-H.MinH.DuP.-F. (2022a). Dda-skf: predicting drug–disease associations using similarity kernel fusion. Front. Pharmacol.12, 784171. 10.3389/fphar.2021.784171

  • 5

    GaoL.CuiH.ZhangT.ShengN.XuanP. (2022b). Prediction of drug–disease associations by integrating common topologies of heterogeneous networks and specific topologies of subnets. Briefings Bioinforma.23, bbab467. 10.1093/bib/bbab467

  • 6

    GottliebA.SteinG. Y.RuppinE.SharanR. (2011). Predict: a method for inferring novel drug indications with application to personalized medicine. Mol. Syst. Biol.7, 496. 10.1038/msb.2011.26

  • 7

    HardingS. M. (2003). Acid reflux and asthma. Curr. Opin. Pulm. Med.9, 4245. 10.1097/00063198-200301000-00007

  • 8

    KangH.HouL.GuY.LuX.LiJ.LiQ. (2023). Drug–disease association prediction with literature based multi-feature fusion. Front. Pharmacol.14, 1205144. 10.3389/fphar.2023.1205144

  • 9

    LavecchiaA. (2015). Machine-learning approaches in drug discovery: methods and applications. Drug Discov. today20, 318331. 10.1016/j.drudis.2014.10.012

  • 10

    LiX.ZhangQ.ZuoZ.YanR.ZhengC.ZhangF. (2022). “Predicting drug-disease associations by self-topological generalized matrix factorization with neighborhood constraints,” in International conference on intelligent computing (Springer), 138149.

  • 11

    LiangX.ZhangP.YanL.FuY.PengF.QuL.et al (2017). Lrssl: predict and interpret drug–disease associations based on data integration using sparse subspace learning. Bioinformatics33, 11871196. 10.1093/bioinformatics/btw770

  • 12

    LiuJ.GuanS.ZouQ.WuH.TiwariP.DingY. (2024). Amdgt: attention aware multi-modal fusion using a dual graph transformer for drug–disease associations prediction. Knowledge-Based Syst.284, 111329. 10.1016/j.knosys.2023.111329

  • 13

    LiuW.TangT.LuX.FuX.YangY.PengL. (2023). Mpclcda: predicting circrna–disease associations by using automatically selected meta-path and contrastive learning. Briefings Bioinforma.24, bbad227. 10.1093/bib/bbad227

  • 14

    LuoH.WangJ.LiM.LuoJ.PengX.WuF.-X.et al (2016). Drug repositioning based on comprehensive similarity measures and bi-random walk algorithm. Bioinformatics32, 26642671. 10.1093/bioinformatics/btw228

  • 15

    MengY.JinM.TangX.XuJ. (2021). Drug repositioning based on similarity constrained probabilistic matrix factorization: covid-19 as a case study. Appl. soft Comput.103, 107135. 10.1016/j.asoc.2021.107135

  • 16

    MengY.WangY.XuJ.LuC.TangX.PengT.et al (2024). Drug repositioning based on weighted local information augmented graph neural network. Briefings Bioinforma.25, bbad431. 10.1093/bib/bbad431

  • 17

    NiuD.XuL.PanS.XiaL.LiZ. (2024). Srr-ddi: a drug–drug interaction prediction model with substructure refined representation learning based on self-attention mechanism. Knowledge-Based Syst.285, 111337. 10.1016/j.knosys.2023.111337

  • 18

    PushpakomS.IorioF.EyersP. A.EscottK. J.HopperS.WellsA.et al (2019). Drug repurposing: progress, challenges and recommendations. Nat. Rev. Drug Discov.18, 4158. 10.1038/nrd.2018.168

  • 19

    SharmaD.SubbaraoG.SaxenaR. (2017). Hepatoblastoma. Seminars diagnostic pathology Elsevier34, 192200. 10.1053/j.semdp.2016.12.015

  • 20

    SinghS.GuptaH.SharmaP.SahiS. (2024). Advances in artificial intelligence (ai)-assisted approaches in drug screening. Artif. Intell. Chem.2, 100039. 10.1016/j.aichem.2023.100039

  • 21

    TangX.ZhouC.LuC.MengY.XuJ.HuX.et al (2023). Enhancing drug repositioning through local interactive learning with bilinear attention networks. IEEE J. Biomed. Health Inf.29, 16441655. 10.1109/JBHI.2023.3335275

  • 22

    ToskalaE.KennedyD. W. (2015). Asthma risk factors. Int. forum allergy and rhinology Wiley Online Libr.5, S11S16. 10.1002/alr.21557

  • 23

    TrottO.OlsonA. J. (2010). Autodock vina: improving the speed and accuracy of docking with a new scoring function, efficient optimization, and multithreading. J. Comput. Chem.31, 455461. 10.1002/jcc.21334

  • 24

    VazE. S.VassiliadesS. V.GiarollaJ.PolliM. C.Parise-FilhoR. (2023). Drug repositioning in the covid-19 pandemic: fundamentals, synthetic routes, and overview of clinical studies. Eur. J. Clin. Pharmacol.79, 723751. 10.1007/s00228-023-03486-4

  • 25

    VeličkovićP.CucurullG.CasanovaA.RomeroA.LioP.BengioY. (2017). Graph attention networks. arXiv Prepr. arXiv:1710.10903. 10.48550/arXiv.1710.10903

  • 26

    WangY.MengY.ZhouC.TangX.ZengP.PanC.et al (2025). Automatic collaborative learning for drug repositioning. Eng. Appl. Artif. Intell.139, 109653. 10.1016/j.engappai.2024.109653

  • 27

    WuZ.PanS.ChenF.LongG.ZhangC.PhilipS. Y. (2020). A comprehensive survey on graph neural networks. IEEE Trans. neural Netw. Learn. Syst.32, 424. 10.1109/tnnls.2020.2978386

  • 28

    YangB.ChenH. (2023). Predicting circrna-drug sensitivity associations by learning multimodal networks using graph auto-encoders and attention mechanism. Briefings Bioinforma.24, bbac596. 10.1093/bib/bbac596

  • 29

    YangH.DingY.TangJ.GuoF. (2021). Drug–disease associations prediction via multiple kernel-based dual graph regularized least squares. Appl. Soft Comput.112, 107811. 10.1016/j.asoc.2021.107811

  • 30

    YangR.FuY.ZhangQ.ZhangL. (2024). Gcngat: drug–disease association prediction based on graph convolution neural network and graph attention network. Artif. Intell. Med.150, 102805. 10.1016/j.artmed.2024.102805

  • 31

    ZengP.ZhangB.LiuA.MengY.TangX.YangJ.et al (2024). Drug repositioning based on tripartite cross-network embedding and graph convolutional network. Expert Syst. Appl.252, 124152. 10.1016/j.eswa.2024.124152

  • 32

    ZhangW.XuH.LiX.GaoQ.WangL. (2020). Drimc: an improved drug repositioning approach using bayesian inductive matrix completion. Bioinformatics36, 28392847. 10.1093/bioinformatics/btaa062

  • 33

    ZhangX.-M.LiangL.LiuL.TangM.-J. (2021). Graph neural networks and their current applications in bioinformatics. Front. Genet.12, 690049. 10.3389/fgene.2021.690049

  • 34

    ZhaoB.-W.HuL.YouZ.-H.WangL.SuX.-R. (2022a). Hingrl: predicting drug–disease associations with graph representation learning on heterogeneous information networks. Briefings Bioinforma.23, bbab515. 10.1093/bib/bbab515

  • 35

    ZhaoB.-W.SuX.-R.HuP.-W.MaY.-P.ZhouX.HuL. (2022b). A geometric deep learning framework for drug repositioning over heterogeneous information networks. Briefings Bioinforma.23, bbac384. 10.1093/bib/bbac384

  • 36

    ZhaoB.-W.SuX.-R.YangY.LiD.-X.LiG.-D.HuP.-W.et al (2025). Regulation-aware graph learning for drug repositioning over heterogeneous biological network. Inf. Sci.686, 121360. 10.1016/j.ins.2024.121360

Summary

Keywords

drug prediction, drug-disease association prediction, graph attention network, contrastive learning, Kolmogorov-Arnold network

Citation

He F, Duan L, Xing G, Chang X, Zhou H and Yu M (2025) AMFGNN: an adaptive multi-view fusion graph neural network model for drug prediction. Front. Pharmacol. 16:1543966. doi: 10.3389/fphar.2025.1543966

Received

12 December 2024

Accepted

15 April 2025

Published

28 April 2025

Volume

16 - 2025

Edited by

Michael Heinrich, University College London, United Kingdom

Reviewed by

Bo-Wei Zhao, Chinese Academy of Sciences (CAS), China

Zeheng Wang, Commonwealth Scientific and Industrial Research Organisation (CSIRO), Australia

Updates

Copyright

*Correspondence: Huixia Zhou, ; Mengnan Yu,

†These authors have contributed equally to this work and share first authorship

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics