ORIGINAL RESEARCH article
Front. Neurosci.
Sec. Brain Imaging Methods
Volume 19 - 2025 | doi: 10.3389/fnins.2025.1658776
This article is part of the Research TopicExploring Neuropsychiatric Disorders Through Multimodal MRI: Network Analysis, Biomarker Discovery, and Clinical InsightsView all articles
A Multi-view Multimodal Deep Learning Framework for Alzheimer's Disease Diagnosis
Provisionally accepted- 1Communication and Network Key Laboratory, Dalian University, Dalian, China
- 2Dalian University Affiliated Xinhua Hospital, Dalian University, Dalian, China
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Early diagnosis of Alzheimer's disease (AD) is challenging, primarily due to the high similarity in characteristics between AD, mild cognitive impairment (MCI), and cognitively normal (CN) individuals, as well as the influence of confounding factors such as population heterogeneity, label noise, and variations in imaging acquisition. Although multimodal neuroimaging techniques such as MRI and PET can provide rich complementary information, existing methods remain inadequate in multimodal feature fusion and multi-scale feature information aggregation. To address this issue, we propose a novel multimodal diagnostic framework called Alzheimer's Disease Multi-View Multimodal Diagnostic Network (ADMV-Net), which aims to improve recognition accuracy across all stages of AD. We first design a dual-pathway structure using a Hybrid Convolution ResNet module that effectively fuses global semantic and local boundary information, providing robust support for extracting feature information from three-dimensional medical images. Additionally, we introduce a Multi-view Fusion Learning mechanism that encompasses a Global Perception Module, Multi-level Local Cross-modal Aggregation Network, and Bidirectional Cross-Attention Module. This mechanism captures and efficiently integrates multimodal features from multiple perspectives, fully exploiting complementary information between modalities. Finally, we incorporate a Regional Interest Perception Module to precisely locate important brain regions associated with AD pathology, enhancing the model's perception of disease characteristics. Comparative experimental results on public datasets demonstrate that the proposed ADMV-Net achieves 94.83% accuracy and 95.97% AUC in AD versus CN classification tasks, significantly outperforming current mainstream methods. The model also exhibits excellent discriminative ability and strong generalization performance in multi-classification tasks.The code is available at https://github.com/zhaoxinyu-1/ADMV-Net.
Keywords: Alzheimer's disease, multimodal fusion, Multi-view learning, cross-modal attention, Neuroimaging
Received: 03 Jul 2025; Accepted: 10 Sep 2025.
Copyright: © 2025 Feng, Zhao, Liu, Ding and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence:
Jianxin Feng, Communication and Network Key Laboratory, Dalian University, Dalian, China
Feng Wang, Dalian University Affiliated Xinhua Hospital, Dalian University, Dalian, China
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.