Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Med.

Sec. Ophthalmology

Volume 12 - 2025 | doi: 10.3389/fmed.2025.1600736

This article is part of the Research TopicEfficient Artificial Intelligence in Ophthalmic Imaging – Volume IIView all 10 articles

Med-DGTN: Dynamic Graph Transformer with Adaptive Wavelet Fusion for Multi-label Medical Image Classification

Provisionally accepted
Guanyu  ZhangGuanyu Zhang1Yan  LiYan Li2*Tingting  WangTingting Wang1Guokun  ShiGuokun Shi1Jing  LiJing Li1Zongyun  GUZongyun GU1,3*
  • 1Anhui University of Chinese Medicine, Hefei, China
  • 2Hefei First People’s Hospital, Hefei, China
  • 3Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei, Anhui Province, China

The final, formatted version of the article will be published soon.

Introduction: Multi-label classification of medical imaging data aims to enable simultaneous identification and diagnosis of multiple diseases, delivering comprehensive clinical decision support for complex conditions. Current methodologies demonstrate limitations in capturing disease co-occurrence patterns and preserving subtle pathological signatures. To address these challenges, we propose Med-DGTN, a dynamically integrated framework designed to advance multi-label classification performance in clinical imaging analytics.The proposed Med-DGTN (Dynamic Graph Transformer Network with Adaptive Wavelet Fusion) introduces three key innovations: 1)A cross-modal alignment mechanism integrating convolutional visual patterns with graph-based semantic dependencies through conditionally reweighted adjacency matrices; 2)Wavelet-transform-enhanced dense blocks (WTDense) employing multi-frequency decomposition to amplify low-frequency pathological biomarkers; 3)An adaptive fusion architecture optimizing multi-scale feature hierarchies across spatial and spectral domains.Results: Validated on two public medical imaging benchmarks, Med-DGTN demonstrates superior performance across modalities:1)Achieving a mean average precision (mAP) of 70.65% on the retinal imaging dataset (MuReD2022), surpassing previous state-of-the-art methods by 2.68 percentage points. 2) On the chest X-ray dataset (ChestXray14), Med-DGTN achieves an average Area Under the Curve (AUC) of 0.841. It outperforms prior state-of-the-art methods in 5 of 14 disease categories.Discussion: This investigation establishes that joint modeling of dynamic disease correlations and wavelet-optimized feature representation significantly enhances multi-label diagnostic capabilities. Med-DGTN's architecture demonstrates clinical translatability by revealing disease interaction patterns through interpretable graph structures, potentially informing precision diagnostics in multi-morbidity scenarios.

Keywords: Dynamic Graph Transformer, Wavelet Transform, multi-label classification, Medical Image Analysis, deep learning

Received: 26 Mar 2025; Accepted: 11 Jul 2025.

Copyright: © 2025 Zhang, Li, Wang, Shi, Li and GU. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence:
Yan Li, Hefei First People’s Hospital, Hefei, China
Zongyun GU, Anhui University of Chinese Medicine, Hefei, China

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.