Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Med.

Sec. Precision Medicine

Volume 12 - 2025 | doi: 10.3389/fmed.2025.1606238

This article is part of the Research TopicIntegrating AI and Machine Learning in Advancing Patient Care: Bridging Innovations in Mental Health and Cognitive NeuroscienceView all 10 articles

Assessing the Adversarial Robustness of Multimodal Medical AI Systems: Insights into Vulnerabilities and Modality Interactions

Provisionally accepted
Ekaterina  MozhegovaEkaterina Mozhegova1*Asad  Masood KhattakAsad Masood Khattak2Adil  KhanAdil Khan3Roman  GaraevRoman Garaev1Bader  RasheedBader Rasheed1
  • 1Innopolis University, Innopolis, Russia
  • 2Zayed University, Abu Dhabi, United Arab Emirates
  • 3University of Hull, Hull, Yorkshire and the Humber, United Kingdom

The final, formatted version of the article will be published soon.

Deep learning systems have been expanding their potential to be used with different data types a4 a wide range of fields. Medicine is one of the most compelling examples of AI applications due to the numerous tasks and various data types associated with them, such as texts, images, and numerical records. Both task-specific single-modality models and generalpurpose multimodal large models have emerged, presenting new opportunities along with new challenges. Despite the potential of AI systems, regardless of the task, they remain vulnerable to adversarial attacks. These attacks are small perturbations, often imperceptible to the human eye, that are intentionally generated to deceive the model. In high-stakes domains, including healthcare as one of the notable examples, they pose significant risks to model's reliability and consequently to their prospects of being applied in real-world tasks. This is why studies focused on adversarial robustness are of such great importance.In our research, we contribute to the fields of multimodality and adversarial robustness, with a specific focus on applications in medicine. As we investigate the behavior of multimodal models under various attack scenarios, this research sheds light on how these models endure attacks and whether the multimodal nature of these models enhances their resilience. We conducted experiments by applying attacks to two modalities: images and texts. We applied attacks to single-modality models, combined these models with modality fusion techniques and applied the same attacks on the combined model. This approach allowed us to validate our hypothesis that multimodality enhances robustness to adversarial attacks. We believe that our findings have the potential to open new horizons in the field of robust multimodality systems. Additionally, this study can initiate further research focused on the data flow in multimodal systems.

Keywords: machine learning (ML), Adversarial attack, multimodal data fusion, Classification, x-ray

Received: 04 Apr 2025; Accepted: 30 Jun 2025.

Copyright: © 2025 Mozhegova, Khattak, Khan, Garaev and Rasheed. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Ekaterina Mozhegova, Innopolis University, Innopolis, Russia

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.