ORIGINAL RESEARCH article
Front. Med.
Sec. Precision Medicine
Volume 12 - 2025 | doi: 10.3389/fmed.2025.1606238
This article is part of the Research TopicIntegrating AI and Machine Learning in Advancing Patient Care: Bridging Innovations in Mental Health and Cognitive NeuroscienceView all 10 articles
Assessing the Adversarial Robustness of Multimodal Medical AI Systems: Insights into Vulnerabilities and Modality Interactions
Provisionally accepted- 1Innopolis University, Innopolis, Russia
- 2Zayed University, Abu Dhabi, United Arab Emirates
- 3University of Hull, Hull, Yorkshire and the Humber, United Kingdom
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Deep learning systems have been expanding their potential to be used with different data types a4 a wide range of fields. Medicine is one of the most compelling examples of AI applications due to the numerous tasks and various data types associated with them, such as texts, images, and numerical records. Both task-specific single-modality models and generalpurpose multimodal large models have emerged, presenting new opportunities along with new challenges. Despite the potential of AI systems, regardless of the task, they remain vulnerable to adversarial attacks. These attacks are small perturbations, often imperceptible to the human eye, that are intentionally generated to deceive the model. In high-stakes domains, including healthcare as one of the notable examples, they pose significant risks to model's reliability and consequently to their prospects of being applied in real-world tasks. This is why studies focused on adversarial robustness are of such great importance.In our research, we contribute to the fields of multimodality and adversarial robustness, with a specific focus on applications in medicine. As we investigate the behavior of multimodal models under various attack scenarios, this research sheds light on how these models endure attacks and whether the multimodal nature of these models enhances their resilience. We conducted experiments by applying attacks to two modalities: images and texts. We applied attacks to single-modality models, combined these models with modality fusion techniques and applied the same attacks on the combined model. This approach allowed us to validate our hypothesis that multimodality enhances robustness to adversarial attacks. We believe that our findings have the potential to open new horizons in the field of robust multimodality systems. Additionally, this study can initiate further research focused on the data flow in multimodal systems.
Keywords: machine learning (ML), Adversarial attack, multimodal data fusion, Classification, x-ray
Received: 04 Apr 2025; Accepted: 30 Jun 2025.
Copyright: © 2025 Mozhegova, Khattak, Khan, Garaev and Rasheed. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence: Ekaterina Mozhegova, Innopolis University, Innopolis, Russia
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.