Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Oncol.

Sec. Cancer Imaging and Image-directed Interventions

Volume 15 - 2025 | doi: 10.3389/fonc.2025.1643208

This article is part of the Research TopicAI Innovations in Neuroimaging: Transforming Brain AnalysisView all 8 articles

Application and Improvement of YOLO11 for Brain Tumor Detection in Medical Images

Provisionally accepted
Weijuan  HanWeijuan Han1*Xinjie  DongXinjie Dong2Guixia  WangGuixia Wang1Yuwen  DingYuwen Ding3Aolin  YangAolin Yang4
  • 1Zhongyuan Institute of Science and Technology, Zhengzhou, China
  • 2Henan Public Security Department, zhengzhou, China
  • 3Henan Medical College, Zhengzhou, China
  • 4Henan University of Economics and Law, Zhengzhou, China

The final, formatted version of the article will be published soon.

Brain tumors pose a critical threat to human health, and early detection is essential for improving patient outcomes. While magnetic resonance imaging (MRI) is the standard diagnostic tool, manual interpretation is labor-intensive and subject to human error. Object detection methods based on deep learning, especially the YOLO (You Only Look Once) series, have shown promise for automated tumor detection. However, challenges remain in balancing detection accuracy, computational efficiency, and real-time applicability, particularly in clinical environments with constrained hardware. This study presents two key enhancements to the YOLOv11 architecture aimed at improving brain tumor detection from MRI images. First, we integrated a set of novel attention modules (Shuffle3D and Dualchannel attention) into the network to enhance its feature extraction capability. Second, we modified the loss function by combining the Complete Intersection over Union (CIoU) with a Hook function (HKCIoU) to better penalize low-quality predictions and accelerate convergence. Experiments conducted on a public Kaggle dataset demonstrated that our improved model reduced parameters and computations by 2.7% and 7.8%, respectively, while achieving mAP50 and mAP50-95 improvements of 1.0% and 1.4%, respectively, over the baseline. Comparative analysis with existing models validated the robustness and accuracy of our approach. The model's balance of detection accuracy and computational efficiency renders it particularly suitable for deployment in clinical environments with hardware constraints.

Keywords: brain tumor, object detection, You Only Look Once (YOLO), Attention, Intersection over Union (IoU), mean average precision (MAP), Giga Floating point Operations Per Second (GFLOPs)

Received: 10 Jun 2025; Accepted: 11 Aug 2025.

Copyright: © 2025 Han, Dong, Wang, Ding and Yang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Weijuan Han, Zhongyuan Institute of Science and Technology, Zhengzhou, China

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.