Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Med.

Sec. Precision Medicine

This article is part of the Research TopicAI, Bioinformatics, and In Silico Based Biomarkers for the Diagnosis and Treatment of Neurodegenerative DiseasesView all articles

Semiautomated breast ultrasound report generation using multimodal large language models and deep learning

Provisionally accepted
Khadija  AzharKhadija Azhar1Byoung-Dai  LeeByoung-Dai Lee2Shi Sub  ByonShi Sub Byon1SeungJae  LeeSeungJae Lee3Kyu Ran  ChoKyu Ran Cho4Sung Eun  SongSung Eun Song4*
  • 1AI Laboratory, HealthHub Co., Ltd, Seoul, Republic of Korea
  • 2Division of AI and Computer Engineering, Kyonggi University, Suwon-si, Republic of Korea
  • 3Dnotitia Inc., Seoul, Republic of Korea
  • 4Department of Radiology, Korea University Anam Hospital, Seoul, Republic of Korea

The final, formatted version of the article will be published soon.

Introduction: Breast ultrasound (US) imaging is essential for early breast cancer detection, yet generating diagnostic reports is labor-intensive, particularly when incorporating multimodal elastography. Methods: This study presents a novel framework that combines multimodal large language models and deep learning to generate semiautomated breast US reports. This framework bridges the gap between manual and fully automated workflows by integrating radiologist annotations with advanced image classification and structured report compilation. A total of 2,119 elastography images and 60 annotated patient cases were retrospectively collected from two US machines. Results: The system demonstrated robust performance in elastography classification, achieving areas under the receiver operating characteristic curve of 0.92, 0.91, and 0.88 for shear-wave, strain, and Doppler images, respectively. In the evaluated dataset, the report generation module correctly identified all suspicious masses across both US machines, achieving 100% sensitivity in lesion detection, with an average report generation time of 31 seconds per patient using the GE Healthcare machine and 36 seconds using the Supersonic Image machine. Discussion: The proposed framework enables accurate, efficient, and device-adaptable breast US report generation by combining multimodal DL and prompt-based LLM inference. It significantly reduces radiologist workload and demonstrates potential for scalable deployment in real-world clinical workflows.

Keywords: breast ultrasound, deep learning, elastography, Large Language Model, semiautomated report generation

Received: 04 Aug 2025; Accepted: 02 Jan 2026.

Copyright: © 2026 Azhar, Lee, Byon, Lee, Cho and Song. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Sung Eun Song

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.