Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Radiol.

Sec. Artificial Intelligence in Radiology

Volume 5 - 2025 | doi: 10.3389/fradi.2025.1672364

Integrating Clinical Indications and Patient Demographics for Multi-label Abnormality Classification and Automated Report Generation in 3D Chest CT Scans

Provisionally accepted
Theo  Di PiazzaTheo Di Piazza1,2,3*Carole  LazarusCarole Lazarus4Olivier  NempontOlivier Nempont5Loic  BousselLoic Boussel2,3
  • 1Univ Lyon, INSA-Lyon, INRAE, BF2I, UMR 203, 69621, Villeurbanne, France
  • 2Centre de Recherche en Acquisition et Traitement de l'Image pour la Sante, Villeurbanne, France
  • 3Hospices Civils de Lyon, Lyon, France
  • 4Philips Healthcare Nederland, Eindhoven, Netherlands
  • 5Philips Innovation Hub Eindhoven, Eindhoven, Netherlands

The final, formatted version of the article will be published soon.

The increasing number of Computed Tomography (CT) scan examinations, combined with the time-intensive nature of manual analysis, necessitates efficient automated methods to assist radiologists to manage their growing workload. While deep learning approaches primarily classify abnormalities from three-dimensional (3D) CT images, radiologists also incorporate clinical indications and patient demographics, such as age and sex, for diagnosis. This study aims to enhance multi-label abnormality classification and automated report generation by integrating imaging and non-imaging data. We propose a multimodal deep learning model that combines 3D chest CT scans, clinical information reports, patient age, and sex to improve diagnostic accuracy. Our method extracts visual features from 3D volumes using a visual encoder, textual features from clinical indications via a pre-trained language model, and demographic features through a lightweight feedforward neural network. These extracted features are projected into a shared representation space, concatenated, and processed by a projection head to predict abnormalities. For the multi-label classification task, incorporating clinical indications and patient demographics into an existing visual encoder, called CT-Net, improves the F1-score to 51.58, representing a +∆6.13% increase over CT-Net alone. In the automated report generation task, we extend two existing methods, CT2Rep and CT-AGRG, by integrating clinical indications and demographic data. This integration enhances Clinical Efficacy metrics, yielding an F1-score improvement of +∆14.78% for the CT2Rep extension and +∆6.69% for the CT-AGRG extension. Our findings suggest that incorporating patient demographics and clinical information into deep learning frameworks can significantly improve automated CT scan analysis. This approach has the potential to enhance radiological workflows and facilitate more comprehensive and accurate abnormality detection in clinical practice.

Keywords: Abnormality classification, Report generation, multimodal, 3d ct scans, Clinical indications, patient demographics

Received: 24 Jul 2025; Accepted: 30 Sep 2025.

Copyright: © 2025 Di Piazza, Lazarus, Nempont and Boussel. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Theo Di Piazza, theo.dipiazza@creatis.insa-lyon.fr

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.