You're viewing our updated article page. If you need more time to adjust, you can return to the old layout.

REVIEW article

Front. Med., 09 January 2026

Sec. Translational Medicine

Volume 12 - 2025 | https://doi.org/10.3389/fmed.2025.1728248

Artificial intelligence and multimodal imaging in orthopaedics: from technological advances to clinical translation

  • 1. Guangdong Medical University, Zhanjiang, China

  • 2. Yue Bei People’s Hospital Postdoctoral Innovation Practice Base, Southern Medical University, Guangzhou, China

Article metrics

View details

1,1k

Views

103

Downloads

Abstract

The integration of multimodal medical imaging with artificial intelligence (AI) is potentially catalysing a paradigm shift in orthopaedic diagnosis and treatment, moving beyond experience-based practices toward intelligent, data-driven precision medicine. This narrative review synthesizes recent key evidence across imaging modalities and AI frameworks, and highlights the translational gap that persists between algorithmic development and real-world clinical implementation. By combining complementary information from X-ray, CT, MRI, PET, ultrasound, and biochemical data, multimodal AI overcomes the inherent limitations of single-modality approaches, enabling more comprehensive structural, functional, and metabolic assessments. Recent advances demonstrate broad applications, including accurate fracture detection and classification, differentiation of benign and malignant bone tumours, quantitative assessment of osteoarthritis, risk prediction for osteoporosis, and intelligent preoperative planning and intraoperative navigation. Moreover, multimodal AI facilitates efficacy prediction and personalised treatment decision-making, positioning future systems as AI-assisted decision-support tools that support surgeons in surgical strategy, implant design, and long-term follow-up. Nevertheless, significant challenges remain, particularly in data heterogeneity, model generalisation, interpretability, and clinical integration. Progress in constructing standardised multimodal databases, developing self-supervised and multi-task learning strategies, and strengthening ethical–regulatory frameworks will be essential for clinical translation. Ultimately, multimodal AI holds immense potential to transition from laboratory validation to routine practice, delivering safer, more efficient, and precise diagnostic and therapeutic solutions for orthopaedic patients.

1 Introduction

Orthopaedic diseases are characterised by high incidence, disability, and recurrence rates, and thus represent a major global public health concern. These conditions frequently lead to functional impairment and impose substantial healthcare and socioeconomic burdens (1, 2). According to the 2023 Global Burden of Disease study published in the Lancet, nearly 170 million people are affected by orthopaedic diseases worldwide each year (3). Among them, fractures, osteoarthritis, and bone tumours are the most prevalent and high-risk conditions. In China, osteoporotic fractures affect 32.1% of individuals aged 65 years and older, while knee osteoarthritis affects over 15.2% of the same population (4, 5). Moreover, the ageing population, coupled with the rising incidence of sports injuries and bone tumours, is intensifying the demand for orthopaedic diagnosis and treatment, thereby generating increasingly complex clinical challenges.

In recent years, AI has begun to play an increasingly important role in orthopaedic practice. Current applications include assisting fracture identification, grading osteoarthritis severity, supporting tumour assessment, and improving preoperative planning. As highlighted by Khojastehnezhad et al. (6), AI has enhanced diagnostic efficiency and consistency across multiple orthopaedic scenarios, although challenges related to data quality, model transparency, and clinical validation still limit widespread adoption. These developments provide essential context for exploring how multimodal imaging integrated with AI may further advance precision orthopaedic care.

Currently, precise diagnosis of orthopaedic diseases remains difficult. Missed or misdiagnosed fractures may result in long-term disability and high societal costs (7). In addition, bone and soft tissue tumours encompass diverse subtypes, and differentiating benign from malignant forms continues to rely heavily on radiologists’ expertise (7). Early-stage osteoarthritis often lacks specific symptoms and reliable diagnostic markers, making timely identification particularly difficult (5). Although DXA-based bone density measurement is the gold standard for osteoporosis assessment, fracture risk is also influenced by factors such as trabecular microarchitecture and bone metabolism (8). Consequently, diagnostic strategies relying on a single modality are increasingly inadequate to meet the requirements of precision treatment.

Medical imaging plays a pivotal role in orthopaedic evaluation, with each modality providing distinct advantages and applications. Specifically, X-rays are simple, rapid, and cost-effective, making them well suited for initial screening and follow-up (9). CT offers high spatial resolution for delineating bony structures and is widely employed in assessing complex fractures and in preoperative 3D reconstruction (10). MRI, with superior sensitivity to soft tissue, is particularly valuable for assessing cartilage, ligaments, and tumour invasion (11). Ultrasound is safe, real-time, and well suited for detecting superficial joint synovial lesions (12). PET/CT integrates anatomical and metabolic information, thereby providing unique value in assessing tumour activity and metastasis (13). DXA remains the standard for diagnosing osteoporosis; however, it measures only bone mass and fails to reflect bone quality or structural heterogeneity (14). Nevertheless, reliance on a single modality cannot comprehensively capture the structural, biochemical, and functional features of complex bone diseases, thereby restricting its utility in precision medicine.

In parallel, the rapid advancement of AI, particularly deep learning, has consequently driven a disruptive transformation in medical image analysis (15). In recent years, applications of AI in orthopaedics have expanded markedly. These applications encompass lesion detection, tissue segmentation, disease classification, and preoperative risk assessment. Moreover, integrating AI has enhanced the efficiency and consistency of image interpretation; however, evidence for direct improvements in patient outcomes remains limited and is largely inferential (16, 17). However, most current AI models are constrained by reliance on single-modality data, leading to redundancy and limited generalisability. For example, fracture detection models trained on X-rays perform poorly on complex 3D anatomy, while MRI-based soft tissue segmentation models cannot be directly transferred to CT environments (18). Where applicable, we explicitly distinguish research-stage multimodal models from FDA/CE-cleared tools and note that evidence for improved patient outcomes remains limited.

To overcome these limitations, a novel fusion strategy combining multimodal imaging with AI has emerged. By integrating heterogeneous imaging information from CT, MRI, PET, and other modalities at the data, feature, or decision levels, AI systems emulate clinicians’ ability to comprehensively interpret multimodal images and clinical data. This approach holds strong potential to enhance disease recognition, predict risk stratification, and guide personalised treatment (19). Notably, multimodal AI has already been applied in clinical contexts, including fracture classification, benign–malignant tumour differentiation, joint degeneration grading, osteoporosis risk prediction, intraoperative navigation, and postoperative complication assessment. These applications have demonstrated promising adaptability and considerable clinical potential (20, 21). At the same time, challenges including cross-modal coordination, model interpretability, generalisation, and clinical feasibility are attracting growing attention (22).

Hence, this review systematically evaluates advancements in multimodal imaging integrated with AI technologies for orthopaedic disease diagnosis and therapeutic interventions. It focuses on core technical frameworks and integration strategies, with particular emphasis on practical applications such as fractures, osteoarthritis, bone tumours, osteoporosis, and intraoperative navigation. The evaluation further considers research quality, clinical translation value, and key challenges. Furthermore, it explores future directions, including model generalisation, fairness, interpretability, and the potential of multimodal AI in advancing smart orthopaedics and precision medicine. Ultimately, this review seeks to provide theoretical insights and practical guidance for clinical applications, thereby promoting the high-quality development and standardised implementation of AI in orthopaedic imaging (Figure 1).

Figure 1

Diagram depicting the integration of AI with multimodal medical imaging. Central AI label surrounded by imaging modalities: X-ray, CT, MRI, Ultrasound, PET-CT, and DXA. Outer ring illustrates applications, including fracture detection, lesion assessment, bone tumor differentiation, osteoporosis screening, and therapeutic decision support.

Schematic representation of multimodal AI integration in orthopaedic diagnosis and treatment, highlighting the role of AI in analyzing imaging data and assisting in decision-making.

2 Literature search and selection

A focused literature search was conducted using the PubMed, Scopus, and Web of Science databases for studies published between 2015 and 2025. Search terms included combinations of “multimodal imaging,” “artificial intelligence,” “orthopaedics,” “deep learning,” “X-ray,” “CT,” and “MRI.” Studies were included if they: (1) reported applications of AI to musculoskeletal imaging or orthopaedic clinical workflows; (2) involved human subjects or clinically relevant imaging datasets; and (3) were published in English. Exclusion criteria included: (1) studies unrelated to orthopaedic conditions; (2) papers without accessible full text; and (3) purely technical reports lacking clinical relevance.

3 Multimodal imaging and artificial intelligence technologies

3.1 Types of multimodal imaging

Multimodal medical imaging integrates data from X-ray, CT, MRI, PET, SPECT, and ultrasound, thereby enabling a comprehensive structural, functional, and metabolic assessment of orthopaedic diseases (23) (Table 1). Each modality provides complementary strengths—ranging from high spatial resolution and advanced tissue discrimination to rapid acquisition and functional imaging—thereby extending diagnostic capability. For instance, X-rays and CT delineate bone structures, MRI excels in detecting marrow lesions and soft-tissue pathology, while PET highlights metabolic activity in bone tumours (24). In clinical practice, registration and fusion align multimodal data for simultaneous structural and functional visualisation, which in turn improves diagnostic accuracy and streamlines surgical planning (25). Notably, CT–MRI fusion is validated for complex spinal navigation (26), whereas PET–CT remains indispensable for tumour staging and metastasis evaluation (11).

Table 1

Imaging modality Spatial resolution Soft tissue contrast Cost & accessibility Scan duration Radiation dose Clinical advantages Major limitations
X-ray High (0.5–1.0 mm) Low Very low; widely available at primary care level ~few minutes Low First-line for fracture diagnosis; fast and cost-effective 2D projection only; poor soft tissue contrast
CT Very high (~0.5 mm) Moderate Moderate; widely available ~10 min Moderate Detailed anatomical structures; essential for complex fractures and preoperative planning Limited soft tissue contrast; higher radiation than X-ray
MRI Moderate (~1–2 mm) Very high High; mainly tertiary hospitals 30–60 min None Excellent soft tissue delineation; evaluation of tumour infiltration and cartilage High cost; long scan time; lower spatial resolution than CT
Ultrasound Moderate (frequency-dependent) Moderate Low; bedside accessible ~30 min None Real-time dynamic assessment; no radiation; suitable for superficial lesions and joint effusion Operator-dependent; cannot assess intraosseous structures
PET/CT Low (PET ~4–6 mm) Very high (functional/metabolic) Very high; limited to specialised centres 90–120 min High Sensitive for tumour staging and bone metastasis detection Low resolution; high cost; significant radiation exposure
DXA Moderate Low Low; widely available ~5–10 min Very low Quantitative bone mineral density; gold standard for osteoporosis diagnosis Cannot assess bone quality or microarchitecture; limited predictive value for fracture risk

Comparative performance of imaging modalities in orthopaedics.

MRI provides superior soft tissue contrast compared with CT, whereas CT achieves higher spatial resolution. PET primarily reflects metabolic activity and requires CT for anatomical localisation. DXA quantifies bone mineral density but does not capture structural or qualitative aspects.

3.2 Alignment and fusion of multimodal images

Robust analysis of multimodal images critically depends on high-precision registration and fusion algorithms. The core objectives are precise spatial alignment (registration) and semantic-level integration (fusion) across modalities. Mutual-information approaches are widely adopted for heterogeneous data (e.g., MRI–CT, PET–CT) and generally robust; however, their performance declines under distortion or poor image quality (27). Moreover, deep-learning registration, particularly unsupervised U-Net or Transformer variants, substantially enhances processing speed and non-rigid accuracy (28, 29). For example, VoxelMorph achieves high-dimensional registration within seconds and is validated for rapid alignment of preoperative spinal CT with intraoperative cone-beam images (30). Because bone–soft-tissue boundaries are indistinct, registration requires boundary-sensitive optimisation; accordingly, attention mechanisms and multi-scale pyramids have become active research directions (31). Fusion strategies have evolved from pixel-level to feature-level and, more recently, to decision-level approaches. While pixel-level fusion enhances detail, it remains sensitive to modality mismatch. By contrast, feature-level fusion preserves complementary information, thereby improving downstream lesion detection. Optimising fusion to balance structural fidelity with semantic compensation is critical for clinically significant tasks such as cartilage-degeneration analysis and tumour-boundary detection.

3.3 Core algorithms in orthopaedic AI

Orthopaedic image analysis relies heavily on deep learning, with convolutional neural networks (CNNs) serving as the mainstay for image recognition (32). Core tasks include lesion detection, anatomical segmentation, disease grading, and decision support before and during surgery. More recently, Vision Transformers have gained momentum due to their superior capacity for global modelling. In fracture detection, YOLO and Faster R-CNN enable rapid and accurate localisation (33, 34). Similarly, U-Net and its variants excel in bone-tumour segmentation, particularly in accuracy and boundary preservation (35). In osteoarthritis grading, multimodal models that integrate imaging with clinical indicators demonstrate promising predictive power (36). Additionally, generative adversarial networks have been increasingly applied to denoising, super-resolution, and few-shot learning in orthopaedic imaging (37).

3.4 Construction of multimodal AI systems

Constructing clinically transferable multimodal AI systems generally involves five stages: data acquisition and preprocessing, feature extraction and alignment, model construction and training, evaluation and validation, and clinical integration. During preprocessing, images from diverse modalities must be normalised, enhanced, and registered to ensure consistency of inputs. Feature extraction must capture complementary spatial resolution and anatomical features across modalities. During training, multi-task learning and attention mechanisms are often employed to improve generalisation. For example, dual-channel networks fusing MRI soft-tissue features with CT bone structures enable structure-aware models through fusion layers (38). Ultimately, final systems require multicentre validation and seamless integration into clinical platforms such as PACS and intraoperative navigation (Figure 2).

Figure 2

Flowchart illustrating a three-step process for clinical decision support using AI. The first step, "Image Acquisition," includes X-ray, CT, and MRI. The second step, "AI Processing," involves data preprocessing, feature extraction, and AI model training and inference. The third step, "Clinical Decision Support," consists of diagnosis, treatment planning, and feedback loops. Each step is connected by arrows.

Workflow of multimodal imaging integration with AI in clinical practice.

4 Applications of multimodal AI in orthopaedic diseases

This section systematically reviews advances in multimodal imaging AI across the entire clinical continuum. Specifically, the scope encompasses disease detection, risk assessment, treatment decision-making, and intraoperative support. The emphasis is placed on core orthopaedic clinical tasks. This section also examines why multimodal fusion outperforms single-modality approaches in diagnostic performance, predictive accuracy, and clinical utility. Each subsection highlights representative clinical scenarios, summarises model strategies, validation methods, and practical applications, and further identifies translational bottlenecks and future directions.

4.1 Fracture detection and classification

Fracture detection and classification represent some of the earliest and most extensively adopted AI applications in orthopaedics. In routine practice, emergency cases are screened using X-rays, whereas complex fractures require CT for detailed anatomical information. However, single-modality imaging has intrinsic limitations, often resulting in missed or misclassified fractures, particularly subtle or complex patterns. Recently, deep-learning models trained on X-rays have been increasingly deployed for fracture recognition. Although these models perform well in selected scenarios, their accuracy declines in the presence of structural overlap or poor image quality (39) (Figure 3). To address these limitations, researchers have fused CT and X-ray data, thereby improving sensitivity and specificity in fracture detection. For example, the FDA-cleared BoneView system automatically annotates suspected fractures on X-rays and cross-validates with CT findings, thereby enhancing detection rates and reporting efficiency in emergency settings. Furthermore, Liu et al. (40) integrated X-rays with demographic and medical history data to predict dissatisfaction after total knee arthroplasty, achieving AUCs of 0.832–0.891 and markedly outperforming models based solely on clinical data. These findings underscore the critical prognostic value of bone-imaging features.

Figure 3

Four X-ray images of a left wrist labeled A to D. Images B and D highlight areas with suspicion of fractures using white squares. Branding for AZMed is visible on images B and D.

Example wrist radiographs illustrating AI-assisted fracture detection. (A,C) Standard radiographs; (B,D) radiographs with AI-based highlighting of suspected fracture regions. Reproduced under terms of the CC-BY license (39). Copyright 2024, BJR.

In the diagnosis of spinal fractures, multimodal strategies likewise demonstrate superior performance. Geng et al. (41) extracted 1,197 radiomics features from multi-sequence MRI (T1/T2) and used an XGBoost classifier to distinguish osteoporotic from metastatic vertebral compression fractures, reaching an external-validation AUC of 0.905; adding clinical features increased AUC to 0.982. Although developed for specific fracture types, the fusion methodology is broadly applicable to general fracture classification. Beyond diagnosis, multimodal AI is progressively expanding into fracture subclassification and treatment decision support. Hip fractures exemplify this trend: X-rays support preliminary classification, whereas CT provides quantitative assessment of displacement and fragment distribution. Fusion AI systems can then estimate the likelihood of successful screw fixation or the need for arthroplasty (42). Several studies also report personalised AI assistants that integrate patient X-rays with basic clinical data to recommend surgical plans, thereby supporting physician decision-making (43).

Notably, some studies extend “multimodal” to multi-view fusion, in which anteroposterior and lateral X-rays serve as complementary inputs within a single modality to enhance performance. In a meta-analysis of seven commercial systems, Husarek et al. found that dual-view inputs significantly increased sensitivity and specificity for long-bone fractures (44). Moreover, collaborative AI–physician interpretation further improves diagnostic accuracy (45). This “generalised multimodal” strategy demonstrates strong feasibility for real-world deployment.

In summary, multimodal AI provides three major advantages in fracture detection and detailed classification: (1) Diagnostic—expanding from single X-ray to multimodal inputs (X-ray + CT/MRI) improves recognition of complex and occult fractures; (2) Predictive—integrating imaging with clinical information increases the accuracy of predicting surgical outcomes and patient responses; and (3) Translational—a small number of AI systems—mostly limited to single-modality or task-specific applications—have obtained FDA/CE clearance, whereas multimodal fusion systems remain primarily at the research or pilot-evaluation stage. Looking forward, advances in data quality and optimisation of multimodal algorithms are expected to enable deeper AI integration across all stages of fracture care.

4.2 Evaluation of bone and joint lesions and intraoperative

Osteoarthritis (OA) is the most common chronic degenerative joint disease, and radiographic assessment continues to play a central role in both diagnosis and monitoring of disease progression. However, conventional radiographic evaluation is inherently subjective and remains heavily dependent on reader expertise. Consequently, these methods often fail to capture early changes, including cartilage degeneration, synovial inflammation, and osteophyte formation (46). Recently, AI, particularly multimodal image fusion, has enabled objective quantitative diagnosis, early prediction, and classification-driven interventions for OA.

For early diagnosis, AI models automatically detect features such as joint-space narrowing, osteophyte formation, and contour changes by analysing X-rays, MRI, and clinical data. As a result, this approach significantly enhances diagnostic consistency. Notably, deep learning–based automated Kellgren–Lawrence (KL) grading models achieve an AUC of 0.92 in external validation, surpassing the diagnostic consistency of most primary radiologists (47) (Figure 4). Moreover, MRI remains pivotal for early OA assessment owing to its sensitivity to soft tissues including cartilage, synovium, and bone marrow lesions. For example, Khan et al. (48) developed a multimodal CNN that integrates T1/T2-weighted MRI with serum inflammatory markers, thereby enabling personalised prediction in high-risk OA populations.

Figure 4

Two panels, A and B, show medical X-rays of knees. Each panel has two rows: the top with X-rays and the bottom with heat maps indicating analysis. Below are ROC (Receiver Operating Characteristic) and Precision-Recall curves in panels C, D, and E. Panel C shows ROC for different KL grades; panel D highlights a single ROC curve; panel E presents a precision-recall curve. Legends identify KL grades and observer roles like radiology residents and musculoskeletal radiologists.

Deep learning–based Kellgren–Lawrence (KL) grading of knee osteoarthritis. (A,B) Example radiographs with corresponding model attention heatmaps and predicted KL probabilities. (C–E) Diagnostic performance of the deep learning model compared with human readers, illustrated by receiver operating characteristic (ROC) and precision–recall curves. Reproduced under terms of the CC-BY-NC-ND license (47). Copyright 2025, Osteoarthritis and Cartilage.

In disease staging and progression monitoring, multimodal AI systems integrating structural and functional imaging (e.g., quantitative MRI, PET–MRI) represent a major advance. These systems track temporal changes in cartilage volume, T2-mapping signals, and bone marrow oedema, thereby allowing prediction of disease progression trajectories (49). Lee et al. (50) developed an OA progression model that integrates MRI with clinical data (e.g., gait velocity), predicting KL grade changes over 4–5 years with an AUC of 0.79–0.82. This model further demonstrated strong temporal stability and generalisability. By contrast, this dynamic approach outperforms static KL grading and facilitates timely identification of rapidly progressing patients.

Furthermore, AI is increasingly applied to personalised treatment decision support in OA. Several models integrate age, BMI, activity, joint alignment, and radiographic severity to predict the efficacy of non-surgical interventions (e.g., intra-articular injections, rehabilitation) or determine the need for knee replacement. Chen et al. (51) reported that multimodal AI–guided treatment achieved significantly greater one-year functional improvement than experience-based decision-making (17.6% improvement, p < 0.05). Importantly, AI applications in OA are now showing potential for future clinical applicability, although most systems remain at the research stage. For instance, the Stanford OA AI platform integrates automated MRI segmentation with OA risk scoring, thereby enabling automated structural assessment. Moreover, several products have already gained FDA approval for OA progression monitoring and preoperative intervention assessment (52, 53). Collectively, these developments highlight multimodal AI as a critical tool for long-term OA management and precise staging.

In summary, multimodal AI has shifted OA care from “static image-assisted judgment” to “dynamic progression prediction,” and ultimately to “personalised intervention planning.” This evolution enhances early detection and provides strong support for establishing personalised and proactive management models for OA.

4.3 Bone tumor detection and benign/malignant diagnosis

Bone tumours, with their low incidence, heterogeneous subtypes, and diverse imaging features, have traditionally depended on the subjective interpretation of senior radiologists (54). Although X-ray, CT, and MRI are essential for localisation and morphological assessment, they remain inadequate for differentiating benign from malignant tumours, grading lesions, and delineating margins (55). Recently, advances in AI-driven multimodal imaging have created new opportunities for early detection and precise characterisation of bone tumours.

At the screening and localisation stage, AI models trained on large-scale bone tumour X-rays or CT scans can rapidly detect and automatically annotate suspected lesions. Numerous studies have applied CNNs to long-bone X-rays, achieving detection sensitivities above 90% (56). AI-based bone scan tools facilitate the detection of multiple bone metastases, thereby reducing misdiagnosis and missed diagnoses. These tools are particularly valuable for patients at high risk of metastasis, such as those with breast or prostate cancer (57).

To differentiate benign from malignant lesions and determine tumour grade, multimodal AI models achieve superior accuracy. Compared with single-modality analysis, combined MRI–CT evaluation captures structural, signal, and margin variations, thereby markedly enhancing classification performance (58). Zhou et al. (59) designed a three-channel deep learning model that integrates T1-weighted MRI, T2-weighted MRI, and contrast-enhanced CT, achieving an AUC of 0.94 for benign–malignant classification and substantially outperforming single-modality models. Furthermore, radiomics approaches integrated with AI extract hundreds of latent features (e.g., texture, morphology, grey-scale distribution) and combine them with clinical variables (e.g., age, lesion site). Gao et al. (60) established an MRI radiomics model that distinguished osteosarcoma from chondroma with >92% accuracy, offering auxiliary diagnostic support in cases with ambiguous benign–malignant boundaries (Figure 5).

Figure 5

A four-step workflow in medical imaging analysis. A shows image acquisition with two MRI scans. B depicts tumor segmentation with highlighted red areas in two images. C illustrates feature extraction with a 3D scatterplot and a histogram. D shows model assessment using ROC curves, with graphs for training and validation groups, displaying AUC values for T1 and T2 sequences.

Schematic representation of the radiomics analysis pipeline for differentiating osteosarcoma and chondrosarcoma based on MRI. The process involves: (A) image acquisition, (B) tumour segmentation, (C) feature extraction, and (D) model assessment using receiver operating characteristic analysis in training and validation cohorts. Reproduced under terms of the CC-BY-NC-ND license (60). Copyright 2024, Scientific Reports.

Moreover, AI is becoming increasingly important for histopathological pre-diagnosis and treatment planning. Several studies have explored the prediction of tumour molecular subtypes (e.g., IDH mutation), malignancy grade, and therapy sensitivity using imaging features, thereby providing critical data support for precision medicine (61). Multimodal AI systems now integrate PET–CT functional imaging into workflows, enabling simultaneous analysis of metabolic activity and structural features to optimise biopsy targeting and predict treatment efficacy (62). In addition, AI is accelerating the clinical translation of bone tumour imaging. For example, OsteoDetect provides real-time marking of potential lesions on X-rays, thereby alerting clinicians to suspicious areas (44). Furthermore, high-risk scoring models have been embedded into PACS systems, automatically generating diagnostic recommendations to enhance efficiency and consistency (63). Importantly, these tools are not intended to replace physicians but to strengthen human–AI collaboration by providing structured, quantifiable data support.

In summary, AI-driven bone tumour imaging is rapidly evolving from lesion identification to benign–malignant classification, molecular prognosis prediction, and precise intervention planning. The integration of multimodal AI not only improves diagnostic reliability for complex lesions but also provides a robust foundation for developing personalised treatment strategies. Ultimately, this advancement is poised to usher bone tumour diagnosis and treatment into a new era of intelligent and precise medicine.

4.4 Osteoporosis risk prediction and screening

Osteoporosis is a systemic metabolic bone disease marked by reduced bone mass and deterioration of bone microarchitecture. This condition poses a major threat to both quality of life and physical safety in elderly populations. Although DXA remains the gold standard, it has clear limitations, including lengthy testing cycles and low sensitivity for subclinical disease (64). Recently, AI, particularly multimodal approaches, has driven breakthroughs in early identification, risk prediction, and personalised management of osteoporosis.

In bone mineral density (BMD) assessment, integrating multimodal AI with imaging and clinical data has markedly improved predictive accuracy. Notably, several studies have demonstrated “imaging-based BMD estimation” without the need for additional radiation exposure. This strategy extracts trabecular texture features from conventional CT images of the lumbar spine and femur and combines them with machine learning models (65). For example, Wang et al. (66) developed an AI model integrating age, sex, BMI, and CT texture features to predict T-scores < −2.5, achieving an AUC of 0.943. This model outperformed those relying solely on DXA, underscoring its potential as a clinical substitute.

Moreover, multimodal AI has also been widely applied to prospective fracture risk prediction. By integrating multidimensional data—including DXA, X-rays, medical history, exercise habits, and biochemical markers—AI identifies high-risk individuals often missed by traditional tools such as FRAX (67, 68). In particular, Kong et al. (69) developed a deep learning network combining spinal X-rays with fracture history, predicting fragility fracture risk within three years with sensitivity above 90%. These models are especially valuable for identifying individuals with “normal BMD but high fragility,” representing a hidden osteoporotic population. Furthermore, advances in AI are enabling personalised management of osteoporosis. By continuously monitoring medication adherence, BMD trajectories, and metabolic indicators, AI models can predict drug response and fracture recurrence risk, thereby enabling dynamic treatment adjustment (70). Zhou et al. (71) demonstrated that a temporal AI model trained on multimodal data predicted one-year BMD trends after six months of medication, achieving accuracy above 85% (Figure 6). Looking ahead, integrating wearable devices and home imaging could make AI-enabled dynamic bone health management more accessible.

Figure 6

Flowchart showing a process for analyzing abdominal X-ray images. It starts with an X-ray and proceeds to a cropped image, followed by U-net model application for soft tissue elimination. Regions of interest (ROIs) are selected as whole, vertical, and horizontal areas. Cropped ROIs undergo multi-threshold addition. Biological parameters and extracted features are inputted into an ANN model to derive a BMD value.

Overall workflow of the proposed U-Net-based deep learning hybrid model for predicting bone mineral density (BMD) from abdominal X-ray images. The process involves image preprocessing using U-Net for soft tissue elimination, region-of-interest (ROI) selection, multi-threshold feature extraction, and integration of biological parameters into an artificial neural network (ANN) to output predicted BMD values. Reproduced under terms of the CC-BY license (71). Copyright 2025, Bioengineering.

In summary, multimodal AI is transforming osteoporosis management from static detection to dynamic prediction and from population-level screening to individualised care. Ultimately, this advancement enhances risk identification and management precision, providing technological support for bone health strategies in ageing societies.

4.5 Intelligent preoperative assessment and intraoperative navigation systems

Preoperative planning and intraoperative navigation are indispensable for precision surgery, particularly in complex orthopaedic procedures, as they improve safety and reduce postoperative complications. Conventional methods, which rely on two-dimensional imaging and surgeon experience, restrict intuitive three-dimensional reconstruction and real-time adaptation to intraoperative changes (72). Recently, the emergence of multimodal AI has driven a paradigm shift in orthopaedic surgery. The integration of CT, MRI, preoperative X-rays, 3D modelling, and real-time intraoperative imaging has enabled intelligent navigation systems defined by precision, adaptability, and personalisation (73).

In preoperative modelling, AI markedly improves the automation and accuracy of three-dimensional reconstruction. By integrating CT and MRI with deep learning segmentation networks, AI swiftly delineates skeletal, soft tissue, and lesion boundaries, thereby generating patient-specific 3D anatomical models (74). For example, Lu et al. (75) developed an AI-assisted system that enabled rapid modelling and force-line assessment in complex pelvic fractures. This system substantially reduced modelling time while maintaining anatomical consistency scores comparable to manual modelling. Moreover, such systems automatically calculate key intraoperative parameters (e.g., osteotomy angles, implant trajectories), thereby providing quantitative support for optimising preoperative plans (76) (Figure 7).

Figure 7

A collection of medical images depicting various perspectives of pelvic anatomy and surgical procedures. Images A to D show different scans of the pelvis, including X-rays and MRI. Images E to H illustrate 3D models of pelvic bones with colored lines indicating surgical points or axes. Images I to L depict diagrams and real-life photos of surgical procedures on the abdomen. Images M to P show detailed diagrams of pelvic surgical interventions with color-coded components. Images Q to T display surgical instruments and models on a green surgical drape. Images U and V show a patient undergoing a surgical procedure. Images W and X include fluoroscopy images of surgical tools in the pelvic region.

AI-assisted preoperative planning and 3D-printing guiding frame for percutaneous screw reconstruction in periacetabular metastatic cancer patients. (A–D) Preoperative imaging (X-ray, CT, 3D reconstruction, MRI). (E–H) Convolutional neural network–based determination of screw entry points and trajectories. (I–L) Skin surface recognition algorithm for personalized anchoring pads. (M–P) Guiding frame generation with trajectory tubes and anchoring feet. (Q–T) 3D-printed components and assembled guiding frame. (U–X) Intraoperative application, CT and fluoroscopic validation, and postoperative screw fixation. Reproduced under terms of the CC-BY license (77). Copyright 2024, Frontiers in Bioengineering and Biotechnology.

Notably, AI-enabled intraoperative navigation systems increasingly achieve real-time fusion of multimodal images. Conventional systems primarily rely on CT registration but suffer from accuracy limitations caused by soft tissue occlusion and bone displacement. By integrating intraoperative fluoroscopy, preoperative CT/MRI, and optical tracking, multimodal AI navigation constructs dynamically updated surgical views. In particular, Zhang et al. (77) developed an AI navigation platform that achieved submillimetre accuracy in pedicle screw placement. The system aligns intraoperative C-arm images with preoperative 3D models and corrects patient drift in real time, thereby reducing penetration failure rates significantly.

Additionally, augmented reality (AR) and virtual reality (VR) (78), enhanced by AI, are promoting seamless integration between preoperative planning and intraoperative execution. Multimodal inputs—including skeletal CT, vascular MRI, and functional imaging—create interactive virtual environments that support preoperative rehearsal and pathway simulation. During surgery, AR devices overlay AI-generated surgical plans onto the patient’s body in real time, thereby delivering navigation consistent with physical reality. Research shows that AR-assisted total knee arthroplasty outperforms conventional methods in osteotomy precision, prosthesis alignment, and operative time, thereby highlighting its clinical potential (79, 80).

In summary, multimodal AI has established an integrated workflow spanning static modelling to dynamic navigation for preoperative planning and intraoperative guidance. Ultimately, this advancement enhances surgical precision and safety and lays the groundwork for future personalised orthopaedic surgery.

4.6 Efficacy prediction and intelligent treatment decision-making

Within precision medicine, efficacy prediction and personalised treatment are becoming increasingly central to orthopaedic clinical decision-making. Conventional efficacy assessments still depend on postoperative imaging and subjective scoring, yet they lack feedback mechanisms to optimise preoperative interventions. The emergence of multimodal AI now enables efficacy prediction by integrating imaging, clinical, and bioinformatics data, thereby facilitating both the formulation and dynamic adjustment of personalised treatment strategies.

AI constructs predictive models by integrating diverse preoperative factors, including imaging data, biochemical markers, and electronic medical records. This approach enables quantitative assessment of postoperative recovery outcomes (81). For example, in hip replacement, Zhang et al. (82) developed a multimodal predictive model that integrates CT/MRI radiomic features with clinical indicators such as demographics and preoperative pain scores. This model accurately predicted postoperative functional improvement (e.g., Harris and WOMAC scores) and demonstrated markedly superior performance compared with single-modality models. Consequently, such preoperative risk–benefit assessment provides valuable data to support individualised surgical planning, recovery timeline estimation, and postoperative management.

Furthermore, AI contributes to postoperative rehabilitation monitoring and supports dynamic adjustment. By integrating postoperative X-rays, MRI, serum inflammatory markers, and gait data, AI models can detect deviations from rehabilitation trajectories, thereby issuing timely alerts for functional impairment or prosthetic malalignment (83). Notably, Lee et al. (84) demonstrated that AI-based camera gait analysis generates quantitative metrics strongly correlated with Harris hip scores, thereby enabling reliable prediction of postoperative recovery. Concurrently, Wang et al. (85) developed an AI framework that combines CT registration with personalised dynamic modelling, improving hip kinematic measurement accuracy and supporting personalised rehabilitation timeline prediction.

Multimodal AI is widely applied in combined radiomics and molecular analyses to assess treatment sensitivity and drug resistance. For example, to predict the efficacy of biologic therapies for rheumatoid arthritis, researchers combined ultrasound data with haematological markers and transcriptomic/multi-omics features in machine learning models. This approach successfully distinguished responders from non-responders, demonstrating high discriminatory power in both prospective and retrospective cohorts (86, 87). It has also shown promise for assessing chemotherapy response in bone tumours and guiding personalised management of bone infections. Radiomics and radiogenomics models support preoperative efficacy evaluation, while AI integration into diagnostic and therapeutic workflows for bone and joint infections (7, 88, 89) may ultimately advance “treatment on demand.” Furthermore, AI aids personalised implant design and surgical pathway planning. By considering patient bone density, alignment parameters, and anatomical structures, AI generates customised implant geometries and performs virtual placement and biomechanical simulations. In addition, when integrated with navigation or robotic systems, AI further enhances intraoperative precision. Moreover, AI enables dynamic adjustment of rehabilitation protocols through closed-loop systems that incorporate wearable or “smart” implants and remote monitoring (90–92) (Figure 8).

Figure 8

(A) Shows a man in black shorts standing front and back, and walking on a treadmill. (B) Displays black sneakers with electronic sensors. (C) MRI images labeled a and b, showing spinal features with measurements. (D) Depicts a person using VR goggles and interacting with a system involving sensors, LiDAR, and motion capture technology.

Applications of U-Net-Based Deep Learning Hybrid Model. (A) Gait analysis dataset generation using self-selected speed treadmill walking for 3D kinematic and dynamic assessment. (B) Wearable six-axis inertial sensors attached to shoes for stride speed and bilateral gait parameter collection. (C) Lumbar T2-weighted MRI analysis for postoperative rehabilitation monitoring, highlighting multifidus muscle cross-sectional area and related parameters. (D) Immersive virtual reality rehabilitation robot integrating VR goggles, force sensors, LiDAR, 3D motion capture, and DC motors for personalized training and real-time feedback. Reproduced under terms of the CC-BY license (91). Copyright 2025, MedComm.

In conclusion, multimodal AI is transforming traditional “experience-driven” treatment into “data-driven, mechanism-oriented, and individualised” approaches. Its integration into preoperative efficacy prediction, intraoperative pathway optimisation, and postoperative rehabilitation tracking provides the foundation for truly personalised orthopaedic treatment.

5 Challenges and perspectives

Despite notable progress in applying multimodal imaging and AI to orthopaedic diagnosis and treatment, widespread clinical adoption continues to face substantial challenges. Future research should therefore prioritise algorithm optimisation, data sharing, model generalisation, and the clinical translation of intelligent orthopaedic systems.

5.1 Data heterogeneity and standardisation challenges

Data heterogeneity is a well-known challenge in the development of AI models, particularly when working with multimodal imaging data from diverse sources. Variability in imaging modalities (e.g., X-ray, CT, MRI, PET), data acquisition protocols (e.g., scanner settings, contrast agents, image resolution), and patient demographics (e.g., age, sex, comorbidities) can create significant biases and inconsistencies in the training and validation datasets (93). These factors can affect the performance of AI models, often leading to overfitting to specific data sets or institutions, reducing the model’s ability to generalize across different clinical environments. Additionally, the lack of data standardization—for example, inconsistent image preprocessing, annotation practices, and quality control measures—further complicates model development. Without uniform standards, AI models may struggle to maintain reliable accuracy when applied to new, unseen data. This issue also poses significant barriers to multicenter validation, where datasets from multiple hospitals or imaging centers must be aggregated to assess model performance across different settings. A model trained on data from a single center or imaging modality may fail to perform optimally when deployed in a real-world clinical setting with diverse sources of imaging data. These challenges have significant implications for clinical translation. Even though AI models may perform well in controlled research environments, the variability of real-world clinical data means that models may not achieve the same level of success when deployed in routine clinical practice. The lack of data standardization across institutions means that clinical adoption will require additional efforts for data harmonization and cross-site validation to ensure that AI models can effectively support decision-making in diverse settings.

5.2 Insufficient model generalisation and robustness

Current AI models frequently achieve high performance on specific datasets. However, in real-world clinical settings spanning diverse devices, institutions, and patient populations, accuracy frequently declines due to variation in image quality and case complexity (94). For example, fracture detection models perform well on data from top-tier hospitals but deteriorate significantly on external datasets (e.g., ankle fracture detection AUC decreased from 0.95 in internal validation to 0.86 in external testing), underscoring substantial out-of-domain generalisation challenges (95). To enhance generalisation, strategies such as transfer learning, multi-task learning, and few-shot learning are indispensable. Concurrently, incorporating active learning and reinforcement learning further enhances model adaptability to “unseen scenarios.”

5.3 Explainability and credibility issues

Notably, AI models in orthopaedic imaging face the “black box” challenge, as they cannot provide clinically acceptable interpretation pathways, thereby hindering adoption (96). Comprehensive diagnosis of orthopaedic diseases necessitates integration of clinical manifestations, imaging features, and intraoperative findings. Consequently, verifying both model explainability and credibility is essential (97, 98). Techniques including heatmaps, attention mechanisms, and visualisation modules can significantly enhance interpretability. Furthermore, interdisciplinary collaboration is crucial for establishing evaluation frameworks that ensure clinical pathway consistency.

5.4 Challenges in clinical integration and intelligent workflow embedding

The value of AI models is greatly diminished if they are not effectively embedded into clinical workflows or seamlessly integrated with PACS, preoperative planning platforms, or intraoperative navigation systems (99). In addition, the physician learning curve and entrenched work habits must be carefully considered when introducing AI systems (100). Consequently, future efforts should focus on refining interface design, interactivity, and real-time performance evaluation, while promoting end-to-end AI integration across workflows—from preoperative decision support and intraoperative navigation to postoperative outcome assessment.

5.5 Issues of regulatory ethics and privacy security

Medical AI must strictly adhere to ethical and regulatory frameworks, particularly when processing personal health information, to ensure anonymisation, privacy protection, and compliant data sharing (101). At present, most AI models have not undergone clinical registration, certification, or ethical review during development. This gap poses a major barrier to their translation into medical products (102, 103). Therefore, future work should strengthen regulatory guidance and support for AI compliance pathways, aiming to establish medical device registration systems and multicentre trial standards.

5.6 Future perspectives

Looking ahead, deeper integration of multimodal imaging and AI into orthopaedic care is expected to drive the development of new intelligent precision medicine models. Recent advances in lightweight neural networks and edge computing have catalysed a paradigm shift, enabling AI integration into mobile devices and intraoperative navigation platforms. This integration could transform healthcare by delivering real-time intelligent assistance at the clinical frontline. Concurrently, task-specific models for fractures, bone tumours, and joint degeneration will further enhance AI specialisation and clinical interpretability. Ultimately, intelligent orthopaedics is expected to evolve from single-purpose diagnostic tools to comprehensive Clinical Decision Support Systems. A defining feature of these systems is their integration of multi-source heterogeneous data, including imaging, pathology, physiological signals, and electronic health records. Such integration will enable early disease detection, dynamic monitoring, and personalised treatment recommendations. In the longer term, digital twins, disease progression modelling, and advanced visualisation platforms may enable the creation of “virtual skeletal individuals.” These could support personalised preoperative planning, predictive analytics, and long-term efficacy assessment. Achieving this goal will require interdisciplinary collaboration and large-scale sharing and integration of real-world data across multiple centres. At the same time, establishing clinical translation standards and robust ethical-regulatory frameworks is essential to ensure safe, effective, and continuously evolving AI applications in orthopaedic diagnosis and treatment.

6 Conclusion

The convergence of multimodal imaging and artificial intelligence is reshaping orthopaedic diagnosis and treatment, shifting from experience-driven practice to intelligent, data-driven precision medicine. Evidence indicates that multimodal AI enhances fracture detection, tumour characterisation, osteoarthritis evaluation, osteoporosis risk prediction, and surgical planning, underscoring its broad clinical potential. Nevertheless, challenges remain in data integration, model robustness, interpretability, and clinical translation. Future systems are expected not only to automate lesion detection and structural measurement but also to support treatment pathway decision-making, potentially evolving into AI-assisted decision-support tools for orthopaedic surgeons. To achieve this transformation, advances in standardised databases, self-supervised learning, model fairness, and real-world validation frameworks will be essential. Ultimately, multimodal AI holds immense promise to transition from laboratory validation to routine clinical application, enabling safer, more efficient, and more precise solutions for orthopaedic patients.

Statements

Author contributions

GL: Conceptualization, Formal analysis, Investigation, Methodology, Visualization, Writing – original draft, Writing – review & editing. ST: Investigation, Resources, Software, Writing – original draft. LL: Data curation, Resources, Validation, Visualization, Writing – review & editing. KH: Resources, Supervision, Validation, Writing – review & editing.

Funding

The author(s) declared that financial support was received for this work and/or its publication. This work was supported by the China Postdoctoral Science Foundation (Certificate Number: 2025M781412).

Acknowledgments

All authors actively participated in the research process, made significant contributions to the revised drafts, and meticulously reviewed and approved the final version of the manuscript.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that Generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1.

    GBD 2019 Diseases and Injuries Collaborators . Global burden of 369 diseases and injuries in 204 countries and territories, 1990-2019: a systematic analysis for the global burden of disease study 2019. Lancet. (2020) 396:120422.

  • 2.

    Kanis JA Cooper C Rizzoli R Reginster JY . European guidance for the diagnosis and management of osteoporosis in postmenopausal women. Osteoporos Int. (2019) 30:344. doi: 10.1007/s00198-018-4704-5,

  • 3.

    Global, regional, and national burden of osteoarthritis, 1990-2020 and projections to 2050: a systematic analysis for the global burden of disease study 2021. Lancet Rheumatol. (2023) 5:e50822.

  • 4.

    Meng S Tong M Yu Y Cao Y Tang B Shi X et al . The prevalence of osteoporotic fractures in the elderly in China: a systematic review and meta-analysis. J Orthop Surg Res. (2023) 18:536. doi: 10.1186/s13018-023-04030-x,

  • 5.

    Cui A Li H Wang D Zhong J Chen Y Lu H . Global, regional prevalence, incidence and risk factors of knee osteoarthritis in population-based studies. EClinicalMedicine. (2020) 29-30:100587. doi: 10.1016/j.eclinm.2020.100587,

  • 6.

    Khojastehnezhad MA Youseflee P Moradi A Ebrahimzadeh MH Jirofti N . Artificial intelligence and the state of the art of orthopedic surgery. Archives Bone Joint Surgery. (2025) 13:17.

  • 7.

    Spaanderman DJ Marzetti M Wan X Scarsbrook AF Robinson P Oei EHG et al . AI in radiological imaging of soft-tissue and bone tumours: a systematic review evaluating against CLAIM and FUTURE-AI guidelines. EBioMedicine. (2025) 114:105642. doi: 10.1016/j.ebiom.2025.105642,

  • 8.

    Ulivieri FM Messina C Vitale FM Rinaudo L Grossi E . Artificial intelligence for predicting the risk of bone fragility fractures in osteoporosis. Eur Radiol Exp. (2025) 9:62. doi: 10.1186/s41747-025-00572-3,

  • 9.

    Requist MR Mills MK Carroll KL Lenz AL . Quantitative skeletal imaging and image-based modeling in pediatric Orthopaedics. Curr Osteoporos Rep. (2024) 22:4455. doi: 10.1007/s11914-023-00845-z,

  • 10.

    Schwarzenberg P Darwiche S Yoon RS Dailey HL . Imaging modalities to assess fracture healing. Curr Osteoporos Rep. (2020) 18:16979. doi: 10.1007/s11914-020-00584-5,

  • 11.

    Papageorgiou PS Christodoulou R Korfiatis P Papagelopoulos DP Papakonstantinou O Pham N et al . Artificial intelligence in primary malignant bone tumor imaging: a narrative review. Diagnostics. (2025) 15. doi: 10.3390/diagnostics15131714,

  • 12.

    D'Agostino V Sorriento A Cafarelli A Donati D Papalexis N Russo A et al . Ultrasound imaging in knee osteoarthritis: current role, recent advancements, and future perspectives. J Clin Med. (2024) 13. doi: 10.3390/jcm13164930,

  • 13.

    Usmani S Ahmed N Gnanasegaran G Marafi F van den Wyngaert T . Update on imaging in chronic kidney disease-mineral and bone disorder: promising role of functional imaging. Skeletal Radiol. (2022) 51:90522. doi: 10.1007/s00256-021-03905-6,

  • 14.

    Alsarraf F Ali DS Brandi ML . The use of bone biomarkers, imaging tools, and genetic tests in the diagnosis of rare bone disorders. Calcif Tissue Int. (2025) 116:32. doi: 10.1007/s00223-024-01323-z,

  • 15.

    Kumar Ale N . The role of artificial intelligence in enhancing test automation: current trends and future directions. Int J Science Res. (2024) 13:19137. doi: 10.21275/SR24717102838

  • 16.

    Mohammed TJ Xinying C Alnoor A Khaw KW Albahri AS Teoh WL et al . A systematic review of artificial intelligence in orthopaedic disease detection: a taxonomy for analysis and trustworthiness evaluation. Int J Comput Intell Syst. (2024) 17. doi: 10.1007/s44196-024-00718-y

  • 17.

    Fritz B Fritz J . Artificial intelligence for MRI diagnosis of joints: a scoping review of the current state-of-the-art of deep learning-based approaches. Skeletal Radiol. (2022) 51:31529. doi: 10.1007/s00256-021-03830-8,

  • 18.

    Zhang Y Ma M Huang X Liu J Tian C Duan Z et al . Machine learning is changing osteoporosis detection: an integrative review. Osteoporos Int. (2025) 36:131326. doi: 10.1007/s00198-025-07541-x,

  • 19.

    da Silva DS Astolfi RS Jagatheesaperumal SK Selvaraj P Russel NS de Albuquerque VHC . Advances in 3D fusion of multimodal medical images: 3D reconstruction of bone, muscle, and ligament structures under load from radiographs and magnetic resonance imaging. Research Biomed Eng. (2025) 41:25. doi: 10.1007/s42600-025-00405-7

  • 20.

    Misir A Yuce A . AI in orthopedic research: a comprehensive review. J Orthop Res. (2025) 43:150827. doi: 10.1002/jor.26109,

  • 21.

    Kumar R Gowda C Sekhar TC Vaja S Hage T Sporn K et al . Advancements in machine learning for precision diagnostics and surgical interventions in interconnected musculoskeletal and visual systems. J Clin Med. (2025) 14. doi: 10.3390/jcm14113669,

  • 22.

    Baghbani S Mehrabi Y Movahedinia M Babaeinejad E Joshaghanian M Amiri S et al . The revolutionary impact of artificial intelligence in orthopedics: comprehensive review of current benefits and challenges. J Robot Surg. (2025) 19:511. doi: 10.1007/s11701-025-02561-5,

  • 23.

    Hussain D Al-Masni MA Aslam M Sadeghi-Niaraki A Hussain J Gu YH et al . Revolutionizing tumor detection and classification in multimodality imaging based on deep learning approaches: methods, applications and limitations. J Xray Sci Technol. (2024) 32:857911. doi: 10.3233/XST-230429,

  • 24.

    Chaudhari AJ Raynor WY Gholamrezanezhad A Werner TJ Rajapakse CS Alavi A . Total-body PET imaging of musculoskeletal disorders. PET Clinics. (2021) 16:99117. doi: 10.1016/j.cpet.2020.09.012,

  • 25.

    Garzia S Capellini K Gasparotti E Pizzuto D Spinelli G Berti S et al . Three-dimensional multi-modality registration for Orthopaedics and cardiovascular settings: state-of-the-art and clinical applications. Sensors. (2024) 24. doi: 10.3390/s24041072,

  • 26.

    Duits AAA van Urk PR Lehr AM Nutzinger D Reijnders MRL Weinans H et al Radiologic assessment of interbody fusion: a systematic review on the use, reliability, and accuracy of current fusion criteria. JBJS Rev2024;12.

  • 27.

    Wang Y Chang W Huang C Kong D . Multiscale unsupervised network for deformable image registration. J Xray Sci Technol. (2024) 32:114. doi: 10.3233/XST-240159,

  • 28.

    Cui X Zhou Y Wei C Suo G Jin F Yang J . Hybrid transformer and convolution iteratively optimized pyramid network for brain large deformation image registration. Sci Rep. (2025) 15:15707. doi: 10.1038/s41598-025-00403-w,

  • 29.

    Anwar M He Z Cao W . Enhancing 3D medical image registration with cross attention, residual skips, and cascade attention. Intell Data Anal. (2024) 29:186201.

  • 30.

    Madge V. Evaluating Voxelmorph: a learning-based 3D non-linear registration algorithm, against the non-linear symmetric normalization technique from ANTs: McGill University (Canada) 2020

  • 31.

    Zeng J Fu Q . A review: artificial intelligence in image-guided spinal surgery. Expert Rev Med Devices. (2024) 21:689700. doi: 10.1080/17434440.2024.2384541,

  • 32.

    Li X Zhang L Yang J Teng FJ . Role of artificial intelligence in medical image analysis: a review of current trends and future directions. J Med Biol Eng. (2024) 44:23143. doi: 10.1007/s40846-024-00863-x

  • 33.

    Tieu A Kroen E Kadish Y Liu Z Patel N Zhou A et al . The role of artificial intelligence in the identification and evaluation of bone fractures. Bioengineering. (2024) 11. doi: 10.3390/bioengineering11040338,

  • 34.

    Guo L Yang H Zhang J Wu H Zhu M Zheng Y et al . The role of deep learning in spinal lesion detection: a systematic review In: 17th international congress on image and signal processing, BioMedical engineering and informatics (CISP-BMEI): IEEE (2024)

  • 35.

    Siouras A Moustakidis S Giannakidis A Chalatsis G Liampas I Vlychou M et al . Knee injury detection using deep learning on MRI studies: a systematic review. Diagnostics. (2022) 12. doi: 10.3390/diagnostics12020537,

  • 36.

    Aladhadh S Mahum RJIA . Knee osteoarthritis detection using an improved CenterNet with pixel-wise voting scheme. IEEE Access. (2023) 11:2228396. doi: 10.1109/ACCESS.2023.3247502

  • 37.

    Rodrigues M Mayo M Patros P . Surgical tool datasets for machine learning research: a survey. Int J Comput Vis. (2022) 130:222248. doi: 10.1007/s11263-022-01640-6

  • 38.

    Fang M Wang Z Pan S Feng X Zhao Y Hou D et al . Large models in medical imaging: advances and prospects. Chin Med J. (2025) 138:164764. doi: 10.1097/CM9.0000000000003699,

  • 39.

    Pauling C Kanber B Arthurs OJ Shelmerdine SC . Commercially available artificial intelligence tools for fracture detection: the evidence. BJR Open. (2024) 6:tzad005. doi: 10.1093/bjro/tzad005,

  • 40.

    Liu X Liu Y Lee ML Hsu W Liow MHL . Identifying who are unlikely to benefit from total knee arthroplasty using machine learning models. NPJ Digit Med. (2024) 7:266. doi: 10.1038/s41746-024-01265-8,

  • 41.

    Geng W Zhu J Li M Pi B Wang X Xing J et al . Radiomics based on multimodal magnetic resonance imaging for the differential diagnosis of benign and malignant vertebral compression fractures. Orthop Surg. (2024) 16:246474. doi: 10.1111/os.14148,

  • 42.

    Gitto S Serpi F Albano D Risoleo G Fusco S Messina C et al . AI applications in musculoskeletal imaging: a narrative review. Eur Radiol Exp. (2024) 8:22. doi: 10.1186/s41747-024-00422-8,

  • 43.

    Liu P Zhang J Liu S Huo T He J Xue M et al . Application of artificial intelligence technology in the field of orthopedics: a narrative review. Artif Intell Rev. (2024) 57. doi: 10.1007/s10462-023-10638-6

  • 44.

    Husarek J Hess S Razaeian S Ruder TD Sehmisch S Muller M et al . Artificial intelligence in commercial fracture detection products: a systematic review and meta-analysis of diagnostic test accuracy. Sci Rep. (2024) 14:23053. doi: 10.1038/s41598-024-73058-8,

  • 45.

    Saravi B Zink A Tabukashvili E Guzel HE Ulkumen S Couillard-Despres S et al . Integrating radiomics with clinical data for enhanced prediction of vertebral fracture risk. Front Bioeng Biotechnol. (2024) 12:1485364. doi: 10.3389/fbioe.2024.1485364,

  • 46.

    Calivà F Namiri NK Dubreuil M Pedoia V Ozhinsky E Majumdar S . Studying osteoarthritis with artificial intelligence applied to magnetic resonance imaging. Nat Rev Rheumatol. (2022) 18:11221. doi: 10.1038/s41584-021-00719-7,

  • 47.

    Vaattovaara E Panfilov E Tiulpin A Niinimäki T Niinimäki J Saarakkala S et al . Kellgren-Lawrence grading of knee osteoarthritis using deep learning: diagnostic performance with external dataset and comparison with four readers. Osteoarthritis Cartilage Open. (2025) 7:100580. doi: 10.1016/j.ocarto.2025.100580,

  • 48.

    Khan ST Huffman N Li X Sharma A Winalski CS Ricchetti ET et al . Pain assessment in osteoarthritis: present practices and future prospects including the use of biomarkers and wearable technologies, and AI-driven personalized medicine. J Orthop Res. (2025) 43:121729. doi: 10.1002/jor.26082,

  • 49.

    Kumar R Sporn K Borole A Khanna A Gowda C Paladugu P et al . Biomarker-guided imaging and AI-augmented diagnosis of degenerative joint disease. Diagnostics. (2025) 15. doi: 10.3390/diagnostics15111418,

  • 50.

    Lee DW Han HS Ro DH Lee YS . Development of the machine learning model that is highly validated and easily applicable to predict radiographic knee osteoarthritis progression. J Orthop Res. (2025) 43:12838. doi: 10.1002/jor.25982,

  • 51.

    Chen J Ma B Hu M Zhai G Sun WQ Yang S et al . Objective bi-modal assessment of knee osteoarthritis severity grades: model and mechanism (2024) 73

  • 52.

    Chaudhari AS Kogan F Pedoia V Majumdar S Gold GE Hargreaves BA . Rapid knee MRI acquisition and analysis techniques for imaging osteoarthritis. J Magnetic Resonance Imaging. (2020) 52:132139. doi: 10.1002/jmri.26991,

  • 53.

    Debs P Fayad LM . The promise and limitations of artificial intelligence in musculoskeletal imaging. Front Radiol. (2023) 3:1242902. doi: 10.3389/fradi.2023.1242902,

  • 54.

    Ocak M Çatak C . Fundamentals and applications of Micro-computed tomography In: Bioimaging modalities in bioengineering: Springer (2025). 16780.

  • 55.

    Chalian M Li X Guermazi A Obuchowski NA Carrino JA Oei EH et al . The QIBA profile for MRI-based compositional imaging of knee cartilage. Radiology. (2021) 301:42332. doi: 10.1148/radiol.2021204587,

  • 56.

    He Y Pan I Bao B Halsey K Chang M Liu H et al . Deep learning-based classification of primary bone tumors on radiographs: a preliminary study (2020) 62

  • 57.

    Lindgren Belal S Larsson M Holm J Buch-Olsen KM Sörensen J Bjartell A et al . Automated quantification of PET/CT skeletal tumor burden in prostate cancer using artificial intelligence: the PET index. Eur J Nucl Med Mol Imaging. (2023) 50:151020. doi: 10.1007/s00259-023-06108-4,

  • 58.

    Song L Li C Tan L Wang M Chen X Ye Q et al . A deep learning model to enhance the classification of primary bone tumors based on incomplete multimodal images in X-ray, CT, and MRI. Cancer Imaging. (2024) 24:135. doi: 10.1186/s40644-024-00784-7,

  • 59.

    Zhou Y Yang K Cui W . Global research trends in bone tumors: a bibliometric analysis of medical imaging. Front Oncol. (2025) 15:1579339. doi: 10.3389/fonc.2025.1579339,

  • 60.

    Gao Z Dai Z Ouyang Z Li D Tang S Li P et al . Radiomics analysis in differentiating osteosarcoma and chondrosarcoma based on T2-weighted imaging and contrast-enhanced T1-weighted imaging. Sci Rep. (2024) 14:26594. doi: 10.1038/s41598-024-78245-1,

  • 61.

    Li G Feng P Lin Y Liang P . Integrating radiomics, artificial intelligence, and molecular signatures in bone and soft tissue tumors: advances in diagnosis and prognostication. Front Oncol. (2025) 15:1613133. doi: 10.3389/fonc.2025.1613133,

  • 62.

    Safarian A Mirshahvalad SA Farbod A Jung T Nasrollahi H Schweighofer-Zwink G et al . Artificial intelligence for tumor [(18)F]FDG PET imaging: advancements and future trends - part II. Semin Nucl Med. (2025) 55:85668. doi: 10.1053/j.semnuclmed.2025.06.012,

  • 63.

    Lyell D Coiera E Chen J Shah P Magrabi F . How machine learning is embedded to support clinician decision making: an analysis of FDA-approved medical devices. BMJ Health Care Informatics. (2021) 28:e100301. doi: 10.1136/bmjhci-2020-100301,

  • 64.

    Liu J Wang H Shan X Zhang L Cui S Shi Z et al . Hybrid transformer convolutional neural network-based radiomics models for osteoporosis screening in routine CT. BMC Med Imaging. (2024) 24:62. doi: 10.1186/s12880-024-01240-5,

  • 65.

    Xue Z Huo J Sun X Sun X Ai ST Zhang L et al . Using radiomic features of lumbar spine CT images to differentiate osteoporosis from normal bone density. BMC Musculoskelet Disord. (2022) 23:336

  • 66.

    Wang S Tong X Fan Y Hu M Cui J Li J et al . Combining deep learning and Radiomics for automated, objective, comprehensive bone mineral density assessment from low-dose chest computed tomography. Acad Radiol. (2024) 31:11808. doi: 10.1016/j.acra.2023.08.030,

  • 67.

    Ogbonna C Onuiri E . Predictive diagnostic model for early osteoporosis detection using deep learning and multimodal imaging data: a systematic review and meta-analysis. Asian J Eng Applied Technol. (2024) 13:2835. doi: 10.70112/ajeat-2024.13.2.4249

  • 68.

    Hong N Whittier DE Glüer CC Leslie WD . The potential role for artificial intelligence in fracture risk prediction. Lancet Diabetes Endocrinol. (2024) 12:596600. doi: 10.1016/S2213-8587(24)00153-0,

  • 69.

    Kong SH Lee JW Bae BU Sung JK Jung KH Kim JH et al . Development of a spine X-ray-based fracture prediction model using a deep learning algorithm. Endocrinol Metabol. (2022) 37:67483. doi: 10.3803/EnM.2022.1461,

  • 70.

    Gatineau G Shevroja E Vendrami C Gonzalez-Rodriguez E Leslie WD Lamy O et al . Development and reporting of artificial intelligence in osteoporosis management. J Bone Miner Res. (2024) 39:155373. doi: 10.1093/jbmr/zjae131,

  • 71.

    Zhou L Nguyen T Choi S Yoon J . U-net-based deep learning hybrid model: research and evaluation for precise prediction of spinal bone density on abdominal radiographs. Bioengineering. (2025) 12. doi: 10.3390/bioengineering12040385,

  • 72.

    Shen H Liu H . Progress in clinical application of computer-assisted orthopedic surgery. Digital Medicine. (2023) 9

  • 73.

    Fan X Zhu Q Tu P Joskowicz L Chen X . A review of advances in image-guided orthopedic surgery. Phys Med Biol. (2023) 68:02TR01. doi: 10.1088/1361-6560/acaae9,

  • 74.

    Zhang B Li W Li M Ding X Huo J Wu T et al . The role of 3-dimensional preoperative planning for primary total hip arthroplasty based on artificial intelligence technology to different surgeons: a retrospective cohort study. Medicine. (2023) 102:e34113. doi: 10.1097/MD.0000000000034113,

  • 75.

    Lu S Li S Zhang L Qi X Yang Y Li B et al . Computer-assisted automatic preoperative path planning method for pelvic fracture reduction surgery based on enlarged RRT* algorithm. IEEE Trans Autom Sci Eng. (2025) 22:1196207. doi: 10.1109/TASE.2024.3361990

  • 76.

    Wang J Zhao Z Liang H Zhang R Liu X Zhang J et al . Artificial intelligence assisted preoperative planning and 3D-printing guiding frame for percutaneous screw reconstruction in periacetabular metastatic cancer patients. Front Bioeng Biotechnol. (2024) 12:1404937. doi: 10.3389/fbioe.2024.1404937,

  • 77.

    Zhang HQ Wang CC Zhang RJ Zhou LP Jia CY Ge P et al . Predictors of accurate intrapedicular screw placement in single-level lumbar (L4-5) fusion: robot-assisted pedicle screw, traditional pedicle screw, and cortical bone trajectory screw insertion. BMC Surg. (2022) 22:284. doi: 10.1186/s12893-022-01733-6,

  • 78.

    Geda MW Tang YM Lee CKM . Applications of artificial intelligence in orthopaedic surgery: a systematic review and meta-analysis. Eng Appl Artif Intell. (2024) 133:108326. doi: 10.1016/j.engappai.2024.108326

  • 79.

    Canton SP Austin CN Steuer F Dadi S Sharma N Kass NM et al . Feasibility and usability of augmented reality Technology in the Orthopaedic Operating Room. Curr Rev Musculoskelet Med. (2024) 17:11728. doi: 10.1007/s12178-024-09888-w,

  • 80.

    Furman AA Hsu WK . Augmented reality (AR) in orthopedics: current applications and future directions. Curr Rev Musculoskelet Med. (2021) 14:397405. doi: 10.1007/s12178-021-09728-1,

  • 81.

    Chowdhury AAA Sultana A Rafi AH Tariq, MJREdDC . AI-driven predictive analytics in orthopedic surgery outcomes (2024) 19:10424.

  • 82.

    Zhang J Zhao Y Chen Y Li H Xing F Liu C et al . A comprehensive predictive model for postoperative joint function in robot-assisted total hip arthroplasty patients: combining radiomics and clinical indicators. J Robot Surg. (2024) 18:347. doi: 10.1007/s11701-024-02102-6,

  • 83.

    Andriollo L Picchi A Iademarco G Fidanza A Perticarini L Rossi SMP et al . The role of artificial intelligence and emerging Technologies in Advancing Total hip Arthroplasty. J Personalized Medicine. (2025) 15. doi: 10.3390/jpm15010021,

  • 84.

    Lee SY Park SJ Gim JA Kang YJ Choi SH Seo SH et al . Correlation between Harris hip score and gait analysis through artificial intelligence pose estimation in patients after total hip arthroplasty. Asian J Surg. (2023) 46:543843. doi: 10.1016/j.asjsur.2023.05.107,

  • 85.

    Wang Y Guo J Tang H Li X Guo S Tian Q . Quantification of soft tissue artifacts using CT registration and subject-specific multibody modeling. J Biomech. (2024) 162:111893. doi: 10.1016/j.jbiomech.2023.111893,

  • 86.

    Matsuo H Kamada M Imamura A Shimizu M Inagaki M Tsuji Y et al . Machine learning-based prediction of relapse in rheumatoid arthritis patients using data on ultrasound examination and blood test. Sci Rep. (2022) 12:7224. doi: 10.1038/s41598-022-11361-y,

  • 87.

    Tao W Concepcion AN Vianen M Marijnissen ACA Lafeber F Radstake T et al . Multiomics and machine learning accurately predict clinical response to Adalimumab and Etanercept therapy in patients with rheumatoid arthritis. Arthritis Rheumatol. (2021) 73:21222. doi: 10.1002/art.41516,

  • 88.

    Zhang L Gao Q Dou Y Cheng T Xia Y Li H et al . Evaluation of the neoadjuvant chemotherapy response in osteosarcoma using the MRI DWI-based machine learning radiomics nomogram. Front Oncol. (2024) 14:1345576. doi: 10.3389/fonc.2024.1345576,

  • 89.

    Imam MA Abdelrahman A Zumla A Ahmed R Satta G Zumla A . Intersection of artificial intelligence, microbes, and bone and joint infections: a new frontier for improving management outcomes. Lancet Microbe. (2025) 6:101008. doi: 10.1016/j.lanmic.2024.101008,

  • 90.

    Guezou-Philippe A Clavé A Maguet E Maintier L Garraud C Fouefack JR et al . Fully automated workflow for designing patient-specific orthopaedic implants: application to total knee arthroplasty. PLoS One. (2025) 20:e0325587. doi: 10.1371/journal.pone.0325587,

  • 91.

    Han F Huang X Wang X Chen YF Lu C Li S et al . Artificial intelligence in orthopedic surgery: current applications, challenges, and future directions. MedComm. (2025) 6:e70260. doi: 10.1002/mco2.70260,

  • 92.

    Kelmers E Szuba A King SW Palan J Freear S Pandit HG et al . Smart knee implants: an overview of current technologies and future possibilities. Indian J Orthop. (2023) 57:63542. doi: 10.1007/s43465-022-00810-5,

  • 93.

    Yang Y Zhang H Gichoya JW Katabi D Ghassemi M . The limits of fair medical imaging AI in real-world generalization. Nat Med. (2024) 30:283848. doi: 10.1038/s41591-024-03113-4,

  • 94.

    Yu AC Mohajer B Eng J . External validation of deep learning algorithms for radiologic diagnosis: a systematic review. Radiology Artificial intelligence. (2022) 4:e210064. doi: 10.1148/ryai.210064,

  • 95.

    Olczak J Prijs J F IJ Wallin F Akbarian E Doornberg J et al . External validation of an artificial intelligence multi-label deep learning model capable of ankle fracture classification. BMC Musculoskelet Disord. (2024) 25:788

  • 96.

    Chen H Gomez C Huang CM Unberath M . Explainable medical imaging AI needs human-centered design: guidelines and evidence from a systematic review. NPJ Digit Med. (2022) 5:156. doi: 10.1038/s41746-022-00699-2,

  • 97.

    Guermazi A Omoumi P Tordjman M Fritz J Kijowski R Regnard NE et al . How AI may transform musculoskeletal imaging. Radiology. (2024) 310:e230764. doi: 10.1148/radiol.230764,

  • 98.

    Lekadir K Frangi AF Porras AR Glocker B Cintas C Langlotz CP et al . FUTURE-AI: international consensus guideline for trustworthy and deployable artificial intelligence in healthcare. BMJ. (2025) 388:e081554. doi: 10.1136/bmj-2024-081554,

  • 99.

    Tejani AS Cook TS Hussain M Sippel Schmidt T O'Donnell KP . Integrating and adopting AI in the radiology workflow: a primer for standards and integrating the healthcare enterprise (IHE) profiles. Radiology. (2024) 311:e232653. doi: 10.1148/radiol.232653,

  • 100.

    Wenderott K Krups J Luetkens JA Weigl M . Radiologists' perspectives on the workflow integration of an artificial intelligence-based computer-aided detection system: a qualitative study. Appl Ergon. (2024) 117:104243. doi: 10.1016/j.apergo.2024.104243,

  • 101.

    van Kolfschooten H van Oirschot J . The EU artificial intelligence act (2024): implications for healthcare. Health Policy. (2024) 149:105152. doi: 10.1016/j.healthpol.2024.105152,

  • 102.

    Kondylakis H Catalan R Alabart SM Barelle C Bizopoulos P Bobowicz M et al . Documenting the de-identification process of clinical and imaging data for AI for health imaging projects. Insights Imaging. (2024) 15:130. doi: 10.1186/s13244-024-01711-x,

  • 103.

    Rempe M Heine L Seibold C Hörst F Kleesiek J . De-identification of medical imaging data: a comprehensive tool for ensuring patient privacy. Eur Radiol. (2025) 35:780918. doi: 10.1007/s00330-025-11695-x,

Summary

Keywords

multimodal imaging, artificial intelligence, deep learning, orthopaedic diagnosis and treatment, clinical translation

Citation

Luo G, Tan S, Luo L and Hu K (2026) Artificial intelligence and multimodal imaging in orthopaedics: from technological advances to clinical translation. Front. Med. 12:1728248. doi: 10.3389/fmed.2025.1728248

Received

19 October 2025

Revised

10 December 2025

Accepted

15 December 2025

Published

09 January 2026

Volume

12 - 2025

Edited by

Edoardo Bori, ECAM Brussels Engineering School, Belgium

Reviewed by

Nafiseh Jirofti, Mashhad University of Medical Sciences, Iran

Mohammed Woyeso Geda, The Hong Kong Polytechnic University, Hong Kong SAR, China

Updates

Copyright

*Correspondence: Konghe Hu, ; Lincong Luo

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics