Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Virtual Real., 05 January 2026

Sec. Virtual Reality in Medicine

Volume 6 - 2025 | https://doi.org/10.3389/frvir.2025.1666614

This article is part of the Research TopicEnabling the Medical Extended Reality ecosystem - Advancements in Technology, Applications and Regulatory ScienceView all 16 articles

From screen to space: evaluating Siemens’ Cinematic Reality application for medical imaging on the Apple Vision Pro

Gijs Luijten,,
Gijs Luijten1,2,3*Lisle Faray de PaivaLisle Faray de Paiva1Sebastian KruegerSebastian Krueger4Alexander BrostAlexander Brost4Laura MazilescuLaura Mazilescu5Ana Sofia SantosAna Sofia Santos1Peter HoyerPeter Hoyer6Jens Kleesiek,,,,,Jens Kleesiek1,7,8,9,10,11Sophia Marie-Therese SchmitzSophia Marie-Therese Schmitz5Ulf Peter Neumann,Ulf Peter Neumann5,12Jan Egger,,,,
Jan Egger1,2,3,8,11*
  • 1Institute for Artificial Intelligence in Medicine (IKIM), Essen University Hospital (AöR), University of Duisburg-Essen, Essen, Germany
  • 2Center for Virtual and Extended Reality in Medicine (ZvRM), University Hospital Essen (AöR), Essen, Germany
  • 3Institute of Computer Graphics and Vision (ICG), Graz University of Technology, Graz, Austria
  • 4Siemens Healthineers, Forchheim, Germany
  • 5Department of General-, Visceral- and Transplant Surgery, Medical Center University Duisburg-Essen, Essen, Germany
  • 6Pediatric Clinic II, University Children’s Hospital Essen, University Duisburg-Essen, Essen, Germany
  • 7Medical Faculty, University of Duisburg-Essen, Essen, Germany
  • 8Cancer Research Center Cologne Essen (CCCE), West German Cancer Center, University Hospital Essen (AöR), Essen, Germany
  • 9German Cancer Consortium (DKTK), Partner site University Hospital Essen (AöR), Essen, Germany
  • 10Department of Physics, Technical University Dortmund, Dortmund, Germany
  • 11Faculty of Computer Science, University of Duisburg-Essen, Essen, Germany
  • 12Department of Surgery, Maastricht University Medical Centre+, Maastricht, Netherlands

Introduction: As one of the first research teams with full access to Siemens’ Cinematic Reality, we evaluated its usability and clinical potential for cinematic volume rendering on the Apple Vision Pro.

Methods: We visualized venous-phase liver computed tomography and magnetic resonance cholangiopancreatography scans from the CHAOS and MRCP_DLRecon public datasets, respectively. Fourteen medical experts assessed usability and anticipated clinical integration potential using standardized questionnaires (System Usability Scale and ISONORM 9242-110-S) and an open-ended survey. Our primary aim was not to validate direct clinical outcomes, but to evaluate the usability of Siemens’ Cinematic Reality on the Apple Vision Pro and gather expert feedback on potential use cases and missing features required for clinical adoption beyond educational purposes.

Results: Their feedback identified feasibility, key usability strengths, and required features to catalyze the adaptation in real-world clinical workflows.

Conclusion: The findings provide insights into the potential of immersive cinematic rendering in medical imaging and the needed features for clinical adoption as suggested by the medical experts. Siemens Cinematic Reality running on the Apple Vision Pro was deemed to have good usability, making it a promising tool.

1 Introduction

Head-mounted displays (HMDs) enable the viewing of three-dimensional (3D) medical images in a 3D environment, rather than on traditional monitors. The first HMD was developed in 1968 by Sutherland (1968). Jaron Lanier later popularized the concept, introducing the promise of virtual reality (Lanier, 1992). However, widespread adoption began only after Palmer Lucky introduced consumer-friendly HMDs in 2012, sparking rapid development (Harley, 2020). The Apple Vision Pro (Apple Inc., Cupertino, California, United States) is currently the newest standalone high-end extended reality (XR) device available. Compared to earlier systems such as the Microsoft HoloLens 1 and 2 (Microsoft Corp., Redmond, WA) or the Meta Quest 3 (Meta Platforms Inc., Menlo Park, CA), the Apple Vision Pro (AVP) offers superior display quality, computational performance, sensors, and more precise hand-eye interactions, potentially overcoming past issues with image clarity and usability (Egger et al., 2024).

In parallel with these hardware developments, there has long been interest in instantly visualizing 2D medical image slices, such as those from computed tomography (CT) or magnetic resonance imaging (MRI), as accurate 3D representations. Ideally, this process would require no manual segmentation efforts (Zhang et al., 2011). Volume rendering, a technique introduced in the late 1980s, enabled the display of 3D representations of CT and MRI data on 2D monitors (Fuchs et al., 1989). HMDs now enable three-dimensional volume rendering (3DVR) of medical scans in immersive environments (Douglas et al., 2017). Ongoing advancements continue expanding its medical and educational applications (Ljung et al., 2016; Zhang et al., 2011; Eid et al., 2017; Dappa et al., 2016), with research indicating that 3DVR improves radiology diagnostics, surgical decision-making, and anatomical understanding by providing superior spatial perception compared to slice-based imaging methods (Duran et al., 2019; Queisner and Eisenträger, 2024).

While standard volume rendering provides crucial 3D spatial perception, photorealistic rendering techniques aim to further enhance anatomical realism and depth cues. Potentially leading to even greater clinical utility, particularly in education or complex surgical planning (Eid et al., 2017; Li et al., 2019; Dappa et al., 2016).

Siemens’ Cinematic Reality (Siemens Healthineers, Medical Imaging Technologies, Erlangen, Germany) (Healthineers, 2017) applies these cinematic rendering techniques to enhance traditional VR, offering photorealistic 3D depictions using global illumination models that simulate natural light, shadows, refraction, and occlusion (Kroes et al., 2012; Ropinski et al., 2010; Csebfalvi and Szirmay-Kalos, 2003; Fellner, 2016). Siemens Healthineers initially developed this Cinematic Reality (CR) application for the computer and later the Microsoft HoloLens (Fellner, 2016; Healthineers, 2024).

To contextualise these advances, it is helpful to outline the limitations of the conventional rendering methods that CR seeks to improve upon. Standard 3DVR on CT provides crucial 3D spatial perception, but typically relies on user-defined transfer functions and local lighting models. In this conventional workflow, a transfer function maps ranges of Hounsfield units to colour and opacity; when set aggressively, this can introduce harsh opacity transitions and stripe-like artefacts, and depth cues remain limited because only local shading around each voxel is considered. Isosurface rendering, commonly used on clinical workstations, further simplifies the volume into a hard surface defined by a single intensity threshold, often yielding unnaturally sharp, “plastic” boundaries between tissues. Photorealistic, global-illumination–based cinematic rendering techniques aim to overcome these limitations by simulating how light is transported, scattered, and absorbed in the volume, thereby providing smoother shading, soft occlusion, and more intuitive depth perception, which can translate into greater clinical utility, particularly for education and complex surgical planning (Eid et al., 2017; Li et al., 2019; Dappa et al., 2016; Fukumoto et al., 2022). These differences across rendering approaches are illustrated in Figure 1, which shows the same CHAOS Subject 5 CT volume rendered with Siemens’ CR, with standard 3DVR using 3D Slicer (version 5.10.0; open-source medical imaging platform - https://www.slicer.org/), with Slicer’s maximum-intensity-projection–style 3D view, and with an isosurface generated using Slicer’s surface tools.

Figure 1
CT scan images show detailed cross-sections of the human chest cavity in various shades, highlighting rib structures and internal organs. Panels A1 and A2 display reddish hues, B1 and B2 feature a brown color, C shows a lighter beige, and D uses a greenish tint. Each panel provides different perspectives and color contrasts to emphasize anatomical details.

Figure 1. Comparison of rendering techniques for a contrast-enhanced abdominal CT scan (CHAOS dataset, Subject 5). Panels A1–A2 show Siemens’ Cinematic Reality renderings that illustrate photorealistic shading and soft-tissue detail. Panels B1–B2 depict standard three-dimensional volume rendering using 3D Slicer, with local lighting and the standard CT-Liver transfer function included in the default settings. Panel C shows a maximum-intensity-projection (MIP) view generated in 3D Slicer’s volume rendering module, and panel D shows an isosurface model created with the Surface tool in 3D Slicer. All views are derived from the same CT volume and demonstrate how CR provides smoother shading and more intuitive depth cues compared with conventional 3DVR, MIP, and isosurface rendering.

Siemens’ CR applies such cinematic rendering techniques to medical CT and MRI data. In contrast to standard 3DVR and isosurface rendering, CR uses a physically based global illumination model that generates soft shadows, ambient occlusion, and view-dependent highlights Kroes et al. (2012); Ropinski et al. (2010); Csebfalvi and Szirmay-Kalos (2003); Fellner (2016). High–dynamic-range lighting combined with tone mapping compresses the simulated radiance into the displayable range while preserving local contrast, and tissue-aware transfer function presets map relevant intensity ranges (e.g., bone, vessels, or parenchyma) to realistic colours and translucency. These tissue-specific presets reduce manual tuning while emphasising clinically relevant anatomy in a visually consistent way. Together, these differences in illumination, tone mapping, and tissue-aware transfer functions explain why CR images are often perceived as more realistic than conventional 3DVR and isosurface rendering and motivate their evaluation on emerging XR platforms such as the Apple Vision Pro.

Prior studies have demonstrated CR’s potential, such as its application on the Microsoft HoloLens for pediatric heart surgery planning (Gehrsitz et al., 2021). Despite these advancements, the clinical benefits of photorealistic rendering techniques such as Siemens’ CR, particularly in routine surgical decision-making, remain unclear (Duran et al., 2019; Gehrsitz et al., 2021).

Since the clinical benefits remain unclear, assessing the usability of such systems (HMDs with cinematic rendering applications) and identifying areas of clinical potential is necessary. It’s especially interesting to look beyond anatomy education. Surgical planning and intervention are an interesting example where minor improvements can still warrant the overall cost. Liver transplantation exemplifies such a domain where precise 3D visualization is paramount due to its intricate and highly variable vascular and biliary anatomy (Kelly et al., 2017; Yeo et al., 2018; Erbay et al., 2003). Furthermore, surgical planning in this context often requires not only identifying anatomical variants but also critically assessing tumour-induced displacement or encasement of vital structures (vessels, bile ducts, parenchyma), tasks where immersive, high-fidelity photorealistic rendering could offer clinical value.

These previously mentioned uncertainties surrounding clinical effectiveness, combined with the recent advancements in XR devices and software, including the introduction of AVP with CR, provide a compelling opportunity to rigorously evaluate the usability and practical integration potential of CR into clinical workflows (Kukla et al., 2023; Lopes et al., 2018; Healthineers, 2024; Egger et al., 2024).

This study evaluates Siemens’ CR on the AVP, focusing on its interactive usability and its anticipated integration into clinical workflows, such as surgical planning, as assessed through expert clinician feedback. Siemens’ CR focuses on photorealistic visualization and standard interactions (Figures 2, 3), but does not yet include workflow-specific features such as measurements, segmentations, or annotations. Therefore, our study aimed not only to evaluate usability but, also to capture expert clinician feedback on which features are essential for future clinical adoption. The goal of this study was therefore twofold: to systematically evaluate the usability of Siemens’ CR on the AVP using validated questionnaires, and to capture clinician feedback identifying essential features for integration into preoperative and intraoperative planning workflows.

Figure 2
Panel A shows detailed anatomical images of a human skull and internal structures in profile and front views. Panel B displays close-ups of the gallbladder, bile ducts and colon. Panel C presents a cross-section of internal organs, highlighting vascular details.

Figure 2. Cinematic 3DVR examples: Two different transfer functions and perspectives for (A) Siemens’ demo case and (B) the MRCP_DLRecon dataset case (Kim et al., 2024), and (C) a single example of subject four from the CHAOS dataset (Kavur et al., 2019).

Figure 3
Radiologic visualization techniques are demonstrated using various CT scan images and 3D models. Buttons for actions like zoom, rotate, move, windowing, scroll, switch plane orientation, and entering/exiting 3D view are highlighted. The image shows overlays explaining these interactive features with close-up views of anatomy and interface elements indicating actions like view direction and light map adjustments.

Figure 3. Core functionalities of Siemens' Cinematic Reality on the Apple Vision Pro. Eye gaze functions as a pointer, and finger pinching acts as a selection tool (click). Top left: Access the library to load scenes (i.e., DICOM files converted to the appropriate scene files). Rendering happens on the AVP, depending on the operator's viewpoint. Top right: 3D model interactions include scrolling/clipping (top left), resizing (top right), windowing (bottom left), and rotation (bottom right). Presets enable tissue-specific transfer functions (e.g., orange rendering in the lower-right model); clipping planes/boxes and light map settings adjust visibility and lighting effects. View orientation can be switched via the 2D/3D overview (axial, sagittal, coronal). Bottom: From left to right, 2D slice interactions include zooming (two hands), windowing, scrolling, and switching between radiologic visualization modes. Rotation supports one or two hands; all other interactions use a single hand.

Direct deployment in clinical workflows was beyond the study’s scope; the evaluation with structured usability questionnaires and domain-specific expertise provides actionable insights into the system’s readiness, features, limitations, and opportunities for clinical adoption prior to clinical trials. While participants reflected on the subjective visual quality of the photorealistic rendering, these impressions were treated as part of the overall user experience rather than an isolated outcome.

To go beyond the standard demo case included in the CR application, public datasets related to abdominal visceral surgery are used. This resembles the realistic world scenario more closely than using an optimized demo case. Both CT and MRI, specifically magnetic resonance cholangiopancreatography (MRCP), scans from the CHAOS (Kavur et al., 2021; Kavur et al., 2019) and MRCP_DLRecon (Kim et al., 2025; Kim et al., 2024) datasets are used. Fourteen medical experts evaluated the system via the System Usability Scale (SUS) (Brooke, 1996; Brooke, 1986; Bangor et al., 2009; Lewis, 2018), ISONORM 9242–110-S (Prümper, 1993; Prümper, 2006; Bevan et al., 2015), and qualitative questionnaires.

This study aims to contribute to a broader understanding of how immersive 3D visualization, specifically through the CR application on the AVP, can enhance medical imaging interpretation. Moreover, the survey intends to provide a perspective on how these technologies can support clinicians, specifically liver surgeons, by identifying usability strengths, limitations, and desired features to catalyze clinical adoption.

2 Materials and methods

2.1 Study design and participants

A mixed-methods usability evaluation was conducted, incorporating both quantitative and qualitative assessments. The quantitative analysis used the System Usability Scale (SUS) score (Brooke, 1986) and ISONORM 9242-110-S (Prümper, 2006), which were chosen because they are validated and used as standards within the field. Qualitative feedback was collected through an open-ended survey that also included demographic information. The study included a cohort of 11 surgeons specializing in general, visceral, vascular, liver and transplant surgery, one doctor assistant, and two medical students.

Preoperative hepatobiliary planning is variable but usually includes initial consultation with review or request of contrast enhanced CT and/or MRI (MRCP), radiology demonstration of findings, and repeated surgeon review of 2D slices to assess vascular and biliary anatomy. Surgeons often formulate a plan with alternatives, which is then confirmed or adapted intraoperatively, most commonly with ultrasound and reinspection of the 2D slices. 3DVR and CR on HMDs have the potential to reduce repeated 2D review and intraoperative plan deviations. Our evaluation therefore focused on usability of CR’s current interaction set (Figure 3) and on eliciting expert-identified features needed for integration into this workflow (Tables 1 and 2).

Table 1
www.frontiersin.org

Table 1. ISONORM 9241-110-S results across the seven usability principles. Values are reported as mean (μ), standard deviation (σ), median (x̃), interquartile range (IQR), and quartiles (Q1, Q3).

Table 2
www.frontiersin.org

Table 2. Positive and negative aspects regarding Siemens’ Cinematic Reality application running on the Apple Vision Pro.

The entire study was conducted at the University Hospital Essen (AöR), Essen, Germany. The Ethics Committee waived the Institutional Review Board (IRB) approval requirement because the study utilized publicly available, anonymized retrospective datasets and posed no risk or harm to the participating professionals and scanned subjects.

2.2 Datasets

The visualized 3D depictions of the AVP using the CR application were based on the following medical datasets:

1. CHAOS dataset (Kavur et al., 2019; Kavur et al., 2021): CT volumes (512×512×number of slices, 2 mm slice thickness) acquired during the portal venous phase after contrast agent injection, enhancing the portal veins and liver parenchyma.

2. MRCP_DLRecon dataset (Kim et al., 2024; Kim et al., 2025): An MRCP scan (480×384×58, 1.2 mm slice thickness, 3T–3D T2w TSE sequence), selected for its detailed depiction of the biliary and pancreatic ducts, particularly relevant for hepatobiliary surgery.

From the CHAOS dataset (Kavur et al., 2021; Kavur et al., 2019), subjects four and five were selected to ensure a representative scan of both a male and a female, and they had 94 and 95 slices, respectively. The MRCP_DLRecon dataset (Kim et al., 2025; Kim et al., 2024) consists of a single volunteer scan with an unidentified gender. To prepare volumetric data for Siemens’ CR on the AVP, the so-called scenes files were generated from DICOM using Siemens’ Cinematic Playground, which requires input in DICOM series format. This process involved only file conversion and scene setup; actual rendering and interaction occur on the AVP based on the user’s viewpoint. The CHAOS dataset was already in DICOM format, while the MRCP_DLRecon dataset, originally stored in HDF5 (.h5) format, was converted using a Python script (Python Software Foundation, Wilmington, DE, United States), utilizing the h5py, numpy, and pydicom libraries. Scenes were loaded onto the AVP via iCloud, with setup times under 1 min. No manual adjustments or transfer function customization were performed; the scans were imported as-is without cropping or editing. Total time per scene preparation was thus under 5 min.

2.3 Data preparation for cinematic reality

Siemens’ CR (v2.0 build 5) for the AVP (VisionOS v2.3.1) was installed via TestFlight (v3.7.1). Unlike the freely available demo version, this version supports the loading of custom-made scenes. The scenes were prepared using Siemens’ Cinematic Playground (v0.19) on a Windows laptop (Intel Core i7-11800H, NVIDIA GeForce RTX 3050) with a required Digital Imaging and Communications in Medicine (DICOM) series. The CHAOS dataset (Kavur et al., 2019) provided the DICOM series directly, while the MRCP_DLRecon dataset (Kim et al., 2024), stored in Hierarchical Data Format 5 (HDF5), was converted to a DICOM series.

The voxel values were normalized and linearly scaled [-1000 to 2000] using Python (version 12) and the following libraries: numpy (v2.2.2), h5py (v3.12.1), nibabel (v5.3.2), pydicom (v3.0.1), and scikit-image (v0.25.1). This scaling adjustment optimized compatibility with Siemens Cinematic Playground and CR’s transfer function. Scene files were then loaded onto the AVP via iCloud.

Siemens’ CR facilitates conventional interactions, including windowing, scrolling, and zooming. The system is also capable of 3D cinematic volume rendering, as illustrated in Figure 3.

2.4 Evaluation

Quantitative data from the SUS (Brooke, 1986) and ISONORM (Prümper, 2006) questionnaires were analyzed using descriptive statistics, including mean scores and standard deviations. Qualitative responses were thematically categorized based on strengths, weaknesses, current clinical applicability, feature requests, and potential future applications. The ISONORM 9241-110-S questionnaire assesses usability across seven principles derived from ISO 9241-110:

1. Suitability for the task: How well the system supports users in completing tasks.

2. Self-descriptiveness: The intuitiveness of system functionality.

3. Conformity with user expectations: Consistency with known interface standards.

4. Learnability: Ease with which new users can become proficient.

5. Controllability: The extent to which users can influence actions and outcomes.

6. Error tolerance: The system’s ability to prevent or recover from user errors.

7. Customizability: How well the system can be adapted to user needs.

These principles provide a comprehensive framework for evaluating the user-system interaction, focusing on effectiveness, efficiency, and satisfaction in a specified context of use (ISO, 2020).

The System Usability Scale (SUS) is a widely used tool for evaluating the usability of various systems, including medical software. It consists of a 10-item questionnaire with five response options for respondents, ranging from strongly agree to strongly disagree. SUS provides a single score ranging from 0 to 100, representing a composite measure of the overall usability of the system being studied (Brooke, 1986). Notably, scores above 68 are considered above average across industries and for digital health applications, with higher scores indicating better usability (Hyzy et al., 2022; Brooke, 2013). Scores between 74 and 85 indicate good to excellent usability (Brooke, 2013).

2.5 Procedure and data collection

For this study, we used a research build of Siemens’ CR for the Apple Vision Pro that supports loading CTs and MRIs beyond the demo case as locally prepared scene files; this build is currently shared only with external groups via research collaborations and is planned for broader release. All sessions took place on site at our institution and were overseen solely by the local research team; no Siemens Healthineers personnel were present during data collection.

Informed Consent: All participants provided written informed consent prior to study participation.

Device Setup: Streaming from the AVP to an iPad was activated via AirPlay to facilitate observation and assistance. Participants then put on the AVP headset and completed the built-in hand–eye calibration before using the Siemens’ CR application. The researcher guided participants in this process.

Application Interaction (Task): Participants launched the Siemens CR application, which was preloaded with CR scenes from the CHAOS and MRCP_DLRecon datasets. The datasets were presented in a fixed order on the AVP: participants first explored the CT volume from the CHAOS dataset, followed by the MRCP scan. Participants thus started with a CT volume from the CHAOS dataset. From this point onward, participants were not guided; help was provided only upon request. The datasets were presented in a fixed order: participants first explored the CT volume from the CHAOS dataset, followed by the MRCP scan. This sequence enabled them to examine both vascular and biliary anatomy and assess the utility of CR in distinct clinical contexts. After viewing both datasets, participants were free to revisit scenes or explore an additional demo head CT case provided by Siemens Healthineers. Participants were tasked with the following instruction: explore key anatomical structures and evaluate the system. Specifically, they were asked to (i) explore anatomical structures relevant to hepatobiliary surgery, including portal veins, bile ducts, and liver parenchyma; (ii) examine image fidelity and photorealism using cinematic volume rendering; (iii) navigate the interface (Figure 3) using standard interaction techniques such as zooming, panning, scrolling, windowing, and rotating the volume and CT slices, as well as adapting the transfer function and switching between scenes (Siemens’ demo scene, CHAOS, MRCP_DLRecon); (iv) verbally articulate their observations, usability impressions, and any encountered challenges as part of a think-aloud protocol; and (v) describe how they would ideally integrate the system into their clinical workflow (e.g., for surgical planning or patient case discussions), including which functionalities they would use and how. These interactions and comments were observed by the researcher in real-time via AirPlay streaming to an iPad. If assistance was required, the researcher provided guidance, utilizing the AirPlay stream to facilitate real-time support.

While no rigid task list or time limit was imposed, the average interaction time during the application interaction was approximately 15–20 min, while hand–eye calibration took less than 5 min. Researchers provided minimal assistance, intervening only when participants requested support.

Data Collection: Following the interaction session, participants completed the System Usability Scale (SUS), the ISONORM 9241-110-S questionnaire, and an open-ended survey capturing qualitative feedback. All responses were collected via Google Forms.

3 Results

3.1 Participant demographics

Fourteen subjects participated in the study, including two medical students, a doctor/surgeon’s assistant, and 11 surgeons, with a gender distribution of eight females and six males. The ages of the female subjects ranged from 22 to 43 years, with a mean of μ=32.0 years, standard deviation σ=6.70, and a median of 32 years (IQR=7.25;Q1=28.75,Q3=36.00). In contrast, males, ranging in age from 35 to 68 years, exhibited a mean age of μ=48.83 years, standard deviation σ=11.92, and a median of 46 years (IQR=13.00;Q1=41.75,Q3=54.75).

The AVP does not permit the use of spectacles within the HMD; rather, it necessitates the acquisition of its proprietary insert lenses. Consequently, the diopter of the subjects is an essential factor. Among the ten participants who wear spectacles, one subject was classified as farsighted (+2.0), while the remaining subjects were nearsighted μ=3.22, σ=2.91, minimum 9.25, maximum 1.0. Of the ten participants who normally wear spectacles, eight used the AVP with compatible insert lenses. The remaining two completed the study without visual correction. While both completed all tasks, they reported reduced clarity when inspecting fine anatomical structures. Although this did not prevent participation, it may have influenced their subjective usability ratings, exemplifying the need for vision correction support in clinical extended reality systems.

Notably, only two subjects had prior experience with HMDs, which was limited to one and 8 h, respectively. The HMDs were the HoloLens 2 and/or Meta Quest 3 (Meta Platforms, Inc., Menlo Park, CA). All subjects were unfamiliar with 3DVR; their familiarity with 3D rendering was limited to segmentation-based 3D rendering.

3.2 System Usability Scale

The SUS is technology-agnostic and combines effectiveness, efficiency and satisfaction with high reliability (Cronbach alpha = 0.91) in a single score (0–100) (Bangor et al., 2009; Bangor et al., 2008; Vlachogianni and Tselios, 2022; Brooke, 1986). Based on over 10 years of empirical evidence, the score can be divided into seven categories, from worst to best imaginable (Bangor et al., 2009; Vlachogianni and Tselios, 2022).

The mean score was σ = 77.68 (σ = 15.01, IQR = 27.50; Q1 = 63.75, Q3 = 91.25), indicating a score between good and excellent. Surgeons and assistant doctors rated the system higher than students. There were no noticeable differences between age or gender groups. An overview of the results for each group is shown in Figure 4, including students, residents, and surgeons.

Figure 4
Box plot showing System Usability Scores (SUS) for surgeons, students, and assistant doctors. Individual scores are plotted with shapes: circles for surgeons, squares for students, and diamonds for assistant doctors. Mean SUS lines and Interquartile Range (IQR) are colored for each group. Overall mean SUS is marked by a solid black line.

Figure 4. Visualization of SUS scores for each participant, grouped by profession. Mean SUS scores are indicated by horizontal lines, with shaded areas representing the Interquartile Range (IQR).

3.3 ISONORM 9242-110-S

ISONORM results across the seven usability principles are summarized in Table 1. Suitability, conformity, and controllability were rated highest, while self-descriptiveness and customizability showed the most room for improvement. Figures 5, 6 provide a visual representation of these results as a heatmap and a stacked bar plot, respectively.

Figure 5
Heatmap showing usability categories versus usability ratings. Categories include Conformity, Controllability, Customizability, Error Tolerance, Learnability, Self-Descriptiveness, and Suitability. Ratings range from Poor to Excellent. Numeric values are displayed in each cell, with colors representing intensity from zero (purple) to eight (yellow), indicating varying levels of usability across categories.

Figure 5. Heatmap of ISONORM responses grouped per usability measurement.

Figure 6
Bar chart depicting usability categories rated from poor to excellent. Categories include customizability, error tolerance, and more, with percentages shown for each rating. Ratings range from negative values for poor to high positive values for excellent, illustrating varying degrees of satisfaction across categories.

Figure 6. Stacked bar plot of ISONORM responses grouped per usability measurement centered around 0%.

3.4 Qualitative feedback

The open-ended questionnaire delved into the advantages and disadvantages of the CR application on the AVP (Table 2), its potential applications in a medical settings, and its current clinical usability (Table 3), as well as specific complex use cases and desired improvements for both software.

Table 3
www.frontiersin.org

Table 3. Potential medical use cases and feasibility in daily clinical routine for the Apple Vision Pro.

Surgeons identified key areas for enhancing the Siemens’ CR application on AVP to optimize its clinical utility. Recommended improvements include integration with electronic patient records, advanced measurement tools for tumor and vessel size assessment, and the ability to insert 3D models of surgical instruments. Enhanced segmentation functionality enabling seamless switching between anatomical structures (e.g., arteries, portal veins, bile ducts), while gesture-based removal of irrelevant areas should provide more focused visualizations. Additionally, multi-user and multi-platform compatibility to facilitate collaborative planning and real-time surgical discussions. Finally, incorporating a voice activation function was suggested to enhance operability in sterile settings or when both hands are occupied with other tasks.

The setup process for the AVP was typically less than 5 min and considered straightforward. This duration includes donning the headset, performing the initial eye and hand calibration, and launching the CR application. Such a setup time is likely more acceptable for preoperative planning sessions, where time constraints are less critical. However, this setup duration may be undesirable for intraoperative applications, but the participants did not see this as a problem. In practice, once the device is calibrated and the CR application is running, re-donning the AVP and waking it from standby can be performed considerably faster than the full initial setup.

In addition to qualitative comments on usability, participants also described their real-time experience with Siemens’ CR on the Apple Vision Pro. Although no performance measurements were recorded, all users who were able to wear the headset reported smooth interaction and perceived the application as running in real-time, consistent with AVP’s reported refresh rate of 90 Hertz. No lag was noted during rotation, clipping or window adjustments. Eye-gaze selection and pinch-to-click gestures remained accurate throughout each session following the brief calibration, and no tracking interruptions occurred. Most participants used the system for about 15–20 min without discomfort, except for one individual who experienced mild nausea near the end. One participant was unable to use the device due to incompatible glasses, and a few noted minor light leakage from imperfect face-sealing, though this did not affect comfort. Some participants needed a short moment to adjust to the hand-eye coordination for certain gestures, particularly wrist-based rotation, but this reflected user familiarity rather than system issues. No crashes or freezes occurred. Overall, these observations indicate that under the evaluated conditions, the AVP provided stable and comfortable performance for real-time exploration of cinematic rendering.

4 Conclusion and discussion

This study presents one of the first empirical evaluations of Siemens’ Cinematic Reality on the Apple Vision Pro for hepatobiliary CT and MR imaging. Clinicians and trainees reported good to excellent usability on standardized instruments (SUS and ISONORM) and provided convergent qualitative feedback about strengths and missing functionality. Taken together, these results indicate that the current AVP–CR implementation is usable in a controlled setting and delineate the feature gaps that must be addressed. Still, they do not yet demonstrate effectiveness or impact on clinical or educational outcomes.

4.1 Contextualizing the findings

In contrast with the current surgical planning, (immersive) 3D visualization may add value by improving spatial understanding of tumor-vessel-bile duct relationships and by enabling rapid exploration of alternative resection planes directly in 3D rather than on 2D slices before intraoperative confirmation. This may reduce intraoperative plan deviations, shorten planning time, and improve surgeon confidence. As highlighted in Tables 1, 2, our experts already emphasized these use cases. Importantly, our study also identifies which functionalities are prerequisites for unlocking this potential in clinical planning: measurements, annotations, segmentation, and multi-modal fusion. Without these, the technology remains mainly useful for education and demonstration; with them, the step toward clinical decision support becomes more feasible.

While the focus of this work is on Siemens’ CR on the AVP, there is a long-standing use of non- and semi-immersive visualization systems for 3D visualization in medical imaging, such as cave automatic virtual environments (CAVEs), multi-projector setups for holographic visualization, and stereoscopic monitors.

Cruz-Neira et al. (1993) pioneered surround-screen VR such as the CAVE which is promising for medical teaching (Cruz-Neira et al., 1993). Yet, the physical space required and the financial cost make it less accessible than current cheaper HMDs (Cruz-Neira et al., 2010), which might be a reason its adoption in, for example, anatomy education is rare (Adnan et al., 2025). An advantage of CAVEs is the direct option for multi-user interaction, which Siemens’ CR on the AVP or HoloLens 2 does not have (Table 2). Additionally, no extra (heavy) head-mounted hardware is needed, allowing users more comfort and the ability to wear their own glasses, unlike in the AVP. The same can be said for holographic projections showing promise, but likely not enough to warrant the investment for clinical purposes (Blanche, 2021; Chu et al., 2009). Not to mention that when moving from surgical planning to in-surgery usages, it requires an entire specialized operating room, therefore we postulate it’s more effective to explore HMDs until the above technology matures more and costs are lowered.

Stereoscopic monitors remain a practical alternative in many operating rooms. They’re easier to deploy than projection setups, can be made glasses-free with eye-tracking and polarization technology, and allow clinicians to maintain sterility (Kang et al., 2022). Studies suggest that stereoscopic displays can improve anatomical understanding, aid novice surgeons in spatial tasks, and enhance diagnostic performance (Held and Hui, 2011; van Beurden et al., 2012). Recent innovations in autostereoscopic displays that integrate eye tracking show promise; for instance, Kang et al. (2022) developed a glasses-free 3D autostereoscopic system using eye tracking for cardiac CT imaging, which expert readers identified key arterial structures faster and without discomfort. Still, these monitors lack the immersive, hands-free 360 spatial interaction of HMDs. While outside the scope of this study, future studies or meta-analyses should compare these approaches when evaluating cost, usability, and above all, surgical outcomes.

While semi-immersive setups deliver valuable stereoscopic perception, they do not provide the fully immersive, high-fidelity interaction achieved by contemporary head-mounted displays. The AVP represents a leap in HMD capabilities (e.g., resolution, interaction fidelity) compared to predecessors like the HoloLens 2 and Meta Quest 3 (Egger et al., 2024). When paired with CR, which enhances traditional volume rendering through photorealistic lighting, this hardware promises clinically actionable 3D visualization. However, prior CR evaluations lacked usability assessment in clinical contexts, particularly on cutting-edge platforms like the AVP.

The applications of cinematic rendering techniques go beyond direct visualization. For example, it can generate photorealistic synthetic data used to train deep learning algorithms for medical image interpretation (Mahmood et al., 2018) or be used in various clinical and educational scenarios (Eid et al., 2017; Li et al., 2019). Cinematic rendering has therefore become an enticing alternative to volume rendering (Dappa et al., 2016). More user-centered studies have applied CR for patient education (Pachowsky et al., 2022) and student learning (Binder et al., 2019; Binder et al., 2021; Elshafei et al., 2019). These studies reported improved patient understanding and student recall compared to traditional CT scans. However, these evaluations were conducted on conventional 2D monitors and did not directly assess software usability.

Combining advanced HMD hardware and cinematic rendering software has led to the development of interactive 3DVR. 3DVR allows medical scans to be displayed and interacted with in true 3D using an HMD (Douglas et al., 2017). Although 3DVR and likewise cinematic rendering technologies hold promise, particularly for complex patient-specific anatomy, the exact clinical benefit remains uncertain (Duran et al., 2019; Queisner and Eisenträger, 2024; Brookmeyer et al., 2024).

Siemens Healthineers previously explored this technology by adapting CR for the HoloLens. One study showed improved surgical decision-making compared to traditional 2D CT images or 3D printed models in cardiac surgery (Gehrsitz et al., 2021). Nevertheless, despite the HoloLens’ popularity in medical research, few clinicians used the commercial CR software (Gsaxner et al., 2023).

While open-source CR alternatives exist, they lack the refined transfer functions and realism of Siemens’ commercial solution (Baseer et al., 2023); clinicians prefer Siemens’ CR application over commercial alternatives. The performance gap likely arises from Siemens Healthineers’ dedicated team working extensively to perfect the transfer functions for each specific use case, which is a highly specialized and time-intensive task. This illustrates a trade-off between open-source and commercial systems: open-source tools allow rapid prototyping of workflow-specific features such as measurement or annotation, whereas Siemens’ CR prioritizes photorealism and refined transfer functions but is currently limited in functionality. This hurdle towards clinical integration and feature desires was consistently noted by expert participants in our study (Tables 1, 2).

Most existing studies on CR focused either on educational uses displayed on monitors or on synthetic data generation. Few studies have explored actual clinical scenarios to demonstrate tangible benefits. This leaves a significant gap in understanding usability and the specific features clinicians require for clinical adoption.

To address this gap, our study specifically examines the usability of the CR software on the AVP, gathering direct clinician feedback and ideas for clinical integration. We selected the latest hardware (AVP) and software (CR application) to ensure the best possible fidelity, realism, and user interaction–factors crucial for future clinical acceptance. This approach also provides insight into the future potential as these technologies become increasingly accessible.

4.2 Key contributions and strengths

This study is the first to evaluate the usability of Siemens’ CR application on the AVP within a clinical context. It provides detailed insights into usability (Figures 4, 6), highlights specific positive and negative aspects (Table 2), and presents potential clinical applications identified by medical experts (Table 3). Using publicly available imaging data instead of Siemens-specific datasets ensured better alignment with clinicians’ expertise and more realistic testing conditions. Additionally, the public datasets increase the reproducibility of this research.

Our results show that CR on the AVP has good-to-excellent usability, even compared with other digital health applications, with an average SUS score of 77.68 (Hyzy et al., 2022; Bevan et al., 2015; Brooke, 1996; Bangor et al., 2008). Usability ratings were positive across participants, ranging from medical students to experienced surgeons.

Medical students rated usability notably lower than medical professionals (Figure 4). While this difference may partly result from the small sample size, several factors likely contribute. The CR application uses features and icons similar to those found in standard radiological software, such as windowing, slice scrolling, zooming, plane selection, and visualization methods like maximum intensity projection (Figure 3). Surgeons, who frequently use these tools, likely felt more comfortable from the start. This interpretation is further supported by the lower ISONORM scores for self-descriptiveness (Figures 5, 6).

Medical professionals may also have built higher frustration tolerance and better learning strategies over time due to exposure to various technologies throughout their careers. Their practical experience likely helped them recognize clinical use cases more readily, possibly influencing their higher appreciation.

The ISONORM results (Figure 6) also suggest improvements are needed in learnability and error tolerance. Importantly, the study had no genuine technical failures (software or hardware malfunctions) observed during the live video stream. Rather, participants occasionally experienced unexpected interactions with the 3D anatomy, likely due to unfamiliarity with the specific hand-eye interactions, interface elements, or icons. Despite these minor issues, users rated the controllability, conformity, and suitability highly, reflecting the system’s consistent performance.

Lastly, customizability received comparatively lower scores. The qualitative feedback (Table 2) confirms that negative aspects mainly consisted of feature requests rather than fundamental problems. The positive feedback aligns well with the overall SUS and ISONORM scores. Thus, exploring requested features and incorporating them without negatively affecting usability, learnability, or self-descriptiveness will be an essential next step.

These findings directly reflect the primary aim of our study: not to demonstrate clinical outcomes, but to assess usability and collect expert feedback on which functionalities must be added for CR on the AVP to transition from promising usability toward practical workflow integration.

4.3 Limitations and challenges

Despite promising results, several limitations must be acknowledged. The current version of the CR application lacks key clinical features such as segmentation, measurement tools, annotations, and multi-modal scan fusion. These features are routinely used in surgical planning and diagnostics. Nevertheless, participating surgeons, who are often familiar with automated 3D reconstructions from surgical navigation software, rated the system’s usability highly. This indicates that the core interaction design is valuable even without these advanced features.

Hardware constraints also posed minor challenges and affect how easily the device can be shared. The AVP requires proprietary vision correction inserts rather than spectacles, which makes each unit effectively personalized. One participant withdrew because their glasses could not be accommodated; most other participants adapted comfortably using contact lenses. Despite two face-interface sizes, slight ambient light leakage was observed when the seal was not perfect, which may subtly influence perceived image quality, although no formal complaints were reported. Participants did not report problems with headset weight during the short 15–20 min sessions, even though the AVP headset itself weighs roughly 750–800 g and the external battery about 350 g, connected via a tethered cable that can restrict movement and compromise sterility if not routed carefully. Only one participant reported mild nausea. While HMD technology has improved substantially, cybersickness remains an important consideration, particularly during prolonged clinical use (Tian et al., 2022).

In practical terms, the AVP therefore behaves more like a personal device that is best suited to single-user preoperative planning or focused intraoperative use, unless streaming to an external display is used for observers. The high purchase cost is less problematic in complex, high-value surgical cases where even small gains in decision-making may be justified. Still, it is a more substantial barrier for broad deployment in education and routine teaching, where many simultaneous viewers would be needed. By contrast, the external battery and cable are more limiting in sterile intraoperative workflows than in educational or non-operative contexts, where restricted movement and cable routing are easier to accommodate.

The sample size of the study was relatively small, yet it included highly experienced surgeons, lending credibility to the findings. Although no direct comparison was made with traditional 2D displays or open-source alternatives, validated usability measures (SUS and ISONORM) allowed for usability assessments. In addition, all participants first viewed the CT dataset and then the MRI dataset on the AVP, a fixed order that may introduce learning effects; future studies will use a counterbalanced design with randomized modality order and explicit baseline comparators, for example, AVP-CR versus standard 2D workstation review in real clinical cases. Thus, while participants identified relevant potential clinical uses ranging from surgical planning to education, these remain speculative until confirmed through real-world clinical studies.

Finally, note that at present, access to the AVP build of Siemens’ CR used in this study requires research collaboration with the vendor. However, broader release is planned, so this is a temporary practical constraint on replication rather than a methodological limitation.

4.4 Integration into clinical practice

The AVP cannot easily be shared among multiple users because prescription glasses and face-fit are personalized for each device owner. This limitation presents a practical and financial hurdle, especially for educational scenarios such as patient education (Egger et al., 2024). Although multiple face-fits can be purchased or users can wear contact lenses, these solutions remain suboptimal for widespread clinical adoption. Ideally, Apple would develop a face-fit option that allows regular prescription glasses to be comfortably worn. Alternatively, the Siemens CR software or open-source alternatives could be adapted for other available or upcoming HMD models. Combining this adaptation with real-time streaming to an iPad would immediately enable effective patient education. Optimally, a multi-user option would also be integrated. In an ideal educational scenario, this would support multiple users simultaneously, both co-located and remotely located (Perz et al., 2024).

Cost considerations are likely less critical for surgical or treatment-related applications. In these scenarios, participants particularly valued the photorealistic rendering, intuitive eye-based navigation, and rapid setup, all beneficial for preoperative planning. Qualitative feedback further supports the view that immersive visualization clarifies anatomical variations and spatial relationships. Such clarity may enable more confident, personalized surgical decisions. Some surgeons even suggested intraoperative applications, like quickly reassessing puncture or biopsy entry points during procedures. While promising, targeted clinical studies are needed to verify this benefit. Such studies could feasibly be executed within relatively short timelines. Studies should focus on measurable clinical impact and integration, prioritizing interventions where the most impact on patient outcome is expected. Even small studies could eventually be combined into a larger meta-analysis.

However, to fully realize the potential of CR on the AVP for preoperative planning, additional features are needed. Examples include drawing lines to visualize entry points or perform precise measurements. Viewing multiple scans simultaneously, or fusing multi-modal scans, would ensure that all relevant imaging data is visible in one cohesive overview. Other desired functionalities include incorporating artificial intelligence-driven segmentations, adding annotations, and removing anatomical structures via hand gestures. In a future scenario, segmentation could even be pre-generated automatically, allowing voice commands to highlight or render specific anatomical structures transparent.

Integrating these advanced features, particularly combined with real-time streaming to an iPad and multi-user capabilities, would also greatly enhance multidisciplinary case discussions. Additionally, incorporating automatic overlays with error indications could make the system suitable for intraoperative guidance. Such functionality might also be highly beneficial for ultrasound training. For instance, clearly indicating ultrasound imaging planes and corresponding anatomical structures could significantly enhance learning and spatial comprehension.

A crucial final feature, requiring close cooperation between hospitals and software developers, is integration with hospital electronic patient record systems. Currently, creating and transferring visualization scenes to the AVP still involves manual steps. Automating these processes would significantly reduce staff workload and associated costs, an important factor considering the current healthcare labor shortages.

From a commercial perspective, especially in insurance-driven healthcare markets, clearly defining a clinical use case with sufficient benefits to justify these additional features is essential. For clinical researchers, commercially superior products like Siemens’ CR offer robust testing environments. However, open-source alternatives might become attractive due to their flexibility for rapidly developing research-specific functionalities (Baseer et al., 2023).

Ultimately, we anticipate that cinematic 3DVR will become a standard feature integrated directly into advanced medical imaging equipment, such as state-of-the-art CT and MRI systems.

4.5 Conclusion and future directions

This study suggests the potential of immersive cinematic 3D volume rendering using Siemens’ CR application on the AVP to enhance medical image interpretation. Participating surgeons and trainees gave high usability ratings, indicating that the current implementation is usable in this pilot setting and may support future clinical adoption if subsequent studies demonstrate effectiveness. The intuitive interface facilitates detailed, photorealistic visualization of CT and MRI data. It was perceived as particularly promising for educational purposes, surgical planning, and intraoperative visualization of complex anatomy and anatomical variations.

The ISONORM profile, which showed high suitability, consistency, and controllability but weaker self-descriptiveness and customizability, was consistent with the open-ended feedback. Clinicians repeatedly requested advanced segmentation capabilities, more robust measurement and annotation tools, multimodal fusion, streaming or multi-user options, voice control, and smoother integration with clinical data systems as prerequisites for routine clinical use. Addressing these feature requests is likely to further enhance both the usability and the perceived clinical relevance of this technology.

Future research should focus on longitudinal, real-world evaluations of CR-assisted clinical workflows. Controlled studies are needed to determine measurable benefits, such as improved surgical accuracy, reduced operative times, or enhanced educational outcomes. Moreover, technical enhancements should emphasize collaborative functionalities, including multi-user capabilities, annotations (lines and notes), remote consultations, and voice-controlled interactions. Implementing these features would likely accelerate clinical acceptance, as suggested by our participants’ feedback.

In summary, while the core usability of CR on the AVP was rated good to excellent, our participants identified key barriers to clinical adoption, including the absence of workflow-specific features, seamless integration with hospital systems, and operability in sterile environments for intraoperative usage.

Finally, while our study focused specifically on hepatobiliary surgery, future investigations should expand into other medical fields, such as orthopedics, cardiology, and interventional radiology. Broadening the scope of evaluation will be necessary to assess the generalizability and potential clinical impact of immersive 3D visualization across modern medicine.

5 Resource identification initiative

At the time of writing, Siemens’ Cinematic Reality and Cinematic Playground, as well as the Apple Vision Pro, are not registered with RRIDs (Research Resource Identifiers). As such, no RRIDs can be provided for the resources used in this study.

6 Life science identifiers

Life Science Identifiers are not applicable to this study, as it does not involve zoological nomenclature or related taxonomic acts.

Data availability statement

The datasets used in this study are publicly available in a publicly accessible repository. This data can be found here: CHAOS (https://doi.org/10.5281/zenodo.3431873, v1.03) and MRCP DLRecon (https://doi.org/10.5281/zenodo.13912092, v3). The questionnaires used in the user study and the scripts used to prepare the public imaging data into formats suitable for the Cinematic Reality application will be made available by the authors upon request, without undue reservation.

Ethics statement

The requirement of ethical approval was waived by Ethics committee of the medical faculty of the university of Duisburg -Essen and university hospital Essen for the studies involving humans because the study involved no significant physical or psychological risk to participants or patients. All participants were medical professionals or students who voluntarily interacted with a non-invasive visualization system (Apple Vision Pro headset).

Author contributions

GL: Conceptualization, Data curation, Formal Analysis, Investigation, Methodology, Project administration, Software, Validation, Visualization, Writing – original draft, Writing – review and editing. LFdP: Validation, Visualization, Writing – original draft, Writing – review and editing. SK: Methodology, Software, Writing – review and editing, Resources. AB: Methodology, Resources, Software, Supervision, Writing – review and editing. LM: Data curation, Investigation, Project administration, Writing – review and editing. AS: Validation, Visualization, Writing – review and editing. PH: Investigation, Methodology, Writing – review and editing. JK: Conceptualization, Methodology, Resources, Writing – review and editing. SS: Conceptualization, Investigation, Methodology, Resources, Supervision, Writing – review and editing. UN: Conceptualization, Methodology, Resources, Writing – review and editing. JE: Conceptualization, Methodology, Resources, Supervision, Writing – original draft, Writing – review and editing.

Funding

The author(s) declared that financial support was received for this work and/or its publication. This work was made possible through a collaboration with Siemens Healthineers, who provided early access to their Cinematic Reality Application on the Apple Vision Pro via TestFlight. We also benefited from access to the Siemens Cinematic Playground, which enabled us to visualize multiple publicly available scans (CHAOS dataset (Kavur et al., 2021; Kavur et al., 2019) and the MRCP_DLRecon dataset (Kim et al., 2025; Kim et al., 2024)) in 3D Cinematic Reality. We thank Siemens Healthineers for their support and commitment to this project. This research was partially supported by the REACT-EU project KITE (grant number EFRE-2920801977, Plattform für KI-Translation Essen, https://kite.ikim.nrw/). We further acknowledge funding from the European Union under Grant Agreement No. 101168715 (INSIDE: INSIGHT, https://inside-insight.eu/). The views and opinions expressed are those of the authors and do not necessarily reflect those of the European Union or the granting authority. Neither the European Union nor the granting authority can be held responsible for them. Finally, we acknowledge support by the Open Access Publication Fund of the University of Duisburg-Essen.

Acknowledgements

We thank the Department of General-, Visceral- and Transplant Surgery, Medical Center University Duisburg-Essen, Essen, Germany, for their support and for enabling this research within their clinical environment. We are grateful to the open-source community for enabling reproducible research in medical imaging. This study made use of publicly available datasets, including the CHAOS dataset (Kavur et al., 2021; Kavur et al., 2019) and the MRCP_DLRecon dataset (Kim et al., 2025; Kim et al., 2024). We also acknowledge Siemens Healthineers for providing early access to the Cinematic Playground and Cinematic Reality software. A preprint version of this manuscript is available on arXiv Luijten et al. (2025). Finally, we acknowledge support by the Open Access Publication Fund of the University of Duisburg-Essen.

Conflict of interest

Authors SK and AB were employed by Siemens Healthineers.

The remaining author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was used in the creation of this manuscript. The authors disclose the intermittent use of generative AI tools between February and July 2025 to assist in refining grammar, phrasing, tone, and writing style. The following tools were used: DeepL Write (free version); DeepL SE, Cologne, Germany, https://www.deepl.com/write. Grammarly: AI Writing and Grammar Checker (paid version, Chrome extension v14.1242.0); Grammarly Inc., San Francisco, CA, United States, https://www.grammarly.com. ChatGPT-4 (via ChatGPT Plus subscription, model GPT-4-turbo); OpenAI, San Francisco, CA, United States, https://chat.openai.com. All content was critically reviewed and edited by the authors to ensure accuracy and adherence to scientific standards.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/frvir.2025.1666614/full#supplementary-material

References

Adnan, S., Benson, A. C., and Xiao, J. (2025). How virtual reality is being adopted in anatomy education in health sciences and allied health: a systematic review. Anat. Sci. Educ. 18, 496–525. doi:10.1002/ase.70027

PubMed Abstract | CrossRef Full Text | Google Scholar

Bangor, A., Kortum, P. T., and Miller, J. T. (2008). An empirical evaluation of the system usability scale. Intl. J. Human–Computer Interact. 24, 574–594. doi:10.1080/10447310802205776

CrossRef Full Text | Google Scholar

Bangor, A., Kortum, P., and Miller, J. (2009). Determining what individual sus scores mean: adding an adjective rating scale. J. Usability Studies 4, 114–123. doi:10.5555/2835587.2835589

CrossRef Full Text | Google Scholar

Baseer, I., Valverde, I., Moustafa, A. H., Blat, J., and Camara, O. (2023). “Evaluation of cinematic volume rendering open-source and commercial solutions for the exploration of congenital heart data,” in 2023 IEEE Visualization and Visual Analytics (VIS), 76–80. doi:10.1109/VIS54172.2023.00024

CrossRef Full Text | Google Scholar

Bevan, N., Carter, J., and Harker, S. (2015). “ISO 9241-11 revised: what have we learnt about usability since 1998?,” in Human-Computer Interaction: Design and Evaluation. Editor M. Kurosu (Springer International Publishing), 143–151.

CrossRef Full Text | Google Scholar

Binder, J., Krautz, C., Engel, K., Grützmann, R., Fellner, F. A., Burger, P. H. M., et al. (2019). Leveraging medical imaging for medical education – a cinematic rendering-featured lecture. Ann. Anat. - Anatomischer Anzeiger 222, 159–165. doi:10.1016/j.aanat.2018.12.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Binder, J. S., Scholz, M., Ellmann, S., Uder, M., Grützmann, R., Weber, G. F., et al. (2021). Cinematic rendering in anatomy: a crossover study comparing a novel 3D reconstruction technique to conventional computed tomography. Anat. Sci. Educ. 14, 22–31. doi:10.1002/ase.1989

PubMed Abstract | CrossRef Full Text | Google Scholar

Blanche, P.-A. (2021). Holography, and the future of 3D display. Light Adv. Manuf. 2, 446–459. doi:10.37188/lam.2021.028

CrossRef Full Text | Google Scholar

Brooke, J. (1986). System usability scale (sus) questionnaire. Digit. Equip. Corp. Available online at: https://rickvanderzwet.nl/trac/personal/export/104/liacs/hci/docs/SUS-questionaire.pdf(Accessed 25 June 2025)

Google Scholar

Brooke, J. (1996). SUS: a ’Quick and Dirty’ usability scale. Usability Eval. Industry. doi:10.1201/9781498710411-35

CrossRef Full Text | Google Scholar

Brooke, J. (2013). Sus: a retrospective. J. Usability Stud. 8, 29–40.

Google Scholar

Brookmeyer, C., Chu, L. C., Rowe, S. P., and Fishman, E. K. (2024). Clinical implementation of cinematic rendering. Curr. Problems Diagnostic Radiology 53, 313–328. doi:10.1067/j.cpradiol.2024.01.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Chu, J. C. H., Gong, X., Cai, Y., Kirk, M. C., Zusag, T. W., Shott, S., et al. (2009). Application of holographic display in radiotherapy treatment planning II: a multi-institutional study. J. Appl. Clin. Med. Phys. 10, 115–124. doi:10.1120/jacmp.v10i3.2902._eprint

PubMed Abstract | CrossRef Full Text | Google Scholar

Cruz-Neira, C., Sandin, D. J., and DeFanti, T. A. (1993). “Surround-screen projection-based virtual reality: the design and implementation of the CAVE,” in Proceedings of the 20th annual conference on Computer graphics and interactive techniques (New York, NY, USA: Association for Computing Machinery), 135–142. doi:10.1145/166117.166134

CrossRef Full Text | Google Scholar

Cruz-Neira, C., Reiners, D., and Springer, J. P. (2010). An affordable surround-screen virtual reality display. J. Soc. Inf. Disp. 18, 836–843. doi:10.1889/JSID18.10.836

CrossRef Full Text | Google Scholar

Csebfalvi, B., and Szirmay-Kalos, L. (2003). Monte carlo volume rendering. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 449–456. doi:10.1109/VISUAL.2003.1250406

CrossRef Full Text | Google Scholar

Dappa, E., Higashigaito, K., Fornaro, J., Leschka, S., Wildermuth, S., and Alkadhi, H. (2016). Cinematic rendering – an alternative to volume rendering for 3D computed tomography imaging. Insights Into Imaging 7, 849–856. doi:10.1007/s13244-016-0518-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Douglas, D. B., Wilke, C. A., Gibson, J. D., Boone, J. M., and Wintermark, M. (2017). Augmented reality: advances in diagnostic imaging. Multimodal Technol. Interact. 1, 29. doi:10.3390/mti1040029

CrossRef Full Text | Google Scholar

Duran, A. H., Duran, M. N., Masood, I., Maciolek, L. M., and Hussain, H. (2019). The additional diagnostic value of the three-dimensional volume rendering imaging in routine radiology practice. Cureus 11, e5579. doi:10.7759/cureus.5579

PubMed Abstract | CrossRef Full Text | Google Scholar

Egger, J., Gsaxner, C., Luijten, G., Chen, J., Chen, X., Bian, J., et al. (2024). Is the apple vision pro the ultimate display? A first perspective and survey on entering the wonderland of precision medicine. JMIR Serious Games 12, e52785. doi:10.2196/52785

PubMed Abstract | CrossRef Full Text | Google Scholar

Eid, M., De Cecco, C. N., Nance, J. W., Caruso, D., Albrecht, M. H., Spandorfer, A. J., et al. (2017). Cinematic rendering in CT: a novel, lifelike 3D visualization technique. Am. J. Roentgenol. 209, 370–379. doi:10.2214/AJR.17.17850

PubMed Abstract | CrossRef Full Text | Google Scholar

Elshafei, M., Binder, J., Baecker, J., Brunner, M., Uder, M., Weber, G. F., et al. (2019). Comparison of cinematic rendering and computed tomography for speed and comprehension of surgical anatomy. JAMA Surg. 154, 738–744. doi:10.1001/jamasurg.2019.1168

PubMed Abstract | CrossRef Full Text | Google Scholar

Erbay, N., Raptopoulos, V., Pomfret, E. A., Kamel, I. R., and Kruskal, J. B. (2003). Living donor liver transplantation in adults: vascular variants important in surgical planning for donors and recipients. Am. J. Roentgenol. 181, 109–114. doi:10.2214/ajr.181.1.1810109

PubMed Abstract | CrossRef Full Text | Google Scholar

Fellner, F. A. (2016). Introducing cinematic rendering: a novel technique for post-processing medical imaging data. J. Biomed. Sci. Eng. 9, 170–175. doi:10.4236/jbise.2016.93013

CrossRef Full Text | Google Scholar

Fuchs, H., Levoy, M., and Pizer, S. M. (1989). Interactive visualization of 3d medical data. Computer 22, 46–51. doi:10.1109/2.35199

CrossRef Full Text | Google Scholar

Fukumoto, W., Kitera, N., Mitani, H., Sueoka, T., Kondo, S., Kawashita, I., et al. (2022). Global illumination rendering versus volume rendering for the forensic evaluation of stab wounds using computed tomography. Sci. Rep. 12, 2452. doi:10.1038/s41598-022-06541-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Gehrsitz, P., Rompel, O., Schöber, M., Cesnjevar, R., Purbojo, A., Uder, M., et al. (2021). Cinematic rendering in mixed-reality holograms: a new 3d preoperative planning tool in pediatric heart surgery. Front. Cardiovasc. Med. 8, 633611. doi:10.3389/fcvm.2021.633611

PubMed Abstract | CrossRef Full Text | Google Scholar

Gsaxner, C., Li, J., Pepe, A., Jin, Y., Kleesiek, J., Schmalstieg, D., et al. (2023). The HoloLens in medicine: a systematic review and taxonomy. Med. Image Anal. 102757, 102757. doi:10.1016/j.media.2023.102757

PubMed Abstract | CrossRef Full Text | Google Scholar

Harley, D. (2020). Palmer luckey and the rise of contemporary virtual reality. Convergence 26, 1144–1158. doi:10.1177/1354856519860237

CrossRef Full Text | Google Scholar

Healthineers, S. (2017). Cinematic rendering in medical imaging. Available online at: https://www.siemens-healthineers.com/digital-health-solutions/cinematic-rendering(Accessed 25 June 2025)

Google Scholar

Healthineers, S. (2024). Siemens healthineers launches cinematic reality app for apple vision pro. Available online at: https://www.siemens-healthineers.com/press/releases/cinematic-reality-applevisionpro(Accessed 25 June 2025)

Google Scholar

Held, R. T., and Hui, T. T. (2011). A guide to stereoscopic 3D displays in medicine. Acad. Radiol. 18, 1035–1048. doi:10.1016/j.acra.2011.04.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Hyzy, M., Bond, R., Mulvenna, M., Bai, L., Dix, A., Leigh, S., et al. (2022). System usability scale benchmarking for digital health apps: meta-Analysis. JMIR mHealth uHealth 10, e37290. doi:10.2196/37290

PubMed Abstract | CrossRef Full Text | Google Scholar

ISO (2020). ISO 9241-110:2020 Ergonomics of human–system interaction – part 110: interaction principles. Standard ISO 9241-110:2020. Geneva, Switzerland: International Organization for Standardization.

Google Scholar

Kang, D., Choi, J.-H., and Hwang, H. (2022). Autostereoscopic 3D display system for 3D medical images. Appl. Sci. 12, 4288. doi:10.3390/app12094288

CrossRef Full Text | Google Scholar

Kavur, A. E., Selver, M. A., Dicle, O., Barış, M., and Gezer, N. S. (2019). Chaos: combined (ct–mr) healthy abdominal organ segmentation challenge data. doi:10.5281/zenodo.3362844

CrossRef Full Text | Google Scholar

Kavur, A. E., Gezer, N. S., Barş, M., Aslan, S., Conze, P.-H., Groza, V., et al. (2021). CHAOS challenge - combined (CT-MR) healthy abdominal organ segmentation. Med. Image Anal. 69, 101950. doi:10.1016/j.media.2020.101950

PubMed Abstract | CrossRef Full Text | Google Scholar

Kelly, P., Fung, A., Qu, J., Greig, P., Tait, G., Jenkinson, J., et al. (2017). Depicting surgical anatomy of the porta hepatis in living donor liver transplantation. J. Visualized Surgery 3, 43. doi:10.21037/jovs.2017.03.08

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, J., Nickel, D., and Knoll, F. (2024). Deep learning-based accelerated MR cholangiopancreatography without fully-sampled data. NMR Biomed. 38, e70002. doi:10.5281/zenodo.13912092

CrossRef Full Text | Google Scholar

Kim, J., Nickel, M. D., and Knoll, F. (2025). Deep learning-based accelerated mr cholangiopancreatography without fully-sampled data. NMR Biomed. 38, e70002. doi:10.1002/nbm.70002

PubMed Abstract | CrossRef Full Text | Google Scholar

Kroes, T., Post, F. H., and Botha, C. P. (2012). Exposure render: an interactive photo-realistic volume rendering framework. PloS One 7, e38586. doi:10.1371/journal.pone.0038586

PubMed Abstract | CrossRef Full Text | Google Scholar

Kukla, P., Maciejewska, K., Strojna, I., Zapał, M., Zwierzchowski, G., and Bak, B. (2023). Extended reality in diagnostic imaging–a literature review. Tomography 9, 1071–1082. doi:10.3390/tomography9030088

PubMed Abstract | CrossRef Full Text | Google Scholar

Lanier, J. (1992). Virtual reality: the promise of the future. Interact. Learn. Int. 8, 275–279.

Google Scholar

Lewis, J. R. (2018). The system usability scale: Past, present, and future. Int. J. Human–Computer Interact. 34, 577–590. doi:10.1080/10447318.2018.1455307

CrossRef Full Text | Google Scholar

Li, K., Yan, R., Ma, H., Zhang, D.-f., Ding, Y., and Li, Z.-h. (2019). Value of the cinematic rendering from volumetric computed tomography data in evaluating the relationship between deep soft tissue sarcomas of the extremities and adjacent major vessels: a preliminary study. J. Comput. Assisted Tomogr. 43, 386–391. doi:10.1097/RCT.0000000000000852

PubMed Abstract | CrossRef Full Text | Google Scholar

Ljung, P., Krüger, J., Groller, E., Hadwiger, M., Hansen, C. D., and Ynnerman, A. (2016). “State of the art in transfer functions for direct volume rendering,” in Computer graphics forum, 669–691.

CrossRef Full Text | Google Scholar

Lopes, D. S., Parreira, P. F., Mendes, A. R., Pires, V. M., Paulo, S. F., Sousa, C., et al. (2018). Explicit design of transfer functions for volume-rendered images by combining histograms, thumbnails, and sketch-based interaction. Vis. Comput. 34, 1713–1723. doi:10.1007/s00371-017-1448-8

CrossRef Full Text | Google Scholar

Luijten, G., Paiva, L. F., Krueger, S., Brost, A., Mazilescu, L., Santos, A. S. F., et al. (2025). From screen to space: evaluating siemens’ cinematic reality. doi:10.48550/arXiv.2506.04972

CrossRef Full Text | Google Scholar

Mahmood, F., Chen, R., Sudarsky, S., Yu, D., and Durr, N. J. (2018). Deep learning with cinematic rendering: fine-tuning deep neural networks using photorealistic medical images. Phys. Med. & Biol. 63, 185012. doi:10.1088/1361-6560/aada93

PubMed Abstract | CrossRef Full Text | Google Scholar

Pachowsky, M. L., Morf, H., Simon, D., Schönau, V., Valor-Mendez, L., Knitza, J., et al. (2022). Cinematic rendering in rheumatic diseases–photorealistic depiction of pathologies improves disease understanding for patients. Front. Med. 9, 946106. doi:10.3389/fmed.2022.946106

PubMed Abstract | CrossRef Full Text | Google Scholar

Perz, M., Luijten, G., Kleesiek, J., Schmalstieg, D., Egger, J., and Gsaxner, C. (2024). MultiAR: a multi-user augmented reality platform for biomedical education. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. 2024, 1–7. doi:10.1109/EMBC53108.2024.10782948

PubMed Abstract | CrossRef Full Text | Google Scholar

Prümper, J. (1993). “Software-evaluation based upon ISO 9241 part 10,” in Human computer interaction. Editors T. Grechenig, and M. Tscheligi (Berlin, Heidelberg: Springer Berlin Heidelberg), 255–265.

CrossRef Full Text | Google Scholar

Prümper, J. (2006). ISONORM 9241/110-S: Kurzfragebogen zur software-evaluation. Presentation at the 45th DGfP Annual Meeting, Nuremberg, September. Available online at: http://www.uselab.tu-berlin.de/wiki/images/6/62/ISONorm_Kurzversion.pdf(Accessed 25 June 2025)

Google Scholar

Queisner, M., and Eisenträger, K. (2024). Surgical planning in virtual reality: a systematic review. J. Med. Imaging 11, 062603. doi:10.1117/1.JMI.11.6.062603

PubMed Abstract | CrossRef Full Text | Google Scholar

Ropinski, T., Döring, C., and Rezk-Salama, C. (2010). “Interactive volumetric lighting simulating scattering and shadowing,” in 2010 ieee Pacific visualization symposium (pacificvis), 169–176.

Google Scholar

Sutherland, I. E. (1968). “A head-mounted three dimensional display,” in Proceedings of the December 9-11, 1968, fall joint computer conference, part I (New York, NY, USA: Association for Computing Machinery), 757–764. AFIPS ’68 (Fall, part I). doi:10.1145/1476589.1476686

CrossRef Full Text | Google Scholar

Tian, N., Lopes, P., and Boulic, R. (2022). A review of cybersickness in head-mounted displays: raising attention to individual susceptibility. Virtual Real. 26, 1409–1441. doi:10.1007/s10055-022-00638-2

CrossRef Full Text | Google Scholar

van Beurden, M. H. P. H., Ijsselsteijn, W. A., and Juola, J. F. (2012). Effectiveness of stereoscopic displays in medicine: a review. 3D Res. 3, 3. doi:10.1007/3DRes.01(2012)3

CrossRef Full Text | Google Scholar

Vlachogianni, P., and Tselios, N. (2022). Perceived usability evaluation of educational technology using the system usability scale (sus): a systematic review. J. Res. Technol. Educ. 54, 392–409. doi:10.1080/15391523.2020.1867938

CrossRef Full Text | Google Scholar

Yeo, C. T., MacDonald, A., Ungi, T., Lasso, A., Jalink, D., Zevin, B., et al. (2018). Utility of 3d reconstruction of 2d liver computed tomography/magnetic resonance images as a surgical planning tool for residents in liver resection surgery. J. Surgical Education 75, 792–797. doi:10.1016/j.jsurg.2017.07.031

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, Q., Eagleson, R., and Peters, T. M. (2011). Volume visualization: a technical overview with a focus on medical applications. J. Digital Imaging 24, 640–664. doi:10.1007/s10278-010-9321-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: extended reality, augmented reality, apple vision pro, 3D medical imaging, volume rendering, clinical integration, usability assessment, cinematic rendering

Citation: Luijten G, Faray de Paiva L, Krueger S, Brost A, Mazilescu L, Santos AS, Hoyer P, Kleesiek J, Schmitz SM-T, Neumann UP and Egger J (2026) From screen to space: evaluating Siemens’ Cinematic Reality application for medical imaging on the Apple Vision Pro. Front. Virtual Real. 6:1666614. doi: 10.3389/frvir.2025.1666614

Received: 15 July 2025; Accepted: 28 November 2025;
Published: 05 January 2026.

Edited by:

Caitlin R. Rawlins, Department of Veterans Affairs, United States

Reviewed by:

Eleftherios Garyfallidis, Indiana University Bloomington, United States
Shreeraj Jadhav, Kitware, United States

Copyright © 2026 Luijten, Faray de Paiva, Krueger, Brost, Mazilescu, Santos, Hoyer, Kleesiek, Schmitz, Neumann and Egger. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Gijs Luijten, R2lqcy5MdWlqdGVuQHVrLWVzc2VuLmRl; Jan Egger, SmFuLkVnZ2VyQHVrLWVzc2VuLmRl

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.