- 1Institute for Artificial Intelligence in Medicine (IKIM), Essen University Hospital (AöR), University of Duisburg-Essen, Essen, Germany
- 2Center for Virtual and Extended Reality in Medicine (ZvRM), University Hospital Essen (AöR), Essen, Germany
- 3Institute of Computer Graphics and Vision (ICG), Graz University of Technology, Graz, Austria
- 4Siemens Healthineers, Forchheim, Germany
- 5Department of General-, Visceral- and Transplant Surgery, Medical Center University Duisburg-Essen, Essen, Germany
- 6Pediatric Clinic II, University Children’s Hospital Essen, University Duisburg-Essen, Essen, Germany
- 7Medical Faculty, University of Duisburg-Essen, Essen, Germany
- 8Cancer Research Center Cologne Essen (CCCE), West German Cancer Center, University Hospital Essen (AöR), Essen, Germany
- 9German Cancer Consortium (DKTK), Partner site University Hospital Essen (AöR), Essen, Germany
- 10Department of Physics, Technical University Dortmund, Dortmund, Germany
- 11Faculty of Computer Science, University of Duisburg-Essen, Essen, Germany
- 12Department of Surgery, Maastricht University Medical Centre+, Maastricht, Netherlands
Introduction: As one of the first research teams with full access to Siemens’ Cinematic Reality, we evaluated its usability and clinical potential for cinematic volume rendering on the Apple Vision Pro.
Methods: We visualized venous-phase liver computed tomography and magnetic resonance cholangiopancreatography scans from the CHAOS and MRCP_DLRecon public datasets, respectively. Fourteen medical experts assessed usability and anticipated clinical integration potential using standardized questionnaires (System Usability Scale and ISONORM 9242-110-S) and an open-ended survey. Our primary aim was not to validate direct clinical outcomes, but to evaluate the usability of Siemens’ Cinematic Reality on the Apple Vision Pro and gather expert feedback on potential use cases and missing features required for clinical adoption beyond educational purposes.
Results: Their feedback identified feasibility, key usability strengths, and required features to catalyze the adaptation in real-world clinical workflows.
Conclusion: The findings provide insights into the potential of immersive cinematic rendering in medical imaging and the needed features for clinical adoption as suggested by the medical experts. Siemens Cinematic Reality running on the Apple Vision Pro was deemed to have good usability, making it a promising tool.
1 Introduction
Head-mounted displays (HMDs) enable the viewing of three-dimensional (3D) medical images in a 3D environment, rather than on traditional monitors. The first HMD was developed in 1968 by Sutherland (1968). Jaron Lanier later popularized the concept, introducing the promise of virtual reality (Lanier, 1992). However, widespread adoption began only after Palmer Lucky introduced consumer-friendly HMDs in 2012, sparking rapid development (Harley, 2020). The Apple Vision Pro (Apple Inc., Cupertino, California, United States) is currently the newest standalone high-end extended reality (XR) device available. Compared to earlier systems such as the Microsoft HoloLens 1 and 2 (Microsoft Corp., Redmond, WA) or the Meta Quest 3 (Meta Platforms Inc., Menlo Park, CA), the Apple Vision Pro (AVP) offers superior display quality, computational performance, sensors, and more precise hand-eye interactions, potentially overcoming past issues with image clarity and usability (Egger et al., 2024).
In parallel with these hardware developments, there has long been interest in instantly visualizing 2D medical image slices, such as those from computed tomography (CT) or magnetic resonance imaging (MRI), as accurate 3D representations. Ideally, this process would require no manual segmentation efforts (Zhang et al., 2011). Volume rendering, a technique introduced in the late 1980s, enabled the display of 3D representations of CT and MRI data on 2D monitors (Fuchs et al., 1989). HMDs now enable three-dimensional volume rendering (3DVR) of medical scans in immersive environments (Douglas et al., 2017). Ongoing advancements continue expanding its medical and educational applications (Ljung et al., 2016; Zhang et al., 2011; Eid et al., 2017; Dappa et al., 2016), with research indicating that 3DVR improves radiology diagnostics, surgical decision-making, and anatomical understanding by providing superior spatial perception compared to slice-based imaging methods (Duran et al., 2019; Queisner and Eisenträger, 2024).
While standard volume rendering provides crucial 3D spatial perception, photorealistic rendering techniques aim to further enhance anatomical realism and depth cues. Potentially leading to even greater clinical utility, particularly in education or complex surgical planning (Eid et al., 2017; Li et al., 2019; Dappa et al., 2016).
Siemens’ Cinematic Reality (Siemens Healthineers, Medical Imaging Technologies, Erlangen, Germany) (Healthineers, 2017) applies these cinematic rendering techniques to enhance traditional VR, offering photorealistic 3D depictions using global illumination models that simulate natural light, shadows, refraction, and occlusion (Kroes et al., 2012; Ropinski et al., 2010; Csebfalvi and Szirmay-Kalos, 2003; Fellner, 2016). Siemens Healthineers initially developed this Cinematic Reality (CR) application for the computer and later the Microsoft HoloLens (Fellner, 2016; Healthineers, 2024).
To contextualise these advances, it is helpful to outline the limitations of the conventional rendering methods that CR seeks to improve upon. Standard 3DVR on CT provides crucial 3D spatial perception, but typically relies on user-defined transfer functions and local lighting models. In this conventional workflow, a transfer function maps ranges of Hounsfield units to colour and opacity; when set aggressively, this can introduce harsh opacity transitions and stripe-like artefacts, and depth cues remain limited because only local shading around each voxel is considered. Isosurface rendering, commonly used on clinical workstations, further simplifies the volume into a hard surface defined by a single intensity threshold, often yielding unnaturally sharp, “plastic” boundaries between tissues. Photorealistic, global-illumination–based cinematic rendering techniques aim to overcome these limitations by simulating how light is transported, scattered, and absorbed in the volume, thereby providing smoother shading, soft occlusion, and more intuitive depth perception, which can translate into greater clinical utility, particularly for education and complex surgical planning (Eid et al., 2017; Li et al., 2019; Dappa et al., 2016; Fukumoto et al., 2022). These differences across rendering approaches are illustrated in Figure 1, which shows the same CHAOS Subject 5 CT volume rendered with Siemens’ CR, with standard 3DVR using 3D Slicer (version 5.10.0; open-source medical imaging platform - https://www.slicer.org/), with Slicer’s maximum-intensity-projection–style 3D view, and with an isosurface generated using Slicer’s surface tools.
Figure 1. Comparison of rendering techniques for a contrast-enhanced abdominal CT scan (CHAOS dataset, Subject 5). Panels A1–A2 show Siemens’ Cinematic Reality renderings that illustrate photorealistic shading and soft-tissue detail. Panels B1–B2 depict standard three-dimensional volume rendering using 3D Slicer, with local lighting and the standard CT-Liver transfer function included in the default settings. Panel C shows a maximum-intensity-projection (MIP) view generated in 3D Slicer’s volume rendering module, and panel D shows an isosurface model created with the Surface tool in 3D Slicer. All views are derived from the same CT volume and demonstrate how CR provides smoother shading and more intuitive depth cues compared with conventional 3DVR, MIP, and isosurface rendering.
Siemens’ CR applies such cinematic rendering techniques to medical CT and MRI data. In contrast to standard 3DVR and isosurface rendering, CR uses a physically based global illumination model that generates soft shadows, ambient occlusion, and view-dependent highlights Kroes et al. (2012); Ropinski et al. (2010); Csebfalvi and Szirmay-Kalos (2003); Fellner (2016). High–dynamic-range lighting combined with tone mapping compresses the simulated radiance into the displayable range while preserving local contrast, and tissue-aware transfer function presets map relevant intensity ranges (e.g., bone, vessels, or parenchyma) to realistic colours and translucency. These tissue-specific presets reduce manual tuning while emphasising clinically relevant anatomy in a visually consistent way. Together, these differences in illumination, tone mapping, and tissue-aware transfer functions explain why CR images are often perceived as more realistic than conventional 3DVR and isosurface rendering and motivate their evaluation on emerging XR platforms such as the Apple Vision Pro.
Prior studies have demonstrated CR’s potential, such as its application on the Microsoft HoloLens for pediatric heart surgery planning (Gehrsitz et al., 2021). Despite these advancements, the clinical benefits of photorealistic rendering techniques such as Siemens’ CR, particularly in routine surgical decision-making, remain unclear (Duran et al., 2019; Gehrsitz et al., 2021).
Since the clinical benefits remain unclear, assessing the usability of such systems (HMDs with cinematic rendering applications) and identifying areas of clinical potential is necessary. It’s especially interesting to look beyond anatomy education. Surgical planning and intervention are an interesting example where minor improvements can still warrant the overall cost. Liver transplantation exemplifies such a domain where precise 3D visualization is paramount due to its intricate and highly variable vascular and biliary anatomy (Kelly et al., 2017; Yeo et al., 2018; Erbay et al., 2003). Furthermore, surgical planning in this context often requires not only identifying anatomical variants but also critically assessing tumour-induced displacement or encasement of vital structures (vessels, bile ducts, parenchyma), tasks where immersive, high-fidelity photorealistic rendering could offer clinical value.
These previously mentioned uncertainties surrounding clinical effectiveness, combined with the recent advancements in XR devices and software, including the introduction of AVP with CR, provide a compelling opportunity to rigorously evaluate the usability and practical integration potential of CR into clinical workflows (Kukla et al., 2023; Lopes et al., 2018; Healthineers, 2024; Egger et al., 2024).
This study evaluates Siemens’ CR on the AVP, focusing on its interactive usability and its anticipated integration into clinical workflows, such as surgical planning, as assessed through expert clinician feedback. Siemens’ CR focuses on photorealistic visualization and standard interactions (Figures 2, 3), but does not yet include workflow-specific features such as measurements, segmentations, or annotations. Therefore, our study aimed not only to evaluate usability but, also to capture expert clinician feedback on which features are essential for future clinical adoption. The goal of this study was therefore twofold: to systematically evaluate the usability of Siemens’ CR on the AVP using validated questionnaires, and to capture clinician feedback identifying essential features for integration into preoperative and intraoperative planning workflows.
Figure 2. Cinematic 3DVR examples: Two different transfer functions and perspectives for (A) Siemens’ demo case and (B) the MRCP_DLRecon dataset case (Kim et al., 2024), and (C) a single example of subject four from the CHAOS dataset (Kavur et al., 2019).
Figure 3. Core functionalities of Siemens' Cinematic Reality on the Apple Vision Pro. Eye gaze functions as a pointer, and finger pinching acts as a selection tool (click). Top left: Access the library to load scenes (i.e., DICOM files converted to the appropriate scene files). Rendering happens on the AVP, depending on the operator's viewpoint. Top right: 3D model interactions include scrolling/clipping (top left), resizing (top right), windowing (bottom left), and rotation (bottom right). Presets enable tissue-specific transfer functions (e.g., orange rendering in the lower-right model); clipping planes/boxes and light map settings adjust visibility and lighting effects. View orientation can be switched via the 2D/3D overview (axial, sagittal, coronal). Bottom: From left to right, 2D slice interactions include zooming (two hands), windowing, scrolling, and switching between radiologic visualization modes. Rotation supports one or two hands; all other interactions use a single hand.
Direct deployment in clinical workflows was beyond the study’s scope; the evaluation with structured usability questionnaires and domain-specific expertise provides actionable insights into the system’s readiness, features, limitations, and opportunities for clinical adoption prior to clinical trials. While participants reflected on the subjective visual quality of the photorealistic rendering, these impressions were treated as part of the overall user experience rather than an isolated outcome.
To go beyond the standard demo case included in the CR application, public datasets related to abdominal visceral surgery are used. This resembles the realistic world scenario more closely than using an optimized demo case. Both CT and MRI, specifically magnetic resonance cholangiopancreatography (MRCP), scans from the CHAOS (Kavur et al., 2021; Kavur et al., 2019) and MRCP_DLRecon (Kim et al., 2025; Kim et al., 2024) datasets are used. Fourteen medical experts evaluated the system via the System Usability Scale (SUS) (Brooke, 1996; Brooke, 1986; Bangor et al., 2009; Lewis, 2018), ISONORM 9242–110-S (Prümper, 1993; Prümper, 2006; Bevan et al., 2015), and qualitative questionnaires.
This study aims to contribute to a broader understanding of how immersive 3D visualization, specifically through the CR application on the AVP, can enhance medical imaging interpretation. Moreover, the survey intends to provide a perspective on how these technologies can support clinicians, specifically liver surgeons, by identifying usability strengths, limitations, and desired features to catalyze clinical adoption.
2 Materials and methods
2.1 Study design and participants
A mixed-methods usability evaluation was conducted, incorporating both quantitative and qualitative assessments. The quantitative analysis used the System Usability Scale (SUS) score (Brooke, 1986) and ISONORM 9242-110-S (Prümper, 2006), which were chosen because they are validated and used as standards within the field. Qualitative feedback was collected through an open-ended survey that also included demographic information. The study included a cohort of 11 surgeons specializing in general, visceral, vascular, liver and transplant surgery, one doctor assistant, and two medical students.
Preoperative hepatobiliary planning is variable but usually includes initial consultation with review or request of contrast enhanced CT and/or MRI (MRCP), radiology demonstration of findings, and repeated surgeon review of 2D slices to assess vascular and biliary anatomy. Surgeons often formulate a plan with alternatives, which is then confirmed or adapted intraoperatively, most commonly with ultrasound and reinspection of the 2D slices. 3DVR and CR on HMDs have the potential to reduce repeated 2D review and intraoperative plan deviations. Our evaluation therefore focused on usability of CR’s current interaction set (Figure 3) and on eliciting expert-identified features needed for integration into this workflow (Tables 1 and 2).
Table 1. ISONORM 9241-110-S results across the seven usability principles. Values are reported as mean
Table 2. Positive and negative aspects regarding Siemens’ Cinematic Reality application running on the Apple Vision Pro.
The entire study was conducted at the University Hospital Essen (AöR), Essen, Germany. The Ethics Committee waived the Institutional Review Board (IRB) approval requirement because the study utilized publicly available, anonymized retrospective datasets and posed no risk or harm to the participating professionals and scanned subjects.
2.2 Datasets
The visualized 3D depictions of the AVP using the CR application were based on the following medical datasets:
1. CHAOS dataset (Kavur et al., 2019; Kavur et al., 2021): CT volumes (512
2. MRCP_DLRecon dataset (Kim et al., 2024; Kim et al., 2025): An MRCP scan (480
From the CHAOS dataset (Kavur et al., 2021; Kavur et al., 2019), subjects four and five were selected to ensure a representative scan of both a male and a female, and they had 94 and 95 slices, respectively. The MRCP_DLRecon dataset (Kim et al., 2025; Kim et al., 2024) consists of a single volunteer scan with an unidentified gender. To prepare volumetric data for Siemens’ CR on the AVP, the so-called scenes files were generated from DICOM using Siemens’ Cinematic Playground, which requires input in DICOM series format. This process involved only file conversion and scene setup; actual rendering and interaction occur on the AVP based on the user’s viewpoint. The CHAOS dataset was already in DICOM format, while the MRCP_DLRecon dataset, originally stored in HDF5 (.h5) format, was converted using a Python script (Python Software Foundation, Wilmington, DE, United States), utilizing the h5py, numpy, and pydicom libraries. Scenes were loaded onto the AVP via iCloud, with setup times under 1 min. No manual adjustments or transfer function customization were performed; the scans were imported as-is without cropping or editing. Total time per scene preparation was thus under 5 min.
2.3 Data preparation for cinematic reality
Siemens’ CR (v2.0 build 5) for the AVP (VisionOS v2.3.1) was installed via TestFlight (v3.7.1). Unlike the freely available demo version, this version supports the loading of custom-made scenes. The scenes were prepared using Siemens’ Cinematic Playground (v0.19) on a Windows laptop (Intel Core i7-11800H, NVIDIA GeForce RTX 3050) with a required Digital Imaging and Communications in Medicine (DICOM) series. The CHAOS dataset (Kavur et al., 2019) provided the DICOM series directly, while the MRCP_DLRecon dataset (Kim et al., 2024), stored in Hierarchical Data Format 5 (HDF5), was converted to a DICOM series.
The voxel values were normalized and linearly scaled [-1000 to 2000] using Python (version 12) and the following libraries: numpy (v2.2.2), h5py (v3.12.1), nibabel (v5.3.2), pydicom (v3.0.1), and scikit-image (v0.25.1). This scaling adjustment optimized compatibility with Siemens Cinematic Playground and CR’s transfer function. Scene files were then loaded onto the AVP via iCloud.
Siemens’ CR facilitates conventional interactions, including windowing, scrolling, and zooming. The system is also capable of 3D cinematic volume rendering, as illustrated in Figure 3.
2.4 Evaluation
Quantitative data from the SUS (Brooke, 1986) and ISONORM (Prümper, 2006) questionnaires were analyzed using descriptive statistics, including mean scores and standard deviations. Qualitative responses were thematically categorized based on strengths, weaknesses, current clinical applicability, feature requests, and potential future applications. The ISONORM 9241-110-S questionnaire assesses usability across seven principles derived from ISO 9241-110:
1. Suitability for the task: How well the system supports users in completing tasks.
2. Self-descriptiveness: The intuitiveness of system functionality.
3. Conformity with user expectations: Consistency with known interface standards.
4. Learnability: Ease with which new users can become proficient.
5. Controllability: The extent to which users can influence actions and outcomes.
6. Error tolerance: The system’s ability to prevent or recover from user errors.
7. Customizability: How well the system can be adapted to user needs.
These principles provide a comprehensive framework for evaluating the user-system interaction, focusing on effectiveness, efficiency, and satisfaction in a specified context of use (ISO, 2020).
The System Usability Scale (SUS) is a widely used tool for evaluating the usability of various systems, including medical software. It consists of a 10-item questionnaire with five response options for respondents, ranging from strongly agree to strongly disagree. SUS provides a single score ranging from 0 to 100, representing a composite measure of the overall usability of the system being studied (Brooke, 1986). Notably, scores above 68 are considered above average across industries and for digital health applications, with higher scores indicating better usability (Hyzy et al., 2022; Brooke, 2013). Scores between 74 and 85 indicate good to excellent usability (Brooke, 2013).
2.5 Procedure and data collection
For this study, we used a research build of Siemens’ CR for the Apple Vision Pro that supports loading CTs and MRIs beyond the demo case as locally prepared scene files; this build is currently shared only with external groups via research collaborations and is planned for broader release. All sessions took place on site at our institution and were overseen solely by the local research team; no Siemens Healthineers personnel were present during data collection.
Informed Consent: All participants provided written informed consent prior to study participation.
Device Setup: Streaming from the AVP to an iPad was activated via AirPlay to facilitate observation and assistance. Participants then put on the AVP headset and completed the built-in hand–eye calibration before using the Siemens’ CR application. The researcher guided participants in this process.
Application Interaction (Task): Participants launched the Siemens CR application, which was preloaded with CR scenes from the CHAOS and MRCP_DLRecon datasets. The datasets were presented in a fixed order on the AVP: participants first explored the CT volume from the CHAOS dataset, followed by the MRCP scan. Participants thus started with a CT volume from the CHAOS dataset. From this point onward, participants were not guided; help was provided only upon request. The datasets were presented in a fixed order: participants first explored the CT volume from the CHAOS dataset, followed by the MRCP scan. This sequence enabled them to examine both vascular and biliary anatomy and assess the utility of CR in distinct clinical contexts. After viewing both datasets, participants were free to revisit scenes or explore an additional demo head CT case provided by Siemens Healthineers. Participants were tasked with the following instruction: explore key anatomical structures and evaluate the system. Specifically, they were asked to (i) explore anatomical structures relevant to hepatobiliary surgery, including portal veins, bile ducts, and liver parenchyma; (ii) examine image fidelity and photorealism using cinematic volume rendering; (iii) navigate the interface (Figure 3) using standard interaction techniques such as zooming, panning, scrolling, windowing, and rotating the volume and CT slices, as well as adapting the transfer function and switching between scenes (Siemens’ demo scene, CHAOS, MRCP_DLRecon); (iv) verbally articulate their observations, usability impressions, and any encountered challenges as part of a think-aloud protocol; and (v) describe how they would ideally integrate the system into their clinical workflow (e.g., for surgical planning or patient case discussions), including which functionalities they would use and how. These interactions and comments were observed by the researcher in real-time via AirPlay streaming to an iPad. If assistance was required, the researcher provided guidance, utilizing the AirPlay stream to facilitate real-time support.
While no rigid task list or time limit was imposed, the average interaction time during the application interaction was approximately 15–20 min, while hand–eye calibration took less than 5 min. Researchers provided minimal assistance, intervening only when participants requested support.
Data Collection: Following the interaction session, participants completed the System Usability Scale (SUS), the ISONORM 9241-110-S questionnaire, and an open-ended survey capturing qualitative feedback. All responses were collected via Google Forms.
3 Results
3.1 Participant demographics
Fourteen subjects participated in the study, including two medical students, a doctor/surgeon’s assistant, and 11 surgeons, with a gender distribution of eight females and six males. The ages of the female subjects ranged from 22 to 43 years, with a mean of
The AVP does not permit the use of spectacles within the HMD; rather, it necessitates the acquisition of its proprietary insert lenses. Consequently, the diopter of the subjects is an essential factor. Among the ten participants who wear spectacles, one subject was classified as farsighted
Notably, only two subjects had prior experience with HMDs, which was limited to one and 8 h, respectively. The HMDs were the HoloLens 2 and/or Meta Quest 3 (Meta Platforms, Inc., Menlo Park, CA). All subjects were unfamiliar with 3DVR; their familiarity with 3D rendering was limited to segmentation-based 3D rendering.
3.2 System Usability Scale
The SUS is technology-agnostic and combines effectiveness, efficiency and satisfaction with high reliability (Cronbach alpha = 0.91) in a single score (0–100) (Bangor et al., 2009; Bangor et al., 2008; Vlachogianni and Tselios, 2022; Brooke, 1986). Based on over 10 years of empirical evidence, the score can be divided into seven categories, from worst to best imaginable (Bangor et al., 2009; Vlachogianni and Tselios, 2022).
The mean score was σ = 77.68 (σ = 15.01, IQR = 27.50; Q1 = 63.75, Q3 = 91.25), indicating a score between good and excellent. Surgeons and assistant doctors rated the system higher than students. There were no noticeable differences between age or gender groups. An overview of the results for each group is shown in Figure 4, including students, residents, and surgeons.
Figure 4. Visualization of SUS scores for each participant, grouped by profession. Mean SUS scores are indicated by horizontal lines, with shaded areas representing the Interquartile Range (IQR).
3.3 ISONORM 9242-110-S
ISONORM results across the seven usability principles are summarized in Table 1. Suitability, conformity, and controllability were rated highest, while self-descriptiveness and customizability showed the most room for improvement. Figures 5, 6 provide a visual representation of these results as a heatmap and a stacked bar plot, respectively.
Figure 6. Stacked bar plot of ISONORM responses grouped per usability measurement centered around 0%.
3.4 Qualitative feedback
The open-ended questionnaire delved into the advantages and disadvantages of the CR application on the AVP (Table 2), its potential applications in a medical settings, and its current clinical usability (Table 3), as well as specific complex use cases and desired improvements for both software.
Table 3. Potential medical use cases and feasibility in daily clinical routine for the Apple Vision Pro.
Surgeons identified key areas for enhancing the Siemens’ CR application on AVP to optimize its clinical utility. Recommended improvements include integration with electronic patient records, advanced measurement tools for tumor and vessel size assessment, and the ability to insert 3D models of surgical instruments. Enhanced segmentation functionality enabling seamless switching between anatomical structures (e.g., arteries, portal veins, bile ducts), while gesture-based removal of irrelevant areas should provide more focused visualizations. Additionally, multi-user and multi-platform compatibility to facilitate collaborative planning and real-time surgical discussions. Finally, incorporating a voice activation function was suggested to enhance operability in sterile settings or when both hands are occupied with other tasks.
The setup process for the AVP was typically less than 5 min and considered straightforward. This duration includes donning the headset, performing the initial eye and hand calibration, and launching the CR application. Such a setup time is likely more acceptable for preoperative planning sessions, where time constraints are less critical. However, this setup duration may be undesirable for intraoperative applications, but the participants did not see this as a problem. In practice, once the device is calibrated and the CR application is running, re-donning the AVP and waking it from standby can be performed considerably faster than the full initial setup.
In addition to qualitative comments on usability, participants also described their real-time experience with Siemens’ CR on the Apple Vision Pro. Although no performance measurements were recorded, all users who were able to wear the headset reported smooth interaction and perceived the application as running in real-time, consistent with AVP’s reported refresh rate of 90 Hertz. No lag was noted during rotation, clipping or window adjustments. Eye-gaze selection and pinch-to-click gestures remained accurate throughout each session following the brief calibration, and no tracking interruptions occurred. Most participants used the system for about 15–20 min without discomfort, except for one individual who experienced mild nausea near the end. One participant was unable to use the device due to incompatible glasses, and a few noted minor light leakage from imperfect face-sealing, though this did not affect comfort. Some participants needed a short moment to adjust to the hand-eye coordination for certain gestures, particularly wrist-based rotation, but this reflected user familiarity rather than system issues. No crashes or freezes occurred. Overall, these observations indicate that under the evaluated conditions, the AVP provided stable and comfortable performance for real-time exploration of cinematic rendering.
4 Conclusion and discussion
This study presents one of the first empirical evaluations of Siemens’ Cinematic Reality on the Apple Vision Pro for hepatobiliary CT and MR imaging. Clinicians and trainees reported good to excellent usability on standardized instruments (SUS and ISONORM) and provided convergent qualitative feedback about strengths and missing functionality. Taken together, these results indicate that the current AVP–CR implementation is usable in a controlled setting and delineate the feature gaps that must be addressed. Still, they do not yet demonstrate effectiveness or impact on clinical or educational outcomes.
4.1 Contextualizing the findings
In contrast with the current surgical planning, (immersive) 3D visualization may add value by improving spatial understanding of tumor-vessel-bile duct relationships and by enabling rapid exploration of alternative resection planes directly in 3D rather than on 2D slices before intraoperative confirmation. This may reduce intraoperative plan deviations, shorten planning time, and improve surgeon confidence. As highlighted in Tables 1, 2, our experts already emphasized these use cases. Importantly, our study also identifies which functionalities are prerequisites for unlocking this potential in clinical planning: measurements, annotations, segmentation, and multi-modal fusion. Without these, the technology remains mainly useful for education and demonstration; with them, the step toward clinical decision support becomes more feasible.
While the focus of this work is on Siemens’ CR on the AVP, there is a long-standing use of non- and semi-immersive visualization systems for 3D visualization in medical imaging, such as cave automatic virtual environments (CAVEs), multi-projector setups for holographic visualization, and stereoscopic monitors.
Cruz-Neira et al. (1993) pioneered surround-screen VR such as the CAVE which is promising for medical teaching (Cruz-Neira et al., 1993). Yet, the physical space required and the financial cost make it less accessible than current cheaper HMDs (Cruz-Neira et al., 2010), which might be a reason its adoption in, for example, anatomy education is rare (Adnan et al., 2025). An advantage of CAVEs is the direct option for multi-user interaction, which Siemens’ CR on the AVP or HoloLens 2 does not have (Table 2). Additionally, no extra (heavy) head-mounted hardware is needed, allowing users more comfort and the ability to wear their own glasses, unlike in the AVP. The same can be said for holographic projections showing promise, but likely not enough to warrant the investment for clinical purposes (Blanche, 2021; Chu et al., 2009). Not to mention that when moving from surgical planning to in-surgery usages, it requires an entire specialized operating room, therefore we postulate it’s more effective to explore HMDs until the above technology matures more and costs are lowered.
Stereoscopic monitors remain a practical alternative in many operating rooms. They’re easier to deploy than projection setups, can be made glasses-free with eye-tracking and polarization technology, and allow clinicians to maintain sterility (Kang et al., 2022). Studies suggest that stereoscopic displays can improve anatomical understanding, aid novice surgeons in spatial tasks, and enhance diagnostic performance (Held and Hui, 2011; van Beurden et al., 2012). Recent innovations in autostereoscopic displays that integrate eye tracking show promise; for instance, Kang et al. (2022) developed a glasses-free 3D autostereoscopic system using eye tracking for cardiac CT imaging, which expert readers identified key arterial structures faster and without discomfort. Still, these monitors lack the immersive, hands-free 360 spatial interaction of HMDs. While outside the scope of this study, future studies or meta-analyses should compare these approaches when evaluating cost, usability, and above all, surgical outcomes.
While semi-immersive setups deliver valuable stereoscopic perception, they do not provide the fully immersive, high-fidelity interaction achieved by contemporary head-mounted displays. The AVP represents a leap in HMD capabilities (e.g., resolution, interaction fidelity) compared to predecessors like the HoloLens 2 and Meta Quest 3 (Egger et al., 2024). When paired with CR, which enhances traditional volume rendering through photorealistic lighting, this hardware promises clinically actionable 3D visualization. However, prior CR evaluations lacked usability assessment in clinical contexts, particularly on cutting-edge platforms like the AVP.
The applications of cinematic rendering techniques go beyond direct visualization. For example, it can generate photorealistic synthetic data used to train deep learning algorithms for medical image interpretation (Mahmood et al., 2018) or be used in various clinical and educational scenarios (Eid et al., 2017; Li et al., 2019). Cinematic rendering has therefore become an enticing alternative to volume rendering (Dappa et al., 2016). More user-centered studies have applied CR for patient education (Pachowsky et al., 2022) and student learning (Binder et al., 2019; Binder et al., 2021; Elshafei et al., 2019). These studies reported improved patient understanding and student recall compared to traditional CT scans. However, these evaluations were conducted on conventional 2D monitors and did not directly assess software usability.
Combining advanced HMD hardware and cinematic rendering software has led to the development of interactive 3DVR. 3DVR allows medical scans to be displayed and interacted with in true 3D using an HMD (Douglas et al., 2017). Although 3DVR and likewise cinematic rendering technologies hold promise, particularly for complex patient-specific anatomy, the exact clinical benefit remains uncertain (Duran et al., 2019; Queisner and Eisenträger, 2024; Brookmeyer et al., 2024).
Siemens Healthineers previously explored this technology by adapting CR for the HoloLens. One study showed improved surgical decision-making compared to traditional 2D CT images or 3D printed models in cardiac surgery (Gehrsitz et al., 2021). Nevertheless, despite the HoloLens’ popularity in medical research, few clinicians used the commercial CR software (Gsaxner et al., 2023).
While open-source CR alternatives exist, they lack the refined transfer functions and realism of Siemens’ commercial solution (Baseer et al., 2023); clinicians prefer Siemens’ CR application over commercial alternatives. The performance gap likely arises from Siemens Healthineers’ dedicated team working extensively to perfect the transfer functions for each specific use case, which is a highly specialized and time-intensive task. This illustrates a trade-off between open-source and commercial systems: open-source tools allow rapid prototyping of workflow-specific features such as measurement or annotation, whereas Siemens’ CR prioritizes photorealism and refined transfer functions but is currently limited in functionality. This hurdle towards clinical integration and feature desires was consistently noted by expert participants in our study (Tables 1, 2).
Most existing studies on CR focused either on educational uses displayed on monitors or on synthetic data generation. Few studies have explored actual clinical scenarios to demonstrate tangible benefits. This leaves a significant gap in understanding usability and the specific features clinicians require for clinical adoption.
To address this gap, our study specifically examines the usability of the CR software on the AVP, gathering direct clinician feedback and ideas for clinical integration. We selected the latest hardware (AVP) and software (CR application) to ensure the best possible fidelity, realism, and user interaction–factors crucial for future clinical acceptance. This approach also provides insight into the future potential as these technologies become increasingly accessible.
4.2 Key contributions and strengths
This study is the first to evaluate the usability of Siemens’ CR application on the AVP within a clinical context. It provides detailed insights into usability (Figures 4, 6), highlights specific positive and negative aspects (Table 2), and presents potential clinical applications identified by medical experts (Table 3). Using publicly available imaging data instead of Siemens-specific datasets ensured better alignment with clinicians’ expertise and more realistic testing conditions. Additionally, the public datasets increase the reproducibility of this research.
Our results show that CR on the AVP has good-to-excellent usability, even compared with other digital health applications, with an average SUS score of 77.68 (Hyzy et al., 2022; Bevan et al., 2015; Brooke, 1996; Bangor et al., 2008). Usability ratings were positive across participants, ranging from medical students to experienced surgeons.
Medical students rated usability notably lower than medical professionals (Figure 4). While this difference may partly result from the small sample size, several factors likely contribute. The CR application uses features and icons similar to those found in standard radiological software, such as windowing, slice scrolling, zooming, plane selection, and visualization methods like maximum intensity projection (Figure 3). Surgeons, who frequently use these tools, likely felt more comfortable from the start. This interpretation is further supported by the lower ISONORM scores for self-descriptiveness (Figures 5, 6).
Medical professionals may also have built higher frustration tolerance and better learning strategies over time due to exposure to various technologies throughout their careers. Their practical experience likely helped them recognize clinical use cases more readily, possibly influencing their higher appreciation.
The ISONORM results (Figure 6) also suggest improvements are needed in learnability and error tolerance. Importantly, the study had no genuine technical failures (software or hardware malfunctions) observed during the live video stream. Rather, participants occasionally experienced unexpected interactions with the 3D anatomy, likely due to unfamiliarity with the specific hand-eye interactions, interface elements, or icons. Despite these minor issues, users rated the controllability, conformity, and suitability highly, reflecting the system’s consistent performance.
Lastly, customizability received comparatively lower scores. The qualitative feedback (Table 2) confirms that negative aspects mainly consisted of feature requests rather than fundamental problems. The positive feedback aligns well with the overall SUS and ISONORM scores. Thus, exploring requested features and incorporating them without negatively affecting usability, learnability, or self-descriptiveness will be an essential next step.
These findings directly reflect the primary aim of our study: not to demonstrate clinical outcomes, but to assess usability and collect expert feedback on which functionalities must be added for CR on the AVP to transition from promising usability toward practical workflow integration.
4.3 Limitations and challenges
Despite promising results, several limitations must be acknowledged. The current version of the CR application lacks key clinical features such as segmentation, measurement tools, annotations, and multi-modal scan fusion. These features are routinely used in surgical planning and diagnostics. Nevertheless, participating surgeons, who are often familiar with automated 3D reconstructions from surgical navigation software, rated the system’s usability highly. This indicates that the core interaction design is valuable even without these advanced features.
Hardware constraints also posed minor challenges and affect how easily the device can be shared. The AVP requires proprietary vision correction inserts rather than spectacles, which makes each unit effectively personalized. One participant withdrew because their glasses could not be accommodated; most other participants adapted comfortably using contact lenses. Despite two face-interface sizes, slight ambient light leakage was observed when the seal was not perfect, which may subtly influence perceived image quality, although no formal complaints were reported. Participants did not report problems with headset weight during the short 15–20 min sessions, even though the AVP headset itself weighs roughly 750–800 g and the external battery about 350 g, connected via a tethered cable that can restrict movement and compromise sterility if not routed carefully. Only one participant reported mild nausea. While HMD technology has improved substantially, cybersickness remains an important consideration, particularly during prolonged clinical use (Tian et al., 2022).
In practical terms, the AVP therefore behaves more like a personal device that is best suited to single-user preoperative planning or focused intraoperative use, unless streaming to an external display is used for observers. The high purchase cost is less problematic in complex, high-value surgical cases where even small gains in decision-making may be justified. Still, it is a more substantial barrier for broad deployment in education and routine teaching, where many simultaneous viewers would be needed. By contrast, the external battery and cable are more limiting in sterile intraoperative workflows than in educational or non-operative contexts, where restricted movement and cable routing are easier to accommodate.
The sample size of the study was relatively small, yet it included highly experienced surgeons, lending credibility to the findings. Although no direct comparison was made with traditional 2D displays or open-source alternatives, validated usability measures (SUS and ISONORM) allowed for usability assessments. In addition, all participants first viewed the CT dataset and then the MRI dataset on the AVP, a fixed order that may introduce learning effects; future studies will use a counterbalanced design with randomized modality order and explicit baseline comparators, for example, AVP-CR versus standard 2D workstation review in real clinical cases. Thus, while participants identified relevant potential clinical uses ranging from surgical planning to education, these remain speculative until confirmed through real-world clinical studies.
Finally, note that at present, access to the AVP build of Siemens’ CR used in this study requires research collaboration with the vendor. However, broader release is planned, so this is a temporary practical constraint on replication rather than a methodological limitation.
4.4 Integration into clinical practice
The AVP cannot easily be shared among multiple users because prescription glasses and face-fit are personalized for each device owner. This limitation presents a practical and financial hurdle, especially for educational scenarios such as patient education (Egger et al., 2024). Although multiple face-fits can be purchased or users can wear contact lenses, these solutions remain suboptimal for widespread clinical adoption. Ideally, Apple would develop a face-fit option that allows regular prescription glasses to be comfortably worn. Alternatively, the Siemens CR software or open-source alternatives could be adapted for other available or upcoming HMD models. Combining this adaptation with real-time streaming to an iPad would immediately enable effective patient education. Optimally, a multi-user option would also be integrated. In an ideal educational scenario, this would support multiple users simultaneously, both co-located and remotely located (Perz et al., 2024).
Cost considerations are likely less critical for surgical or treatment-related applications. In these scenarios, participants particularly valued the photorealistic rendering, intuitive eye-based navigation, and rapid setup, all beneficial for preoperative planning. Qualitative feedback further supports the view that immersive visualization clarifies anatomical variations and spatial relationships. Such clarity may enable more confident, personalized surgical decisions. Some surgeons even suggested intraoperative applications, like quickly reassessing puncture or biopsy entry points during procedures. While promising, targeted clinical studies are needed to verify this benefit. Such studies could feasibly be executed within relatively short timelines. Studies should focus on measurable clinical impact and integration, prioritizing interventions where the most impact on patient outcome is expected. Even small studies could eventually be combined into a larger meta-analysis.
However, to fully realize the potential of CR on the AVP for preoperative planning, additional features are needed. Examples include drawing lines to visualize entry points or perform precise measurements. Viewing multiple scans simultaneously, or fusing multi-modal scans, would ensure that all relevant imaging data is visible in one cohesive overview. Other desired functionalities include incorporating artificial intelligence-driven segmentations, adding annotations, and removing anatomical structures via hand gestures. In a future scenario, segmentation could even be pre-generated automatically, allowing voice commands to highlight or render specific anatomical structures transparent.
Integrating these advanced features, particularly combined with real-time streaming to an iPad and multi-user capabilities, would also greatly enhance multidisciplinary case discussions. Additionally, incorporating automatic overlays with error indications could make the system suitable for intraoperative guidance. Such functionality might also be highly beneficial for ultrasound training. For instance, clearly indicating ultrasound imaging planes and corresponding anatomical structures could significantly enhance learning and spatial comprehension.
A crucial final feature, requiring close cooperation between hospitals and software developers, is integration with hospital electronic patient record systems. Currently, creating and transferring visualization scenes to the AVP still involves manual steps. Automating these processes would significantly reduce staff workload and associated costs, an important factor considering the current healthcare labor shortages.
From a commercial perspective, especially in insurance-driven healthcare markets, clearly defining a clinical use case with sufficient benefits to justify these additional features is essential. For clinical researchers, commercially superior products like Siemens’ CR offer robust testing environments. However, open-source alternatives might become attractive due to their flexibility for rapidly developing research-specific functionalities (Baseer et al., 2023).
Ultimately, we anticipate that cinematic 3DVR will become a standard feature integrated directly into advanced medical imaging equipment, such as state-of-the-art CT and MRI systems.
4.5 Conclusion and future directions
This study suggests the potential of immersive cinematic 3D volume rendering using Siemens’ CR application on the AVP to enhance medical image interpretation. Participating surgeons and trainees gave high usability ratings, indicating that the current implementation is usable in this pilot setting and may support future clinical adoption if subsequent studies demonstrate effectiveness. The intuitive interface facilitates detailed, photorealistic visualization of CT and MRI data. It was perceived as particularly promising for educational purposes, surgical planning, and intraoperative visualization of complex anatomy and anatomical variations.
The ISONORM profile, which showed high suitability, consistency, and controllability but weaker self-descriptiveness and customizability, was consistent with the open-ended feedback. Clinicians repeatedly requested advanced segmentation capabilities, more robust measurement and annotation tools, multimodal fusion, streaming or multi-user options, voice control, and smoother integration with clinical data systems as prerequisites for routine clinical use. Addressing these feature requests is likely to further enhance both the usability and the perceived clinical relevance of this technology.
Future research should focus on longitudinal, real-world evaluations of CR-assisted clinical workflows. Controlled studies are needed to determine measurable benefits, such as improved surgical accuracy, reduced operative times, or enhanced educational outcomes. Moreover, technical enhancements should emphasize collaborative functionalities, including multi-user capabilities, annotations (lines and notes), remote consultations, and voice-controlled interactions. Implementing these features would likely accelerate clinical acceptance, as suggested by our participants’ feedback.
In summary, while the core usability of CR on the AVP was rated good to excellent, our participants identified key barriers to clinical adoption, including the absence of workflow-specific features, seamless integration with hospital systems, and operability in sterile environments for intraoperative usage.
Finally, while our study focused specifically on hepatobiliary surgery, future investigations should expand into other medical fields, such as orthopedics, cardiology, and interventional radiology. Broadening the scope of evaluation will be necessary to assess the generalizability and potential clinical impact of immersive 3D visualization across modern medicine.
5 Resource identification initiative
At the time of writing, Siemens’ Cinematic Reality and Cinematic Playground, as well as the Apple Vision Pro, are not registered with RRIDs (Research Resource Identifiers). As such, no RRIDs can be provided for the resources used in this study.
6 Life science identifiers
Life Science Identifiers are not applicable to this study, as it does not involve zoological nomenclature or related taxonomic acts.
Data availability statement
The datasets used in this study are publicly available in a publicly accessible repository. This data can be found here: CHAOS (https://doi.org/10.5281/zenodo.3431873, v1.03) and MRCP DLRecon (https://doi.org/10.5281/zenodo.13912092, v3). The questionnaires used in the user study and the scripts used to prepare the public imaging data into formats suitable for the Cinematic Reality application will be made available by the authors upon request, without undue reservation.
Ethics statement
The requirement of ethical approval was waived by Ethics committee of the medical faculty of the university of Duisburg -Essen and university hospital Essen for the studies involving humans because the study involved no significant physical or psychological risk to participants or patients. All participants were medical professionals or students who voluntarily interacted with a non-invasive visualization system (Apple Vision Pro headset).
Author contributions
GL: Conceptualization, Data curation, Formal Analysis, Investigation, Methodology, Project administration, Software, Validation, Visualization, Writing – original draft, Writing – review and editing. LFdP: Validation, Visualization, Writing – original draft, Writing – review and editing. SK: Methodology, Software, Writing – review and editing, Resources. AB: Methodology, Resources, Software, Supervision, Writing – review and editing. LM: Data curation, Investigation, Project administration, Writing – review and editing. AS: Validation, Visualization, Writing – review and editing. PH: Investigation, Methodology, Writing – review and editing. JK: Conceptualization, Methodology, Resources, Writing – review and editing. SS: Conceptualization, Investigation, Methodology, Resources, Supervision, Writing – review and editing. UN: Conceptualization, Methodology, Resources, Writing – review and editing. JE: Conceptualization, Methodology, Resources, Supervision, Writing – original draft, Writing – review and editing.
Funding
The author(s) declared that financial support was received for this work and/or its publication. This work was made possible through a collaboration with Siemens Healthineers, who provided early access to their Cinematic Reality Application on the Apple Vision Pro via TestFlight. We also benefited from access to the Siemens Cinematic Playground, which enabled us to visualize multiple publicly available scans (CHAOS dataset (Kavur et al., 2021; Kavur et al., 2019) and the MRCP_DLRecon dataset (Kim et al., 2025; Kim et al., 2024)) in 3D Cinematic Reality. We thank Siemens Healthineers for their support and commitment to this project. This research was partially supported by the REACT-EU project KITE (grant number EFRE-2920801977, Plattform für KI-Translation Essen, https://kite.ikim.nrw/). We further acknowledge funding from the European Union under Grant Agreement No. 101168715 (INSIDE: INSIGHT, https://inside-insight.eu/). The views and opinions expressed are those of the authors and do not necessarily reflect those of the European Union or the granting authority. Neither the European Union nor the granting authority can be held responsible for them. Finally, we acknowledge support by the Open Access Publication Fund of the University of Duisburg-Essen.
Acknowledgements
We thank the Department of General-, Visceral- and Transplant Surgery, Medical Center University Duisburg-Essen, Essen, Germany, for their support and for enabling this research within their clinical environment. We are grateful to the open-source community for enabling reproducible research in medical imaging. This study made use of publicly available datasets, including the CHAOS dataset (Kavur et al., 2021; Kavur et al., 2019) and the MRCP_DLRecon dataset (Kim et al., 2025; Kim et al., 2024). We also acknowledge Siemens Healthineers for providing early access to the Cinematic Playground and Cinematic Reality software. A preprint version of this manuscript is available on arXiv Luijten et al. (2025). Finally, we acknowledge support by the Open Access Publication Fund of the University of Duisburg-Essen.
Conflict of interest
Authors SK and AB were employed by Siemens Healthineers.
The remaining author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declared that generative AI was used in the creation of this manuscript. The authors disclose the intermittent use of generative AI tools between February and July 2025 to assist in refining grammar, phrasing, tone, and writing style. The following tools were used: DeepL Write (free version); DeepL SE, Cologne, Germany, https://www.deepl.com/write. Grammarly: AI Writing and Grammar Checker (paid version, Chrome extension v14.1242.0); Grammarly Inc., San Francisco, CA, United States, https://www.grammarly.com. ChatGPT-4 (via ChatGPT Plus subscription, model GPT-4-turbo); OpenAI, San Francisco, CA, United States, https://chat.openai.com. All content was critically reviewed and edited by the authors to ensure accuracy and adherence to scientific standards.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/frvir.2025.1666614/full#supplementary-material
References
Adnan, S., Benson, A. C., and Xiao, J. (2025). How virtual reality is being adopted in anatomy education in health sciences and allied health: a systematic review. Anat. Sci. Educ. 18, 496–525. doi:10.1002/ase.70027
Bangor, A., Kortum, P. T., and Miller, J. T. (2008). An empirical evaluation of the system usability scale. Intl. J. Human–Computer Interact. 24, 574–594. doi:10.1080/10447310802205776
Bangor, A., Kortum, P., and Miller, J. (2009). Determining what individual sus scores mean: adding an adjective rating scale. J. Usability Studies 4, 114–123. doi:10.5555/2835587.2835589
Baseer, I., Valverde, I., Moustafa, A. H., Blat, J., and Camara, O. (2023). “Evaluation of cinematic volume rendering open-source and commercial solutions for the exploration of congenital heart data,” in 2023 IEEE Visualization and Visual Analytics (VIS), 76–80. doi:10.1109/VIS54172.2023.00024
Bevan, N., Carter, J., and Harker, S. (2015). “ISO 9241-11 revised: what have we learnt about usability since 1998?,” in Human-Computer Interaction: Design and Evaluation. Editor M. Kurosu (Springer International Publishing), 143–151.
Binder, J., Krautz, C., Engel, K., Grützmann, R., Fellner, F. A., Burger, P. H. M., et al. (2019). Leveraging medical imaging for medical education – a cinematic rendering-featured lecture. Ann. Anat. - Anatomischer Anzeiger 222, 159–165. doi:10.1016/j.aanat.2018.12.004
Binder, J. S., Scholz, M., Ellmann, S., Uder, M., Grützmann, R., Weber, G. F., et al. (2021). Cinematic rendering in anatomy: a crossover study comparing a novel 3D reconstruction technique to conventional computed tomography. Anat. Sci. Educ. 14, 22–31. doi:10.1002/ase.1989
Blanche, P.-A. (2021). Holography, and the future of 3D display. Light Adv. Manuf. 2, 446–459. doi:10.37188/lam.2021.028
Brooke, J. (1986). System usability scale (sus) questionnaire. Digit. Equip. Corp. Available online at: https://rickvanderzwet.nl/trac/personal/export/104/liacs/hci/docs/SUS-questionaire.pdf(Accessed 25 June 2025)
Brooke, J. (1996). SUS: a ’Quick and Dirty’ usability scale. Usability Eval. Industry. doi:10.1201/9781498710411-35
Brookmeyer, C., Chu, L. C., Rowe, S. P., and Fishman, E. K. (2024). Clinical implementation of cinematic rendering. Curr. Problems Diagnostic Radiology 53, 313–328. doi:10.1067/j.cpradiol.2024.01.010
Chu, J. C. H., Gong, X., Cai, Y., Kirk, M. C., Zusag, T. W., Shott, S., et al. (2009). Application of holographic display in radiotherapy treatment planning II: a multi-institutional study. J. Appl. Clin. Med. Phys. 10, 115–124. doi:10.1120/jacmp.v10i3.2902._eprint
Cruz-Neira, C., Sandin, D. J., and DeFanti, T. A. (1993). “Surround-screen projection-based virtual reality: the design and implementation of the CAVE,” in Proceedings of the 20th annual conference on Computer graphics and interactive techniques (New York, NY, USA: Association for Computing Machinery), 135–142. doi:10.1145/166117.166134
Cruz-Neira, C., Reiners, D., and Springer, J. P. (2010). An affordable surround-screen virtual reality display. J. Soc. Inf. Disp. 18, 836–843. doi:10.1889/JSID18.10.836
Csebfalvi, B., and Szirmay-Kalos, L. (2003). Monte carlo volume rendering. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 449–456. doi:10.1109/VISUAL.2003.1250406
Dappa, E., Higashigaito, K., Fornaro, J., Leschka, S., Wildermuth, S., and Alkadhi, H. (2016). Cinematic rendering – an alternative to volume rendering for 3D computed tomography imaging. Insights Into Imaging 7, 849–856. doi:10.1007/s13244-016-0518-1
Douglas, D. B., Wilke, C. A., Gibson, J. D., Boone, J. M., and Wintermark, M. (2017). Augmented reality: advances in diagnostic imaging. Multimodal Technol. Interact. 1, 29. doi:10.3390/mti1040029
Duran, A. H., Duran, M. N., Masood, I., Maciolek, L. M., and Hussain, H. (2019). The additional diagnostic value of the three-dimensional volume rendering imaging in routine radiology practice. Cureus 11, e5579. doi:10.7759/cureus.5579
Egger, J., Gsaxner, C., Luijten, G., Chen, J., Chen, X., Bian, J., et al. (2024). Is the apple vision pro the ultimate display? A first perspective and survey on entering the wonderland of precision medicine. JMIR Serious Games 12, e52785. doi:10.2196/52785
Eid, M., De Cecco, C. N., Nance, J. W., Caruso, D., Albrecht, M. H., Spandorfer, A. J., et al. (2017). Cinematic rendering in CT: a novel, lifelike 3D visualization technique. Am. J. Roentgenol. 209, 370–379. doi:10.2214/AJR.17.17850
Elshafei, M., Binder, J., Baecker, J., Brunner, M., Uder, M., Weber, G. F., et al. (2019). Comparison of cinematic rendering and computed tomography for speed and comprehension of surgical anatomy. JAMA Surg. 154, 738–744. doi:10.1001/jamasurg.2019.1168
Erbay, N., Raptopoulos, V., Pomfret, E. A., Kamel, I. R., and Kruskal, J. B. (2003). Living donor liver transplantation in adults: vascular variants important in surgical planning for donors and recipients. Am. J. Roentgenol. 181, 109–114. doi:10.2214/ajr.181.1.1810109
Fellner, F. A. (2016). Introducing cinematic rendering: a novel technique for post-processing medical imaging data. J. Biomed. Sci. Eng. 9, 170–175. doi:10.4236/jbise.2016.93013
Fuchs, H., Levoy, M., and Pizer, S. M. (1989). Interactive visualization of 3d medical data. Computer 22, 46–51. doi:10.1109/2.35199
Fukumoto, W., Kitera, N., Mitani, H., Sueoka, T., Kondo, S., Kawashita, I., et al. (2022). Global illumination rendering versus volume rendering for the forensic evaluation of stab wounds using computed tomography. Sci. Rep. 12, 2452. doi:10.1038/s41598-022-06541-9
Gehrsitz, P., Rompel, O., Schöber, M., Cesnjevar, R., Purbojo, A., Uder, M., et al. (2021). Cinematic rendering in mixed-reality holograms: a new 3d preoperative planning tool in pediatric heart surgery. Front. Cardiovasc. Med. 8, 633611. doi:10.3389/fcvm.2021.633611
Gsaxner, C., Li, J., Pepe, A., Jin, Y., Kleesiek, J., Schmalstieg, D., et al. (2023). The HoloLens in medicine: a systematic review and taxonomy. Med. Image Anal. 102757, 102757. doi:10.1016/j.media.2023.102757
Harley, D. (2020). Palmer luckey and the rise of contemporary virtual reality. Convergence 26, 1144–1158. doi:10.1177/1354856519860237
Healthineers, S. (2017). Cinematic rendering in medical imaging. Available online at: https://www.siemens-healthineers.com/digital-health-solutions/cinematic-rendering(Accessed 25 June 2025)
Healthineers, S. (2024). Siemens healthineers launches cinematic reality app for apple vision pro. Available online at: https://www.siemens-healthineers.com/press/releases/cinematic-reality-applevisionpro(Accessed 25 June 2025)
Held, R. T., and Hui, T. T. (2011). A guide to stereoscopic 3D displays in medicine. Acad. Radiol. 18, 1035–1048. doi:10.1016/j.acra.2011.04.005
Hyzy, M., Bond, R., Mulvenna, M., Bai, L., Dix, A., Leigh, S., et al. (2022). System usability scale benchmarking for digital health apps: meta-Analysis. JMIR mHealth uHealth 10, e37290. doi:10.2196/37290
ISO (2020). ISO 9241-110:2020 Ergonomics of human–system interaction – part 110: interaction principles. Standard ISO 9241-110:2020. Geneva, Switzerland: International Organization for Standardization.
Kang, D., Choi, J.-H., and Hwang, H. (2022). Autostereoscopic 3D display system for 3D medical images. Appl. Sci. 12, 4288. doi:10.3390/app12094288
Kavur, A. E., Selver, M. A., Dicle, O., Barış, M., and Gezer, N. S. (2019). Chaos: combined (ct–mr) healthy abdominal organ segmentation challenge data. doi:10.5281/zenodo.3362844
Kavur, A. E., Gezer, N. S., Barş, M., Aslan, S., Conze, P.-H., Groza, V., et al. (2021). CHAOS challenge - combined (CT-MR) healthy abdominal organ segmentation. Med. Image Anal. 69, 101950. doi:10.1016/j.media.2020.101950
Kelly, P., Fung, A., Qu, J., Greig, P., Tait, G., Jenkinson, J., et al. (2017). Depicting surgical anatomy of the porta hepatis in living donor liver transplantation. J. Visualized Surgery 3, 43. doi:10.21037/jovs.2017.03.08
Kim, J., Nickel, D., and Knoll, F. (2024). Deep learning-based accelerated MR cholangiopancreatography without fully-sampled data. NMR Biomed. 38, e70002. doi:10.5281/zenodo.13912092
Kim, J., Nickel, M. D., and Knoll, F. (2025). Deep learning-based accelerated mr cholangiopancreatography without fully-sampled data. NMR Biomed. 38, e70002. doi:10.1002/nbm.70002
Kroes, T., Post, F. H., and Botha, C. P. (2012). Exposure render: an interactive photo-realistic volume rendering framework. PloS One 7, e38586. doi:10.1371/journal.pone.0038586
Kukla, P., Maciejewska, K., Strojna, I., Zapał, M., Zwierzchowski, G., and Bak, B. (2023). Extended reality in diagnostic imaging–a literature review. Tomography 9, 1071–1082. doi:10.3390/tomography9030088
Lewis, J. R. (2018). The system usability scale: Past, present, and future. Int. J. Human–Computer Interact. 34, 577–590. doi:10.1080/10447318.2018.1455307
Li, K., Yan, R., Ma, H., Zhang, D.-f., Ding, Y., and Li, Z.-h. (2019). Value of the cinematic rendering from volumetric computed tomography data in evaluating the relationship between deep soft tissue sarcomas of the extremities and adjacent major vessels: a preliminary study. J. Comput. Assisted Tomogr. 43, 386–391. doi:10.1097/RCT.0000000000000852
Ljung, P., Krüger, J., Groller, E., Hadwiger, M., Hansen, C. D., and Ynnerman, A. (2016). “State of the art in transfer functions for direct volume rendering,” in Computer graphics forum, 669–691.
Lopes, D. S., Parreira, P. F., Mendes, A. R., Pires, V. M., Paulo, S. F., Sousa, C., et al. (2018). Explicit design of transfer functions for volume-rendered images by combining histograms, thumbnails, and sketch-based interaction. Vis. Comput. 34, 1713–1723. doi:10.1007/s00371-017-1448-8
Luijten, G., Paiva, L. F., Krueger, S., Brost, A., Mazilescu, L., Santos, A. S. F., et al. (2025). From screen to space: evaluating siemens’ cinematic reality. doi:10.48550/arXiv.2506.04972
Mahmood, F., Chen, R., Sudarsky, S., Yu, D., and Durr, N. J. (2018). Deep learning with cinematic rendering: fine-tuning deep neural networks using photorealistic medical images. Phys. Med. & Biol. 63, 185012. doi:10.1088/1361-6560/aada93
Pachowsky, M. L., Morf, H., Simon, D., Schönau, V., Valor-Mendez, L., Knitza, J., et al. (2022). Cinematic rendering in rheumatic diseases–photorealistic depiction of pathologies improves disease understanding for patients. Front. Med. 9, 946106. doi:10.3389/fmed.2022.946106
Perz, M., Luijten, G., Kleesiek, J., Schmalstieg, D., Egger, J., and Gsaxner, C. (2024). MultiAR: a multi-user augmented reality platform for biomedical education. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. 2024, 1–7. doi:10.1109/EMBC53108.2024.10782948
Prümper, J. (1993). “Software-evaluation based upon ISO 9241 part 10,” in Human computer interaction. Editors T. Grechenig, and M. Tscheligi (Berlin, Heidelberg: Springer Berlin Heidelberg), 255–265.
Prümper, J. (2006). ISONORM 9241/110-S: Kurzfragebogen zur software-evaluation. Presentation at the 45th DGfP Annual Meeting, Nuremberg, September. Available online at: http://www.uselab.tu-berlin.de/wiki/images/6/62/ISONorm_Kurzversion.pdf(Accessed 25 June 2025)
Queisner, M., and Eisenträger, K. (2024). Surgical planning in virtual reality: a systematic review. J. Med. Imaging 11, 062603. doi:10.1117/1.JMI.11.6.062603
Ropinski, T., Döring, C., and Rezk-Salama, C. (2010). “Interactive volumetric lighting simulating scattering and shadowing,” in 2010 ieee Pacific visualization symposium (pacificvis), 169–176.
Sutherland, I. E. (1968). “A head-mounted three dimensional display,” in Proceedings of the December 9-11, 1968, fall joint computer conference, part I (New York, NY, USA: Association for Computing Machinery), 757–764. AFIPS ’68 (Fall, part I). doi:10.1145/1476589.1476686
Tian, N., Lopes, P., and Boulic, R. (2022). A review of cybersickness in head-mounted displays: raising attention to individual susceptibility. Virtual Real. 26, 1409–1441. doi:10.1007/s10055-022-00638-2
van Beurden, M. H. P. H., Ijsselsteijn, W. A., and Juola, J. F. (2012). Effectiveness of stereoscopic displays in medicine: a review. 3D Res. 3, 3. doi:10.1007/3DRes.01(2012)3
Vlachogianni, P., and Tselios, N. (2022). Perceived usability evaluation of educational technology using the system usability scale (sus): a systematic review. J. Res. Technol. Educ. 54, 392–409. doi:10.1080/15391523.2020.1867938
Yeo, C. T., MacDonald, A., Ungi, T., Lasso, A., Jalink, D., Zevin, B., et al. (2018). Utility of 3d reconstruction of 2d liver computed tomography/magnetic resonance images as a surgical planning tool for residents in liver resection surgery. J. Surgical Education 75, 792–797. doi:10.1016/j.jsurg.2017.07.031
Keywords: extended reality, augmented reality, apple vision pro, 3D medical imaging, volume rendering, clinical integration, usability assessment, cinematic rendering
Citation: Luijten G, Faray de Paiva L, Krueger S, Brost A, Mazilescu L, Santos AS, Hoyer P, Kleesiek J, Schmitz SM-T, Neumann UP and Egger J (2026) From screen to space: evaluating Siemens’ Cinematic Reality application for medical imaging on the Apple Vision Pro. Front. Virtual Real. 6:1666614. doi: 10.3389/frvir.2025.1666614
Received: 15 July 2025; Accepted: 28 November 2025;
Published: 05 January 2026.
Edited by:
Caitlin R. Rawlins, Department of Veterans Affairs, United StatesReviewed by:
Eleftherios Garyfallidis, Indiana University Bloomington, United StatesShreeraj Jadhav, Kitware, United States
Copyright © 2026 Luijten, Faray de Paiva, Krueger, Brost, Mazilescu, Santos, Hoyer, Kleesiek, Schmitz, Neumann and Egger. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Gijs Luijten, R2lqcy5MdWlqdGVuQHVrLWVzc2VuLmRl; Jan Egger, SmFuLkVnZ2VyQHVrLWVzc2VuLmRl
Lisle Faray de Paiva1