ORIGINAL RESEARCH article

Front. Digit. Health, 14 January 2025

Sec. Health Technology Implementation

Volume 6 - 2024 | https://doi.org/10.3389/fdgth.2024.1500677

Targeting accuracy of neuronavigation: a comparative evaluation of an innovative wearable AR platform vs. traditional EM navigation

  • 1. Department of Information Engineering, University of Pisa, Pisa, Italy

  • 2. EndoCAS Interdipartimental Center, University of Pisa, Pisa, Italy

  • 3. Department of Neurosurgery, Azienda Ospedaliero Universitaria Pisana, Pisa, Italy

  • 4. Department of Translational Research and New Technologies in Medicine and Surgery, University of Pisa, Pisa, Italy

Article metrics

View details

14

Citations

4,1k

Views

839

Downloads

Abstract

Wearable augmented reality in neurosurgery offers significant advantages by enabling the visualization of navigation information directly on the patient, seamlessly integrating virtual data with the real surgical field. This ergonomic approach can facilitate a more intuitive understanding of spatial relationships and guidance cues, potentially reducing cognitive load and enhancing the accuracy of surgical gestures by aligning critical information with the actual anatomy in real-time. This study evaluates the benefits of a novel AR platform, VOSTARS, by comparing its targeting accuracy to that of the gold-standard electromagnetic (EM) navigation system, Medtronic StealthStation® S7®. Both systems were evaluated in phantom and human studies. In the phantom study, participants targeted 13 predefined landmarks using identical pointers to isolate system performance. In the human study, three facial landmarks were targeted in nine volunteers post-brain tumor surgery. The performance of the VOSTARS system was superior to that of the standard neuronavigator in both the phantom and human studies. In the phantom study, users achieved a median accuracy of 1.4 mm (IQR: 1.2 mm) with VOSTARS compared to 2.9 mm (IQR: 1.4 mm) with the standard neuronavigator. In the human study, the median targeting accuracy with VOSTARS was significantly better for selected landmarks in the outer eyebrow (3.7 mm vs. 6.6 mm, p = 0.05) and forehead (4.5 mm vs. 6.3 mm, p = 0.021). Although the difference for the pronasal point was not statistically significant (2.7 mm vs. 3.5 mm, p = 0.123), the trend towards improved accuracy with VOSTARS is clear. These findings suggest that the proposed AR technology has the potential to significantly improve surgical outcomes in neurosurgery.

1 Introduction

Neurosurgery is a complex and delicate field that requires high precision and accuracy. Surgeons now have access to various computer-assisted surgery (CAS) technologies that enable surgical planning based on preoperative imaging and, during surgery, provide intraoperative instrument guidance. Navigation technologies work by spatially linking the patient and surgical tools to the image data through “patient-to-image” registration. Neuronavigators have been widely used to guide brain biopsy (1), transphenoidal surgery for the resection of pituitary adenomas (2), and brain tumor resection (3).

According to the literature, neuronavigation offers several advantages to surgeons, such as precise planning of the incision and craniotomy, and the identification of small subcortical lesions (4). In addition to anatomical data, functional magnetic resonance imaging (5) and tractography information (6) are possible as overlays available during surgery.

Clinical demonstrations of the neuronavigation benefit include the study from Wirtz et al. (7) focusing on glioblastoma surgery. The study compared the impact of neuronavigation on time consumption, the extent of tumor removal, and survival. It showed that absolute and relative residual tumor volumes are significantly lower with neuronavigation, and patients operated with neuronavigation had longer survival (median 13.4 vs. 11.1 months). Additionally, neuronavigation increased the extent of tumor removal in glioblastoma resection without prolonging operating time.

Two main modalities of navigator tracking are currently adopted, namely optical (OTS) and electromagnetic tracking systems (EMTS). Both OTS and EMTS techniques have proven their value in neurosurgical navigators where they are equally popular. The main limitation of the OTS is the requirement for a direct line of sight between the camera, the patient reference frame (used for continuous tracking of anatomy during registration and navigation), and the probe during navigation. If there is a “line-of-sight occlusion,” the system cannot track and navigate. EMTS overcomes this limitation; indeed, no line of sight is needed between the transmitter (electromagnetic field generator) and the receiver (sensor coil integrated in the surgical tool). In addition, electromagnetic navigation probes have a more compact design, making them easier to use under a microscope. However, a disadvantage of EMTS is the need to position the electromagnetic field generator near the surgical target. Furthermore, the accuracy of EMTS systems can be affected by the proximity of ferromagnetic instruments; however, current evidence suggests that this issue rarely arises in clinical settings (3, 8).

In addition to the technical accuracy related to the tracking technologies employed, the accuracy of the neuronavigation can be affected by various sources of errors. Wang et al. (9) identified two groups of errors based on the neuronavigator working principle. Error Type I includes errors caused by differences between the anatomical structures in the images and the actual patient, such as brain deformation and low image resolution. Type II error involves errors in the transformation of the position of surgical tools from the patient space to the image space, including tracking errors and surgical toll calibration inaccuracy, as well as image-to-patient registration errors. In the past, neuronavigation relied on direct point matches of bone-implanted fiducials, which provided high accuracy but caused discomfort. Modern systems use paired point registration with adhesive markers and surface matching algorithms, which, although less accurate than registration based on implanted cranial markers, have proven to be suitable for daily use in most neurosurgical cases with a reported accuracy between 1.8 mm and 5 mm (10, 11).

Conventional navigators are also limited in terms of visualization ergonomics: they display the guidance information on an external screen, which means that surgeons often shift their attention between the surgical field and the monitor. Augmented reality (AR) visual interfaces provide a solution by presenting the surgeon with virtual information seamlessly blended with the patient’s anatomy (e.g., this can include displaying a planned cutting line to guide an osteotomy superimposed on the patient’s bony anatomy). This approach reduces cognitive load and improves information management in image-guided surgery. Head-mounted displays (HMDs) provide an ergonomic interface that maintains the user’s egocentric view of the surgical field. For this reason, they are considered the most ergonomic and effective medium for guiding procedures that are performed manually under the surgeon’s direct vision, such as procedures involving manual manipulation of exposed human tissue, including incisions in epithelial, muscle, or bone structures.

In recent years, burgeoning research interests have been devoted to developing AR-based neuronavigators (1214), for example for surgical resection of intracranial meningiomas (15), providing insight into the disruptive potential of AR in neurosurgery. However, most studies on HMD are “proofs of concept” trials based on using a Microsoft HoloLens (16), a self-contained Optical See Through (OST) headset, outside its indication and despite the technological and human-factor limits that prevent achieving high accuracy levels. To list the most relevant: the perceptual conflicts between the view of the real world and the VR image (17), the small field of view (FoV), the sub-optimal ergonomics, and calibration issues to attain a robust VR-to-real alignment (1820).

To the best of the authors’ knowledge, today, there is no neurosurgery-specific HMD designed to comply with medical device regulation that has been validated in a relevant environment for guiding high-precision tasks. To address this gap, we aim to test VOSTARS, a novel hybrid video and optical see-through HMD designed for precision surgery applications, in the field of neurosurgical oncology for targeting supratentorial tumors, both intraparenchymal and extra-axial tumors, including plaque cranial vault meningiomas. Developed within the Horizon 2020 project framework, VOSTARS has already demonstrated promise in guiding maxillofacial osteotomies (21). We further validated its navigation performance through recent in-vitro studies using patient-specific phantoms (22, 23). These studies yielded impressive results, with a mean real-to-virtual 3D target visualization error (TVE3D) of just 1.3 mm and a standard deviation of 0.6 mm. Additionally, user studies showed that subjects guided by VOSTARS could trace a remarkable 97% of a planned craniotomy trajectory within a 1.5 mm error margin, with an outstanding 92% achieving a 1 mm margin. These results were obtained using a skin-fixed dynamic reference frame (DRF) for real-time registration.

In this study, we aim to evaluate the targeting accuracy achievable with the VOSTARS system in combination with the aforementioned DRF, and compare its performance to that of a traditional commercial navigation system. The most widely used neuronavigation systems on the market are produced by leading companies such as Stryker, Brainlab, and Medtronic, with the latter two being particularly notable for their strong clinical presence (24). A literature review suggests that the navigational accuracy of Medtronic’s StealthStation, BrainLab’s VectorVision, and Stryker’s iNtellect systems is generally comparable (25). For this study, we have selected the Medtronic StealthStation® S7® as the benchmark, as it is available in the neurosurgery unit participating in the trial.

Tests were performed first on a patient-specific phantom, then we proceeded with a noninvasive study on nine volunteers. Trials are focused on the targeting of superficial landmarks because the ultimate goal is to demonstrate the applicability of the VOSTARS system in guiding complex craniotomy procedures. Recent literature studies suggest that planning and executing an appropriately positioned and sized craniotomy, is the central role of a neuronavigational system in neurosurgery (3).

2 Material and methods

2.1 Navigation systems

The StealthStation® S7® System (Medtronic Inc., Louisville, CO, USA) is a popular commercial navigator (Figure 1A) that provides real-time surgical guidance by combining radiological images of the patient with real-time surgical tool tracking using optical or electromagnetic technology. The application software allows loading patient-specific CT or MR images acquired before surgery, or fluoroscopic images captured during surgery and displays them on the screen from various perspectives (such as axial, sagittal, coronal, and oblique). The surgeon can plan and save one or more surgical trajectories in the preoperative phase. Additionally, the surgeon can create and manipulate one or more 3D anatomical models to aid visualization. During surgery, the system continuously updates the position of the instruments on these radiological images by tracking the position of specialized surgical instruments in or on the patient’s anatomy using optical or electromagnetic tracking.

Figure 1

In 2013 (26), the technical accuracy of the StealthStation® S7® was assessed in a hospital setting using an ad-hoc designed phantom containing 51 target points. The OTS and EMTS measurement performances were estimated in a volume of 120 mm 120 mm 100 mm, roughly mimicking the size of the human head. The accuracy is calculated by evaluating the error in the distance of each target point to a reference point, both acquired with a navigated tooltip. The protocol does not require performing a registration procedure and repurges measured errors from human performance related to interface ergonomics and intrinsic user accuracy. Results showed that the technical accuracies of OTS and EMTS over the pre-determined volume are nearly equal: 0.20 mm 0.10 mm and 0.30 mm 0.13 mm, respectively.

Depending on the assessed device and methods, other studies evaluating the OTS and the EMTS technical accuracies have reported values up to 1.4 mm (26).

In our work, we utilized the EMTS functions of the StealthStation® S7® System. Indeed, this modality does not require rigid immobilization of the surgical area (e.g., using skull clamps) making it suitable for non-invasive testing on human volunteers in a non-surgical environment. This is allowed by the dynamic referencing function of the navigator: a proprietary patient reference frame (Figure 1C) is attached to the patient for the real-time update of the image-to-patient registration, thus ensuring accurate navigation.

The VOSTARS AR headset (Figure 1D) was created by modifying a commercial OST visor (ARS.30® by Trivisio). This headset can provide optical and video see-through augmentations using two liquid-crystal (LC) optical shutters (FOS model by LC-Tec®) placed on top of the semi-transparent optical combiners of the visor. By adjusting the drive voltage, users can switch between a regular optical see-through (OST) view (with shutters open) and a video see-through (VST) camera-assisted view (with shutters closed). The ARS.30® visor is equipped with dual SXGA OLED panels with resolution, a 30° diagonal field of view, and a 3 cm eye relief. The OST display has an average angular resolution of approximately 1.11 arcmin/pixel, which is comparable to human visual acuity. The visor’s collimation optics were redesigned to have a focal length of about 40 cm. Additionally, the two optical engines of the visor are slightly toed-in, meaning that the optical axes of the two displays converge at approximately the focal length of the collimation optics. These features are crucial for reducing issues like vergence-accommodation conflict and focus rivalry (17) when the headset is used in the peripersonal space.

The visor is made up of a 3D-printed plastic frame that incorporates two LC shutters and houses a pair of world-facing RGB cameras (two USB 3.0 LI-OV4689 cameras by Leopard Imaging, both equipped with 1/3” OmniVision CMOS 4M pixels sensor (pixel size: m) and an M12 lens with 6 mm focal length). These cameras are used for inside-out tracking and to provide the VST view. The cameras have a horizontal field of view of approximately , corresponding to an average angular resolution of about 2.2 arcmin/pixel. The stereo camera pair is mounted on the top of the visor with an anthropometric interaxial distance of about m to minimize the effect of camera-to-eye parallax. This setup achieves a quasi-orthostereoscopic perception of the scene under VST view.

For neurosurgical applications, we designed a custom-made DRF (Figures 2A–D) to facilitate registration during procedures. This frame utilizes colored fiducial markers embedded within a 3D-printed patient-specific template. The biocompatible and sterilizable material used for printing, like MED610 from Stratasys, ensures patient safety. Pre-operative MRI scans guide the creation of this template, ensuring a customized fit for each patient’s face. Three strategically placed spherical markers, each 12 mm in diameter, serve as tracking markers during the registration process. The template’s design offers an intuitive registration experience due to its snug fit and the clear positioning of the markers on the face. The template’s shape and placement are designed based on facial anatomical landmarks that experience minimal deformation due to the thin underlying soft tissue layer, as detailed in (22).

Figure 2

2.2 Patient-specific phantom

In vitro experiments were performed on a patient-specific 3D-printed mannequin (Figure 1B). The 3D model of the mannequin was generated from a real MRI dataset (an axial spoiled gradient recalled acquisition in the steady-state (SPGR) sequence with a 0.5 mm 0.5 mm 0.6 mm resolution), segmented with a semi-automatic pipeline (27) to extract the head surface. 13 holes (5 points in the frontal region, 4 in the temporal region, 2 in the orbital region, and 2 in the nasal region) 1 mm in diameter were designed on the phantom surface to be used as targets for accuracy evaluation (Figure 3). Two DRFs were designed for the phantom and used during the navigation trials for aiding the VOSTARS navigation (Figures 2A–D). A 3D printer (Dimension Elite®) was used to turn the virtual model into a tangible replica made of acrylonitrile butadiene styrene.

Figure 3

Figure 4

2.3 Volunteers for human study

Nine volunteers were recruited among patients who underwent surgery through a standard craniotomy for brain tumor (grade I meningioma according to W.H.O. classification) exeresis at the Pisana University Hospital (Pisa, Italy). Inclusion criteria comprised a recent (less than 6 months) postoperative MRI performed, with an axial spoiled gradient recalled acquisition in the steady-state (SPGR) sequence with a 0.5 mm 0.5 mm 0.6 mm resolution. All recruited subjects were self-sufficient adults, able to provide informed consent, with no signs of recurrence on MRI. Table 1 reports the data of recruited subjects. Recruited volunteers’ anthropometric data for head and face dimensions include 4 out of the 5 (80%) “face type” classified in (28) that cover all the anthropometric characteristics of human beings. The distribution is shown in Figure 5.

Table 1

General infoValues
Gender (number of male; number of female; number of non-binary)(0; 9; 0)
Age (min; max; mean; STD)(39; 76; 58.7; 12.4)
Time in months since last MRI acquisition (min; max; mean; STD)(2; 6; 4.8; 1.9)

Demographics of volunteers recruited for the human-study.

Figure 5

2.4 Subjects recruited to test the navigators

Four subjects aged between 28 and 42 with normal visual acuity or corrected-to-normal visual acuity (with the aid of contact lenses) were recruited. Table 2 reports the participants’ demographics) to perform navigation trials with the VOSTARS system and the StealthStation® S7. The subjects were four biomedical engineers skilled in surgical navigation systems development and testing. They all signed a dedicated informed consent containing general information about the commercial navigation system used and the VOSTARS AR platform as well as the aim of the trial.

Table 2

General infoValues
Gender (male; female; non-binary)(0; 4; 0)
Age (min; max; mean; STD)(28; 42; 35.8; 6.8)
Visual Acuity (normal; corrected to normal)(4; 1)
Experience with VOSTARS in-vitro (none; limited; familiar; experienced)(0; 0; 1; 3)
Experience with StealthStation in-vitro (none; limited; familiar; experienced)(0; 0; 2; 2)
Experience with EM navigation in-vitro (none; limited; familiar; experienced)(0; 0; 2; 2)

Demographics of users recruited for testing the navigation systems in the phantom-study.

2.5 Phantom experiment protocol

Subjects were randomly assigned to use the VOSTARS system or the StealthStation® first and were instructed to target the 13 phantom landmarks (target holes) using the two navigators. In both cases, subjects were instructed to use the pointer of the StealthStation® to avoid introducing distortions related to the dimensional characteristics of the targeting tool. The pointer features a spherical tip with a 1 mm size, matching the target holes.

The protocol for both navigation methods is reported below and shown in

Figure 4

.

  • Step 1: Cover the phantom landmarks with adhesive tape to hide their position (Figure 6A).

  • Step 2: Perform registration.

  • Step 3: Check the accuracy of the registration and repeat step 2 until the registration is successfully performed.

  • Step 4: Instruct the user to dip the pointer tip into liquid dye, target each landmark as shown by the navigator, and make a colored mark at each landmark.

  • Step 5: Repeat Step 4 for each subject.

  • Step 6: Use a sharp instrument to uncover the actual position of each landmark.

  • Step 7: An experimenter, blind to the navigation method, uses a caliper to measure the distance of each mark from the actual landmark. For marks completely inside the holes (not visible on the adhesive tape after Step 5), record an accuracy of 0.5 mm (Figure 6B).

Figure 6

During Step 2, in the trials using StealthStation® navigation, the Medtronic patient reference frame was attached to the phantom’s forehead. Initial registration was achieved using surface-matching, which aligns a cloud of digitized points on the patient’s scalp with a volume rendering of the imaging data (29). This registration method is popular in neuronavigation due to its advantages over the more accurate but less convenient procedure based on point-pair matching with adhesive markers. These advantages include ease of use, not requiring additional imaging with markers in place and cost-effectiveness. In VOSTARS navigation trials, the DRF was positioned on the phantom’s face and held in place by elastic bands. The registration was obtained implicitly, as the DRF fits the phantom’s face, providing pose registration (22).

As for Step 3, the registration accuracy estimated by the StealthStation® navigation software was utilized. A 1 mm error was selected as a threshold to consider the registration accurate enough for the trials. Additionally, a registration “sanity-check” procedure was performed (30): the experimenter “touched” some easily identifiable anatomical landmarks with the tracked probe, and the relative positioning of the probe and the anatomical model was checked on the navigator screen. The landmarks used were the canthi of the eyes and the pronasal point at the anterior apex of the nose. During VOSTARS trials, a sanity check procedure was employed to verify the proper placement of the DRF. AR spheres were designed at the canthi of the eyes, allowing the experimenter to visually estimate the template placement from different viewpoints. The positioning of the template was corrected until the AR spheres appeared perfectly aligned with the corresponding anatomical landmarks.

2.6 Human study protocol

The human study was focused on evaluating the accuracy of a single user in targeting 3 landmarks on the face of the 10 recruited volunteers. The pronasal point (PN), a point on the outer eyebrow (OE), and a point on the forehead (FH) were used as navigation targets (Figure 1C). Each recruited subject was provided with a custom mask with holes at the three target points to pinpoint their exact locations based on their selection on the MRI dataset. The custom mask, as the DRF, was designed starting from the segmentation of the MRI images to easily fit the patient’s face, in a unique and stable position. Compared with DFR, the masks are characterized by a larger fitting area. The starting navigation modality was randomly chosen for each patient, and the StealthStation® pointer was used during both navigation trials.

The protocol of the study is reported below and depicted in

Figure 7

.

  • Step 1: Start with the first navigation modality and perform the registration procedure.

  • Step 2: Check the accuracy of the registration and repeat step 1 until the registration is successfully performed.

  • Step 3: Instruct the user to dip the pointer tip into liquid green dye, target each landmark as shown by the navigator, and make a mark at each landmark.

  • Step 4: Switch to the second navigation mode and perform the registration procedure.

  • Step 5: Check the accuracy of the registration and repeat step 4 until the registration is successfully performed.

  • Step 6: Instruct the user to dip the pointer tip into red liquid dye, target each landmark as shown by the navigator, and make a mark at each landmark.

  • Step 7: Fit the patient-specific mask on the volunteer’s face.

  • Step 8: Use the mask holes as a guide to mark the volunteer’s face with blue, indicating the planned position of the three targets.

  • Step 9: An experimenter, unaware of the randomization order, uses a caliper to measure the distance of each green and red mark from the corresponding blue mark.

Figure 7

Registration and sanity check procedures were the same as those used in the phantom tests.

2.7 Statistical analysis

The GNU PSPP 2.0.1 software was used to perform statistical analysis of data. The results of the targeting accuracy estimation were summarized in terms of median and interquartile range (IQR). A Wilcoxon signed-rank test was used to assess whether there was a significant difference in the users’ targeting accuracy based on the navigator used. A p-value less than or equal to 0.05

3 Results

3.1 Phantom study results

Table 3 shows the performance results of four subjects who tested the VOSTARS system and the StealthStation® S7® for targeting 13 different landmarks on the phantom. These landmarks include five points in the frontal region (FR), four in the temporal region (TR), two in the orbital region (OR), and two in the nasal region (NR) (Figure 3).

Table 3

LandmarkVostarsStealthStation®
User 1User 2User 3User 4MedianIQRUser 1User 2User 3User 4MedianIQR
FR 11.41.41.71.61.50.22.13.02.52.42.50.3
FR 21.41.82.71.91.90.43.13.13.13.03.10.0
FR 30.50.50.50.50.50.02.13.31.82.32.20.5
FR 40.50.50.50.50.50.02.02.03.12.52.30.7
FR 50.50.51.50.71.60.62.02.42.42.33.90.1
TR 10.52.51.51.61.60.63.94.23.93.83.80.8
TR 20.51.62.71.62.21.22.93.64.74.04.40.0
TR 31.91.74.52.52.20.84.44.44.44.23.00.1
TR 41.22.03.62.31.10.32.93.03.02.82.20.3
NR 10.51.11.51.00.80.71.42.22.32.11.60.2
NR 20.50.51.61.01.01.41.71.71.41.52.00.3
OR 10.50.53.31.41.50.71.72.12.11.93.60.3
OR 20.51.72.11.30.60.43.23.54.13.62.40.2

Results of phantom-study: comparison of Vostars and StealthStation® across various landmarks and users.

Overall, users achieved 1.4 mm median accuracy (IQR:1.2 mm) with VOSTARS and 2.9 mm (IQR:1.4 mm) with the standard neuronavigator. The results of the Wilcoxon signed-rank test indicate a statistically significant difference in the performance of all users according to the navigation modality (User1 p = 0.001, User2 p = 0.001, User3 p = 0.033, User4 p = 0.001).

3.2 Human study results

Table 4 shows the performance results of a single subject involved in testing the VOSTARS system and the StealthStation® S7® for targeting three facial landmarks (pronasal point (PN), a point on the outer eyebrow (OE), and a point on the forehead (FH)) in nine volunteers.

Table 4

Subject #Pronasal pointOuter eyebrowForehead
VostarsStealthStation®VostarsStealthStation®VostarsStealthStation®
15.810.16.26.79.110.8
23.03.54.816.64.55.8
31.61.32.61.43.30.3
44.84.02.54.42.84.4
50.72.42.35.71.21.7
65.810.12.76.68.710.4
72.72.43.87.62.06.3
82.04.43.74.07.010.6
92.02.04.08.58.815.2
Median2.73.53.76.64.56.3
IQR2.82.01.43.25.96.2

Results of human-study: comparison of Vostars and StealthStation® for various facial points.

The median targeting accuracy using the VOSTARS system was better than that obtained with the StealthStation® for all three facial landmarks (2.7 mm vs. 3.5 for PN; 3.7 mm vs. 6.6 for OE; 4.5 mm vs. 6.3 for OE). The Wilcoxon signed-rank test results revealed a statistically significant difference in targeting accuracy only for the OE and FH landmarks (PN p = 0.123; OE p = 0.05; FH p = 0.021).

4 Discussion

This work focuses on a comparative evaluation of an innovative HMD AR navigation system vs. traditional EM navigation for neurosurgery.

Literature studies on AR-HMD for neuronavigation mainly concentrate on the use of the Microsoft HoloLens, a general-purpose headset, not specifically designed for surgery. A recent clinical trial in neuro-oncology with this device showed that it can enable the surgeon to understand the relationship of the pathology with the surrounding structures (31). Table 5 summarizes the accuracy results of previous literature studies on wearable AR neuronavigation, focalized on cranial procedures. The experimental setups and error metrics reported in the different studies vary significantly, making direct comparisons with our results challenging. Nevertheless, it is worth highlighting that in prior studies involving the Microsoft HoloLens, the average display error consistently exceeded 4 mm for both fiducial markers (Fiducial Registration Error, FRE) and targets (Target Registration Error, TVE). Additionally, the TDE, which is further influenced by the user’s skill level, has similarly not demonstrated lower values in those studies.

Table 5

StudyAR technologyStudy typologyMethod of accuracy evaluationReported accuracy
Maruyama et al. (32)Epson Moverio (BT-200)Phantom studyTRE over four target points at the border of a tumorThe mean and standard deviation were 2.1 and 1.1 mm respectively
Incekara et al. (33)Microsoft HoloLens®Patient studyTDE measured using a BrainLab neuronavigator as a gold standardThe overall median deviation between the two modalities was 4 mm with an interquartile range 0–0.8 mm
van Doormaal et al. (34)Microsoft HoloLens®Phantom and patient studyFRE calculated as the root-mean-square of the distance between skin fiducialsPhantom study: mean and standard deviation were 7.2 and 1.8 mm respectively. Patients: Mean and standard deviation were 4.4 and 2.5 mm respectively
McJunkin et al. (35)Microsoft HoloLens®Cadaver study and patient studyTRE over 7 pre-specified landmarksPhantom study: the mean and standard deviation were 5.76 and 0.54 mm respectively
Li et al. (36)Microsoft HoloLens®Patient studyPostoperative CT scan used to measure the TDE in guiding the external ventricular drainMean and standard deviation of 4.34 and 1.63 mm
Fick et al. (37)Microsoft HoloLens®Patient studyFRE over 6 registrations8.5 mm
Qi et al. (38)Microsoft HoloLens®Patient studyTDE measured using a BrainLab neuronavigator as a gold standardThe overall median deviation between the two modalities was 4.1 mm (IQR 3.0 mm–4.7 mm)

FRE, Fiducial registration error as root mean square distance between real fiducial positions and the associated virtual homologous; TVE2D, Target visualization error estimated as reprojection error (commonly in px) in 2D onto the displayed AR image; TDE, Target deviation error as the Euclidean distance between the planned and real (achieved) targets; TRE, Target registration error as Euclidean distance between the real-physical target point and its virtual counterpart (in mm); TVE3D, Target visualization error in 3D between the real-physical target point and its virtual counterpart (in mm).

In this work, we compared the VOSTARS AR HMD to the StealthStation® S7 EM navigation for accuracy in targeting superficial landmarks. Unlike most of the studies mentioned before, the performance of the AR HMD is not reported in terms of deviation from the traditional navigator measurements. The accuracy of both was evaluated by measuring the distance between the point targeted by the users and the real target with a caliper.

According to the study result, the VOSTARS platform yields a better performance both in vitro and in vivo.

The users involved in the phantom study performed significantly better with the VOSTARS system (median accuracy of 1.4 mm with VOSTARS and 2.9 mm with the StealthStation® S7). Results obtained with VOSTARS align with our previous findings concerning estimating the real-to-virtual 3D target visualization error in a similar setup: the TVE3D mean and standard deviation were 1.3 and 0.6 mm, respectively.

Results obtained from in-vivo experiments showed a lower targeting accuracy with both guidance systems. This was to be expected since, for example, the patient’s anatomy at the time of testing may differ from the face model extracted from the MRI data set [see Type I error according to (9)]. In our specific protocol, this error may be higher than in traditional neuronavigation flow, as our study included healed patients who had MRIs up to six months before our experiments. Any changes that occurred between the time of the MRI data acquisition and the experiments could have reduced the accuracy of both navigation systems’ patient-to-image registration (as for the VOSTARS system, this could have determined modification to the fitting of the DRF on the patient’s facial anatomy).

Moreover, a limitation of the human study, is that only a a single user was recruited to perform the accuracy tests, thus reducing the number of tests performed to minimize patient discomfort. The test may also be biased due to its non-invasive nature. During the test, recruited volunteers were awake and their heads were not secured with a head clamp or holder. While subjects were instructed to remain as still as possible, any small movements that occurred during the test could have caused slight changes in the positioning of the reference frame of the StealthStation®/ the VOSTARS DRF on the subject face.

Additionally, both the targeting trial and the error measurements were more challenging in the in-vivo setup compared to the in-vitro setup. This is because both the user performing the navigation trial and the user measuring the errors needed to minimize contact with the volunteer’s face (with their hands and the pointer/caliper) to avoid discomfort.

In light of these factors, caution is advised when interpreting the absolute error values. The values obtained cannot be used to predict the accuracy of the VOSTARS system in a real surgical scenario. However, since all the errors mentioned affected both navigation systems in our protocol, the results obtained are useful for comparing the performance of the two systems.

The median targeting accuracy with the VOSTARS system was significantly better than that obtained with the StealthStation® for selected landmarks at the level of the outer eyebrow (3.7 mm vs. 6.6, p = 0.05) and forehead (4.5 mm vs. 6.3, p = 0.021), while a non-significant difference was found for the pronasal point (2.7 mm vs. 3.5 mm, p = 0.123). Our findings suggest that the lower error in identifying this landmark might be attributed to its ease of anatomical localization. Conversely, the forehead landmark yielded the highest error, with the StealthStation® tests reaching a maximum of 15.2 mm. This significant discrepancy is likely due, in part, to the inherent mobility of the skin on the forehead during targeting.

5 Conclusion and future scope

This study represents a critical preliminary step in validating the potential of the VOSTARS augmented reality (AR) system for neurosurgical navigation, towards its clinical use for guiding complex craniotomies. By comparing the VOSTARS platform to the established Medtronic StealthStation® S7® in both phantom and human studies, we have demonstrated the superior targeting accuracy of the AR system in superficial landmark localization. The phantom trials showed significantly higher accuracy with VOSTARS, confirmed in human volunteers.

Despite the positive results, this work is only an initial validation of the system’s capabilities. The in-vivo human study was limited in scope, with only one operator and no invasive surgical interventions. Furthermore, the tests were conducted under conditions more challenging than those outlined in the protocol planned for the future in vivo surgical trial. Specifically, the subject was awake and free to move, introducing significant variability and reducing accuracy compared to a controlled operative setting. Furthermore, the radiological images used for augmented reality data extraction could be up to six months old, potentially affecting registration accuracy due to anatomical or physiological changes over time.

In conclusion, despite the limitations of the study, the results obtained encourage further development of the VOSTARS platform. Thanks to this study’s results, in the coming months, we will expand this research with a comprehensive clinical trial designed to evaluate the in-vivo performance of VOSTARS during live craniotomies. This trial will assess the system’s accuracy in real-time surgical environments, focusing on complex craniotomies. We aim to explore its effectiveness in improving surgical precision, reducing operating time, and ultimately enhancing patient outcomes.

Statements

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by “Comitato Etico di Area Vasta Nord Ovest 316 (CEAVNO)” (CEAVNO). The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

MC: Conceptualization, Formal Analysis, Investigation, Methodology, Writing – original draft, Writing – review & editing. NM: Data curation, Formal Analysis, Validation, Writing – original draft, Writing – review & editing. NC: Data curation, Investigation, Validation, Writing – review & editing. MA: Investigation, Validation, Writing – review & editing. FC: Investigation, Methodology, Validation, Writing – review & editing. VF: Conceptualization, Resources, Supervision, Writing – review & editing. EC: Supervision, Writing – review & editing. SC: Conceptualization, Funding acquisition, Investigation, Methodology, Supervision, Validation, Writing – original draft, Writing – review & editing.

Funding

The authors declare financial support was received for the research, authorship, and/or publication of this article. This work was supported by: the Italian Ministry of Health, under the call for proposal “Finalized Research” 2021, project “Clinical testing of augmented reality navigation for brain tumor surgery,” Project Code GR-2021-12373198; the PNRR national plan, Spoke 9 of the Tuscany Health Ecosystem (THE).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1.

    VeiceschiPLocatelliDDarioAAgrestaG. Frameless neuronavigation-assisted brain biopsy with electromagnetic tracking: how I do it?Acta Neurochir. (2022) 164:331722. 10.1007/s00701-022-05252-4

  • 2.

    ThomaleU-WStoverJFUnterbergAW. The use of neuronavigation in transnasal transsphenoidal pituitary surgery. Zentralbl Neurochir. (2005) 66:12632. 10.1055/s-2005-836602

  • 3.

    OrringerDAGolbyAJoleszF. Neuronavigation in the surgical management of brain tumors: current and future trends. Expert Rev Med Devices. (2012) 9:491500. 10.1586/erd.12.42

  • 4.

    ChandraRVKingJAJ. Advanced imaging of brain tumors. In: KayeAHLawsER, editors. Brain Tumors. Edinburgh: Elsevier (2012). p. 188213.

  • 5.

    BraunVDempfSTomczakRWunderlichAWellerRRichterH-P. Multimodal cranial neuronavigation: direct integration of functional magnetic resonance imaging and positron emission tomography data: technical note. Neurosurgery. (2001) 48:117882. 10.1097/00006123-200105000-00050

  • 6.

    HlatkyRJacksonEFWeinbergJSMcCutcheonIE. Intraoperative neuronavigation using diffusion tensor MR tractography for the resection of a deep tumor adjacent to the corticospinal tract. Stereotact Funct Neurosurg. (2005) 83:22832. 10.1159/000091954

  • 7.

    WirtzCRAlbertFKSchwadererMHeuerCStaubertATronnierVMet al. The benefit of neuronavigation for neurosurgery analyzed by its impact on glioblastoma surgery. Neurol Res. (2000) 22:35460. 10.1080/01616412.2000.11740684

  • 8.

    MascottCR. Comparison of magnetic tracking and optical tracking by simultaneous use of two independent frameless stereotactic systems. Oper Neurosurg. (2005) 57:ONS-295301. 10.1227/01.NEU.0000176411.55324.1E

  • 9.

    WangMNSongZJ. Classification and analysis of the errors in neuronavigation. Neurosurgery. (2011) 68:113143. 10.1227/neu.0b013e318209cc45

  • 10.

    MascottCRSolJ-CBousquetPLagarrigueJLazorthesYLauwers-CancesV. Quantification of true in vivo (application) accuracy in cranial image-guided surgery: influence of mode of patient registration. Oper Neurosurg. (2006) 59:ONS-14656. 10.1227/01.neu.0000220089.39533.4e

  • 11.

    PinggeraDKerschbaumerJBauerMRiedmannMConradMBrennerEet al. Accuracy validation of neuronavigation comparing headholder-based system with head-mounted array—a cadaveric study. World Neurosurg. (2018) 120:e3137. 10.1016/j.wneu.2018.08.059

  • 12.

    CampisiBMCostanzoRGulinoVAvalloneCNotoMBonosiLet al. The role of augmented reality neuronavigation in transsphenoidal surgery: a systematic review. Brain Sci. (2023) 13:1695. 10.3390/brainsci13121695

  • 13.

    FickTvan DoormaalJAHovingEWWillemsPWvan DoormaalTP. Current accuracy of augmented reality neuronavigation systems: systematic review and meta-analysis. World Neurosurg. (2021) 146:17988. 10.1016/j.wneu.2020.11.029

  • 14.

    HeyGGuyotMCarterALucke-WoldB. Augmented reality in neurosurgery: a new paradigm for training. Medicina. (2023) 59:1721. 10.3390/medicina59101721

  • 15.

    Gómez AmarilloDFOrdóñez-RubianoEGRamírez-SanabriaADFigueredoLFVargas-OsorioMPRamonJFet al. Augmented reality for intracranial meningioma resection: a mini-review. Front Neurol. (2023) 14:1269014. 10.3389/fneur.2023.1269014

  • 16.

    FrantzTJansenBDuerinckJVandemeulebrouckeJ. Augmenting microsoft’s hololens with vuforia tracking for neuronavigation. Healthc Technol Lett. (2018) 5:2215. 10.1049/htl.2018.5079

  • 17.

    CondinoSCarboneMPiazzaRFerrariMFerrariV. Perceptual limits of optical see-through visors for augmented reality guidance of manual tasks. IEEE Trans Biomed Eng. (2020) 67:4119. 10.1109/TBME.2019.2914517

  • 18.

    ChenLZhaoSChenWShiZXieLYanYet al. Trajectory-based alignment for optical see-through HMD calibration. Multimed Tools Appl. (2024) 83:7192550. 10.1007/s11042-024-18252-6

  • 19.

    HuXBaenaFRYCutoloF. Alignment-free offline calibration of commercial optical see-through head-mounted displays with simplified procedures. IEEE Access. (2020) 8:22366174. 10.1109/ACCESS.2020.3044184

  • 20.

    TangAZhouJOwenC. Evaluation of calibration procedures for optical see-through head-mounted displays. In: The Second IEEE and ACM International Symposium on Mixed and Augmented Reality, 2003. Proceedings. IEEE Computer Society (2003). ISMAR-03. 10.1109/ismar.2003.1240699

  • 21.

    CarboneMCutoloFCondinoSCercenelliLD’AmatoRBadialiGet al. Architecture of a hybrid video/optical see-through head-mounted display-based augmented reality surgical navigation platform. Information. (2022) 13:81. 10.3390/info13020081

  • 22.

    CondinoSMontemurroNCattariND’AmatoRThomaleUFerrariVet al. Evaluation of a wearable AR platform for guiding complex craniotomies in neurosurgery. Ann Biomed Eng. (2021) 49:2590605. 10.1007/s10439-021-02834-8

  • 23.

    MontemurroNCondinoSCattariND’AmatoRFerrariVCutoloF. Augmented reality-assisted craniotomy for parasagittal and convexity en plaque meningiomas and custom-made cranio-plasty: a preliminary laboratory report. Int J Environ Res Public Health. (2021) 18:9955. 10.3390/ijerph18199955

  • 24.

    ResearchEM. Data from: Neuronavigation systems market: global industry trends, share, size, growth, opportunity and forecast 2023–2028 (2024). (accessed November 25, 2024).

  • 25.

    ChartrainAGKellnerCPFargenKMSpiottaAMCheslerDAFiorellaDet al. A review and comparison of three neuronavigation systems for minimally invasive intracerebral hemorrhage evacuation. J Neurointerv Surg. (2018) 10:6674. 10.1136/neurintsurg-2017-013091

  • 26.

    KoivukangasTKatiskoJPKoivukangasJP. Technical accuracy of optical and the electromagnetic tracking systems. Springerplus. (2013) 2:90. 10.1186/2193-1801-2-90

  • 27.

    FerrariVCarboneMCappelliCBoniLMelfiFFerrariMet al. Value of multidetector computed tomography image segmentation for preoperative planning in general surgery. Surg Endosc. (2011) 26:61626. 10.1007/s00464-011-1920-x

  • 28.

    ZhuangZBensonSViscusiD. Digital 3D headforms with facial features representative of the current us workforce. Ergonomics. (2010) 53:66171. 10.1080/00140130903581656

  • 29.

    MongenMAWillemsPWA. Current accuracy of surface matching compared to adhesive markers in patient-to-image registration. Acta Neurochir. (2019) 161:86570. 10.1007/s00701-019-03867-8

  • 30.

    CondinoSCutoloFCarboneMCercenelliLBadialiGMontemurroNet al. Registration sanity check for ar-guided surgical interventions: experience from head and face surgery. IEEE J Transl Eng Health Med. (2024) 12:25867. 10.1109/jtehm.2023.3332088

  • 31.

    JainSGaoYYeoTTNgiamKY. Use of mixed reality in neuro-oncology: a single centre experience. Life. (2023) 13(2):398. 10.3390/life13020398

  • 32.

    MaruyamaKSuyamaYNakamuraH. Accuracy of augmented reality-assisted percutaneous spine surgery. Oper Neurosurg. (2017) 16:11725. 10.1093/ons/opx279

  • 33.

    IncekaraFSmitsMDirvenCVincentA. Clinical feasibility of a wearable mixed-reality device in neurosurgery. World Neurosurg. (2018) 118:e4227. 10.1016/j.wneu.2018.06.208

  • 34.

    van DoormaalTPCvan DoormaalJAMMensinkT. Clinical accuracy of holographic navigation using point-based registration on augmented-reality glasses. Oper Neurosurg. (2019) 17:58893. 10.1093/ons/opz094

  • 35.

    McJunkinJLJiramongkolchaiPChungWSouthworthMDurakovicNBuchmanCAet al. Development of a mixed reality platform for lateral skull base anatomy. Otol Neurotol. (2018) 39:e113742. 10.1097/MAO.0000000000001995

  • 36.

    LiYChenXWangNZhangWLiDZhangLet al. A wearable mixed-reality holographic computer for guiding external ventricular drain insertion at the bedside. J Neurosurg. (2018) 131:1599606. 10.3171/2018.4.JNS18124

  • 37.

    FickTvan DoormaalJAMHovingEWRegliLvan DoormaalTPC. Holographic patient tracking after bed movement for augmented reality neuronavigation using a head-mounted display. Acta Neurochir. (2021) 163:87984. 10.1007/s00701-021-04707-4

  • 38.

    QiZLiYXuXZhangJLiFGanZet al. Holographic mixed-reality neuronavigation with a head-mounted device: technical feasibility and clinical application. Neurosurg Focus. (2021) 51:E22. 10.3171/2021.5.focus21175

Summary

Keywords

neuronavigation, augmented reality navigation, targeting accuracy, head mounted displays, neurosurgery, surgical navigation

Citation

Carbone M, Montemurro N, Cattari N, Autelitano M, Cutolo F, Ferrari V, Cigna E and Condino S (2025) Targeting accuracy of neuronavigation: a comparative evaluation of an innovative wearable AR platform vs. traditional EM navigation. Front. Digit. Health 6:1500677. doi: 10.3389/fdgth.2024.1500677

Received

23 September 2024

Accepted

30 December 2024

Published

14 January 2025

Volume

6 - 2024

Edited by

Roshan Joy Martis, Manipal Institute of Technology Bengaluru, India

Reviewed by

Kausik Basak, JIS Institute of Advanced Studies and Research, India

Chandrakala HT, Government of Karnataka, India

Updates

Copyright

*Correspondence: Nadia Cattari

†These authors share first authorship

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics