Skip to main content

ORIGINAL RESEARCH article

Front. Bioinform., 31 January 2022
Sec. Data Visualization
Volume 1 - 2021 | https://doi.org/10.3389/fbinf.2021.777101

New Approach to Accelerated Image Annotation by Leveraging Virtual Reality and Cloud Computing

www.frontiersin.orgCorentin Guérinot1,2,3 www.frontiersin.orgValentin Marcon1 www.frontiersin.orgCharlotte Godard1,4 www.frontiersin.orgThomas Blanc3,5 www.frontiersin.orgHippolyte Verdier1,6,7 www.frontiersin.orgGuillaume Planchon1 www.frontiersin.orgFrancesca Raimondi1,8,9,10 www.frontiersin.orgNathalie Boddaert9,10 www.frontiersin.orgMariana Alonso2 www.frontiersin.orgKurt Sailor2 www.frontiersin.orgPierre-Marie Lledo2 www.frontiersin.orgBassam Hajj3,4 www.frontiersin.orgMohamed El Beheiry1* www.frontiersin.orgJean-Baptiste Masson1*
  • 1Decision and Bayesian Computation, USR 3756 (C3BI/DBC) & Neuroscience Department CNRS UMR 3751, Université de Paris, Institut Pasteur, Paris, France
  • 2Perception and Memory Unit, CNRS UMR3571, Institut Pasteur, Paris, France
  • 3Sorbonne Université, Collège Doctoral, Paris, France
  • 4École Doctorale Physique en Île-de-France, PSL University, Paris, France
  • 5Laboratoire Physico-Chimie, Institut Curie, PSL Research University, CNRS UMR168, Paris, France
  • 6Histopathology and Bio-Imaging Group, Sanofi R&D, Vitry-Sur-Seine, France
  • 7Université de Paris, UFR de Physique, Paris, France
  • 8Unité Médicochirurgicale de Cardiologie Congénitale et Pédiatrique, Centre de Référence des Malformations Cardiaques Congénitales Complexes M3C, Hôpital Universitaire Necker-Enfants Malades, Université de Paris, Paris, France
  • 9Pediatric Radiology Unit, Hôpital Universitaire Necker-Enfants Malades, Université de Paris, Paris, France
  • 10UMR-1163 Institut Imagine, Hôpital Universitaire Necker-Enfants Malades, AP-HP, Paris, France

Three-dimensional imaging is at the core of medical imaging and is becoming a standard in biological research. As a result, there is an increasing need to visualize, analyze and interact with data in a natural three-dimensional context. By combining stereoscopy and motion tracking, commercial virtual reality (VR) headsets provide a solution to this critical visualization challenge by allowing users to view volumetric image stacks in a highly intuitive fashion. While optimizing the visualization and interaction process in VR remains an active topic, one of the most pressing issue is how to utilize VR for annotation and analysis of data. Annotating data is often a required step for training machine learning algorithms. For example, enhancing the ability to annotate complex three-dimensional data in biological research as newly acquired data may come in limited quantities. Similarly, medical data annotation is often time-consuming and requires expert knowledge to identify structures of interest correctly. Moreover, simultaneous data analysis and visualization in VR is computationally demanding. Here, we introduce a new procedure to visualize, interact, annotate and analyze data by combining VR with cloud computing. VR is leveraged to provide natural interactions with volumetric representations of experimental imaging data. In parallel, cloud computing performs costly computations to accelerate the data annotation with minimal input required from the user. We demonstrate multiple proof-of-concept applications of our approach on volumetric fluorescent microscopy images of mouse neurons and tumor or organ annotations in medical images.

1 Introduction

Continuous technological advances in optical and electron microscopy have enhanced our ability to discern three-dimensional (3D) biological structures via slice-based tomography (Zheng et al., 2018; Driscoll et al., 2019; Gao et al., 2019; Hörl et al., 2019; Hoffman et al., 2020). Entire structures from organelles to whole organisms can be imaged at the nanometric resolution, allowing the exploration of complex interplay between 3D geometry and biological activity (Gao et al., 2019). Furthermore, large-scale recordings capturing entire organisms provide a new means for understanding biology at multiple spatial and temporal scales. Three-dimensional medical imaging has been accessible for many years (typically at the millimetric resolution), primarily acquired from computed tomography (CT) scans, magnetic resonance imaging (MRI), and, more recently, numerically processed ultrasound recordings. Medical image analysis is based on the specialized exploration of the slices along the principal axes of recording, i.e., the sagittal, coronal, and axial planes. These last 10 years have seen numerous machine learning-based approaches to assist and automate medical image analysis (Esteva et al., 2021).

Gaining an intuitive understanding from these complex raw data remains a challenge. Due to noise and statistical variability in the recordings, biological researchers often encounter difficulties in probing the geometry of organelles. It is also challenging in the medical imaging domain, where surgeons and clinicians lacking radiology training have difficulties in mentally transforming information in 2D image slices into a 3D representation of an organ, tumor or region of interest. In addition, natural modes of 3D visualization are missing, as most analyses rely on viewing 3D data on a computer monitor while simultaneously using a mouse to interact and extract information from the data.

Virtual reality (VR) technology has recently reemerged, in part due to low-cost consumer headsets and increasingly powerful graphics cards. The efficient integration of stereoscopy, immersion, and motion tracking in VR allows the user to visualize 3D structures in a physically realistic computer-generated environment. Interactions in this artificial environment rely on handheld VR controllers that allow physically-based actions to be performed on virtual objects.

Numerous initiatives have focused on taking advantage of this technology in the domains of education and scientific research (Dede et al., 2017; Balo et al., 2017; O’Connor et al., 2018; Johnston et al., 2018; Matthews, 2018; El Beheiry et al., 2019; Safaryan and Mehta, 2021). Recent studies have additionally highlighted the benefits of immersive viewing for handling 3D data, including efficiency and enhanced intuition relative to standard monitor-based visualization (Johnston et al., 2018; El Beheiry et al., 2019). It is worth pointing out that multiple companies have focused their efforts on developing state-of-the-art processes for high-quality image rendering. Examples of these active initiatives are found in Arivis AG and syGlass (see Table 1 in El Beheiry et al., 2020). Within the context of medical applications, initiatives have also focused on education (Djukic et al., 2013; Fertleman et al., 2018; Bouaoud et al., 2020; Shao et al., 2020; Venkatesan et al., 2021), surgery planning and diagnosis (Reitinger et al., 2006; Ong et al., 2018; Pfeiffer et al., 2018; Ayoub and Pulijala, 2019; Lee and Wong, 2019; Pinter et al., 2020; Wake et al., 2020; Boedecker et al., 2021; Chheang et al., 2021; Laas et al., 2021; Lau et al., 2021; Raimondi et al., 2021; Ruiz et al., 2021; Venkatesan et al., 2021).

Experimental three-dimensional image recordings (e.g., microscopy and medical) are typically acquired in limited quantities (Matthews, 2018). Additionally, these few acquisitions are subject to variability which make for difficult streamlining of data analysis. To address this reality we require, first, the appropriate means to visualize, interact with, and manipulate data and, second, an ability to rapidly perform quantitative assessments on these data.

The first challenge can be tackled via visualization with VR. By rendering image stacks into a VR environment, users can easily navigate and interact with their 3D data. In turn, VR enables the user to grasp an intuitive understanding of the dataset being visualized. However, multiple issues are associated with this task: 1) finding proper ways to represent diverse image stacks originating from different imaging modalities with varying signal to noise ratios, 2) providing versatile tools to explore and interact with the VR representation, and 3) finding procedures that can handle large data sets.

The second challenge is addressed by employing human-in-the-loop (Patel et al., 2019) data treatment procedures. The idea here is to couple user interactions with data analysis for extracting relevant information from the data. In the context of this work, this implies 1) defining procedures to select data within the VR environment, 2) performing the required computations for analysis without significantly impacting the VR rendering performance, and 3) allowing corrections to be performed in an iterative fashion.

In the following sections, we discuss related works involving VR software for image stack visualization. We introduce our approach, DIVA Cloud, which allows visualization and interaction in VR combined with cloud computing. Finally, we show how this approach can be effectively utilized in data annotation for microscopy and medical images.

2 Quick Introduction to Related Works

Affordable VR headsets, efficient graphics cards and easily accessible software development platforms (OpenXR, Unity, Unreal Engine etc.) have widened access to VR developments. These factors have promoted initiatives combining imaging techniques with VR in order to address topics in cell biology. As a result, VR applications are often forecasted to become essential components of the experimental research environment (Matthews, 2018; El Beheiry et al., 2019).

Image stack visualization in the VR environment is at the center of numerous initiatives. These include ConfocalVR (Stefani et al., 2018) and Scenery (Gunther et al., 2019), which can generate volumetric reconstructions of microscopy images. Neuroscience is a domain with a great need to visualize and manipulate data in 3D. Large images of entire nervous systems can now be acquired with optical and electron microscopy (EM). Tracing complex neuronal structures is essential as there is a link between structure, connectivity and functions of neural circuits. Some initiatives already use VR to address this (Peng et al., 2010; Usher et al., 2017). Applications at the frontiers of microscopy and neurosurgery have also been demonstrated in literature (de Mauro et al., 2009; Wang et al., 2019; Wisotzky et al., 2019).

VR interactions from multiple users on the same data have been introduced in Naviscope (Shattuck, 2018). Additionally, some projects address topic-specific challenges in microscopy, such as colocalization (Theart et al., 2017).

Companies developing software for microscopy image analysis are now adding VR compatibility for visualization and treatment. Major advances are found with Arivis AG (Dekkers et al., 2019; Conrad et al., 2020) syGlass, which include an optimized data interaction interface. Other research applications focus on biomolecule structural information visualizations and interactions (Doutreligne et al., 2014; Balo et al., 2017; Goddard et al., 2018; Cassidy et al., 2020).

Not all microscopy image analysis software involves raw, full-stack image analyses but instead a deconvolved output (Betzig et al., 2006; Manzo and Garcia-Parajo, 2015; Qi et al., 2016). It is especially the case for single-molecule microscopy, where signals from individual biomolecules are captured and processed to deduce their nanometric positions and dynamic behavior. In these cases, the microscopy image stacks are reduced to point clouds. Two recent open-source software tools have been introduced to visualize and interact with single-molecule experiments: vLUME (Spark et al., 2020) and ours, Genuage (Blanc et al., 2020). Both software offers interfaces to interact with the point clouds and to perform various forms of data analysis (measuring, counting, cropping, etc.). Other initiatives on point clouds relate to data tagging for machine learning (Berge et al., 2016; Stets et al., 2018; Ramirez et al., 2019; Wirth et al., 2019; Liu et al., 2020). Mixed applications can be found involving astronomy, such as Gaia Sky (Sellés, 2013). General visualization and interaction software include PointCloud XR and developments centered on compression to ensure visualization in VR and Augmented Reality (AR) (Pavez et al., 2018).

In medicine, VR has found applications in surgery-specific topics, notably education. While VR may be useful for radiology (Uppot et al., 2019; Elsayed et al., 2020), radiologists are trained to perform 3D mental reconstructions of medical images, limiting their interest in immersive visualization modalities. Craniofacial trauma education (Bouaoud et al., 2020), neurosurgical training (Bernardo, 2017), spinal surgery (Pfandler et al., 2017), anatomy education (Uruthiralingam and Rea, 2020), orthopedic surgery (Bartlett et al., 2018; Walbron et al., 2019; Yoo et al., 2019; Lohre et al., 2020) and patient education (Dyer et al., 2018) have been demonstrated in this regard. Furthermore, clinical results hint towards uses of VR in surgical applications spanning heart diseases (Ayerbe et al., 2020; Sadeghi et al., 2020; Hattab et al., 2021; Raimondi et al., 2021), breast cancer (Tomikawa et al., 2010; Laas et al., 2021), liver surgery (Reitinger et al., 2006; Quero et al., 2019; Golse et al., 2020; Lang and Huber, 2020; Boedecker et al., 2021), pediatric surgery (Wang et al., 2012; Ruiz et al., 2021; Salvatore et al., 2021) and orthopaedic surgery (Bartlett et al., 2018; Yoo et al., 2019; Verhey et al., 2020). Multiple new companies are now investigating the potential of VR for surgical planning such as ImmersiveTouch®, PrecisionOS or SurgicalTheater.

3 Visualizing and Interacting With Image Stacks Without Pre-processing in VR

We recently introduced DIVA software (El Beheiry et al., 2020), a user-friendly platform that generates volumetric reconstructions from raw 3D microscopy image stacks and enables efficient visualization, analysis and quantification. The software is available at https://diva.pasteur.fr.

DIVA was developed using the Unity game engine (UnityTechnologies), and is based on what we term a lean mapper software architecture. Furthermore, the software uses the Windows-based SteamVR standard, making it compatible with most PC VR headsets, such as the HTC Vive and Oculus Rift S. DIVA renders image stacks and hyperstacks instantaneously as 3D volumes through via GPU-based volume ray-casting (Engel et al., 2004, 2017). DIVA offers a dual interface allowing the user to interact both on a standard computer monitor (i.e., desktop mode) and in an immersive artificial environment (i.e., VR mode). However, spending a significant amount of time in VR can lead to discomfort among many users. Therefore, the desktop mode allows users to set optimal visualization parameters before switching to the VR mode, which is dedicated to visually interpreting, analyzing, and navigating the data.

In the desktop interface, the user can modify scaling and lighting, voxel color and opacity in real-time through a user-friendly transfer function interface. This transfer function allows a simple association of color and opacity to visualized voxels based on their raw intensity values, as shown in Figure 1A. Configuration of the transfer function and interaction with the volume (rotation, translation, and scaling) can be controlled with the mouse. Transfer functions in DIVA can be saved and loaded in Javascript Object Notation (JSON) format.

FIGURE 1
www.frontiersin.org

FIGURE 1. The DIVA dual interface presented on an example of a light-sheet microscopy image of a human fetus lung with pulmonary alveoli (in red), trachea (in blue) and vascular system (in green) (Belle et al., 2017). (A) Desktop interface with raw data in the bottom right corner and transfer function interface in the top right corner with curves for voxel opacity (white arrow) and color (red arrow). (B) VR interface with VR controller in orange. (C) Clipping tool with the VR controller to navigate inside the volume. (D) Flashlight tool with the VR controller to highlight a spherical area of interest. (E) Counter tool with the VR controller to enumerate elements of interest.

In DIVA’s VR mode, the user visualizes image stacks rendered as live physical objects as a result of stereoscopy. Physical manipulation of the volume, such as grasping, rotation, or navigation, can be done with the VR controller, which acts as a 3D mouse for interaction (see Figure 1B). Hence, the data can be observed at any arbitrary angle to understand its structure in detail. To ease the navigation in complex and dense biological images, a clipping tool, presented in Figure 1C, can be activated. It dynamically removes a planar portion of the rendered volume allowing deep structures in the image to be revealed. Tagging, counting, and distance measurement tools are included for basic quantitative measurements (see Figures 1D,E). Users can extract all measurement results in a CSV file as well as through screen and movie captures.

As shown in Figures 2A–C, users can utilize DIVA for various microscopy modalities with up to four different channels associated with individual transfer functions. The user can reveal structures of interest by modulating the voxel transparency and colors in the transfer function. He can also discard undesired voxels without losing information, which can happen when using segmentation techniques. For example, users can easily remove the significant background noise of EM images. DIVA can also be used to compare raw and segmented data by merging the image stacks together as TIFF files with multiple channels.

FIGURE 2
www.frontiersin.org

FIGURE 2. DIVA application examples on the desktop interface with their corresponding raw image in the bottom left corner. (A–C) TIFF image stacks of (A) Mouse hippocampus imaged by two-photon serial endblock imaging (SEBI, Thy-1-GFP mouse) (Sun et al., 2013). (B) Mouse embryonic brain slices from spinning disk microscope (Brault et al., 2016). (C) Focused ion beam scanning EM of components of an adult mouse neuron: Golgi apparatus and mitochondria (Gao et al., 2019). (D–F) DICOM images of (D) Post CT-scan of craniofacial fractures (Bouaoud et al., 2020). (E) MRI of an adult heart with ventricular D-loop and septal defect (Raimondi et al., 2021). (F) CT-scan of lung with COVID-19 infection (Cohen et al., 2020).

With VR, the perception of 3D structures in complex data (e.g., EM images) is enhanced, and measurements are performed quicker than in standard 2D stack viewers. Examples of advantages in 3D perception of VR are found for histological sample examination (Lobachev et al., 2021), surgery simulation or planning (Seymour, 2008; Guerriero et al., 2018; Thomsen et al., 2017; Chen et al., 2020), motion or gaze precision (Martirosov et al., 2021; Pastel et al., 2021), and 3D data labeling prior to machine learning training (Ramirez et al., 2020). Medical experts and undergraduate students have reported better visualization of 3D anatomical structures in VR using DIVA, when compared with typical 3D renderings, see Table 1 in Bouaoud et al. (2020) and Raimondi et al. (2021).

Medical images are most often stored in the Digital Imaging and Communications in Medicine (DICOM) format and analyzed through a 2D interface (i.e., a DICOM Viewer). In Figures 2D–F, we show examples of medical images visualized within DIVA. As DIVA is data agnostic, the user can experience both medical and microscopy images.

4 Implementation

In this work, we seek to use the DIVA VR visualization context with 3D image analysis. We focused on the acceleration of image annotation with the objective of exploit new data without prior information or a pre-trained model. We aimed at reducing the burden of data tagging, which can require a large amount of user interaction. Our procedure consists of rapid 3D tagging in VR, simple classifier training and inference on the entire dataset with iterative corrections performed within VR. The complete procedure, Voxel Learning, is described schematically in Figure 3A and in the included Supplementary Video S1.

FIGURE 3
www.frontiersin.org

FIGURE 3. Voxel Learning and its application on a confocal image stack of mouse olfactory bulb interneurons. (A) Schematic of the analysis pipeline. After having set visualization parameters, the user performs the VR tagging and selects the model to be used. Training and inference steps are performed on the cloud, as indicated by the pictogram. (B) Data tagging step with the VR controller in orange. The positive and negative tags are colored in cyan and magenta, respectively. (C) Voxel Learning interface in DIVA with the output probabilities overlaid on the original image (0 corresponds to blue; 1 to red).

4.1 Annotation in VR and Feature Extraction

In this updated version of the DIVA software, we implemented a VR tagging functionality which allows voxel annotation i.e., associating an identifier (ID) to individual voxels. Tagged voxels appear in different colors, in a transparent wireframe mesh around the tagged voxel, allowing simultaneous visualization of the voxel and its tag. In the application shown in Figure 3B, the two colors are cyan and magenta as they highly visible in most situations. Tags can be updated or erased if necessary. The clipping plane tool (CPT) (El Beheiry et al., 2020) is also available to ensure more precision in ambiguous situations (see Figure 1C) and fluid tagging within the volume. Most importantly, the CPT allows annotating data at the frontier between different domains with geometries that do not align along the natural axis of data acquisition. Inside the VR environment, the properties of the interface and the transfer function are instrumental in accelerating the annotation process. We demonstrate the tagging procedure in Supplementary Video S2 on a medical example.

We additionally integrated the ability to calculate image features with the DIVA software. An efficient feature evaluation was implemented for each voxel, using a small subset of features (Arganda-Carreras et al., 2017). It includes a wide variety of spatial filters (Gaussian, median, mean, etc.) with different kernel sizes for gathering simple multi-scale features in the vicinity of the voxels. Features are then associated with a unique voxel ID, and the list of annotated voxels is stored. In the case of iterative tagging, the iteration number is also stored and associated with the voxel IDs.

4.2 Training and Inference

Our application here consists of accelerating data annotation using a simple one-shot learning procedure based on a few VR controller “strokes” performed by the user on the image stack. We follow the same principles as those used in ilastik (Berg et al., 2019), Weka (Arganda-Carreras et al., 2017) or behavior detection in larva (Masson et al., 2020), by tagging limited sets of data and stacking simple learners in order to train a collectively stronger classifier. Specifically, features are associated with tagged voxels, and learning is performed using robust classifiers in limited amounts of data. Furthermore, data tagging iterations allow the correction of anomalies in the learning to process the data.

We updated the DIVA desktop user interface to allow users to create quickly, load, and export classifiers. Basic classification approaches were used since they are known to provide robust classification on small datasets, such as Random Forest Classification (RFC), Multi-Layer Perceptron (MLP), Gradient Boosting (XGB), Support Vector Machine (SVM), and Naive Bayes (NBC) as implemented in the Scikit-learn (Pedregosa et al., 2011) Python package. In addition, hyperparameters were tuned to adapt to the problem being investigated and set to ensure rapid learning. Note that in this application, the usual problem of overfitting (Mostafa, 2012) is less present, as the goal is to annotate the data being explored and not to find a general learning scheme. Once the user has finished the annotation step, features associated to voxels are evaluated locally. They are then transferred in JSON format to the cloud, where a model is trained to classify all voxels in the 3D data stack. The models and their associated parameters are saved locally in a Pickle (Python) format. They can afterwards be loaded to perform inference on the entire dataset or, if found robust, provide initial annotations on new, previously unseen, data. The resulting inference is then broadcasted back from the cloud for local rendering in DIVA.

VR provides a significant advantage in the data annotation task by properly overlaying the classification result on the raw data in a volumetric representation. The representation of both the raw data and the annotated data provided by the classifiers helps correct errors and ensures proper annotation. We integrated a channel-based representation to DIVA in which raw data and classifier-generated data are associated with different channels allowing separate and fused views of both raw and classified data. A transfer function interface is associated for each channel. In most of our applications, the raw data was fused with the voxel probability (or log-probability) of belonging to a specific class. An example is shown in Figure 3C and in Supplementary Video S3.

Overall classifier robustness can be improved by using an ensemble of weak classifiers, whose resulting probabilities are added to the list of features before final model training. In addition, classifiers may also be iteratively updated with additional tagging rounds to correct for sub-optimal performance and false detection. Stacked learners are efficient in performing intuitive segmentation (Sommer et al., 2011). We denote here the gradient boost classifier with four weak learners as the strong learner.

4.3 DIVA Cloud

DIVA Cloud allows users to interact with data in VR and perform the analysis via Python scripts (see Section 6) whose calculations are performed on a remote web server. This development is motivated by the computational costs required for detailed VR volumetric rendering, leaving limited calculation bandwidth. Accordingly, to ensure fluid interactions and precise annotation dynamics, computationally costly operations are performed on the remote web server (i.e. cloud). We used Django, a Python Framework, as an application programming interface (API) provider for this project. Django features a system of data models and serializer links to a PostgreSQL database, enabling the management of jobs and file objects to track the life cycle of the learning jobs and various input and output files. This system is instrumental, as it permits performing a limited set of queries to interact with the cloud component.

The REST web service is used to provide an API with specific endpoints accessible from DIVA using HTTP methods. Celery, a task queue intermediate between the web server and the Python scripts, enables jobs to run asynchronously in a multi-threaded fashion. At the same time, the broker, Redis, allows communication with Celery.

We implemented DIVA Cloud within a Docker container to ensure portability on most platforms since it does not depend on the user installation. The layers implemented in the container consist of 1) the Django webserver, 2) the Celery task queue that integrates learning scripts (with their associated packages), 3) the Redis broker, and 4) the PostgreSQL database. Note that if the user runs DIVA Cloud on a powerful computation station, the entire pipeline can be executed locally on the same computer already running DIVA. A typical interaction workflow for the DIVA Cloud application is shown in Figure 4A in the context of data tagging. A visual representation of the relation between DIVA and DIVA Cloud is shown in Figure 4B.

FIGURE 4
www.frontiersin.org

FIGURE 4. (A) DIVA Cloud interaction workflow through a data tagging experiment to output a classifier that is visualized in DIVA. (B) Interaction between DIVA and DIVA Cloud in 6 steps: 1) POST request to the/jobs endpoint. It initializes a job entry in Django. Get in return the job ID 2) POST request to the/jobs/jobid/file endpoint with the inputs files. It creates a file entry and returns the file ID 3) PUT request to the/learning/jobid endpoint with the type of learning. It launches the job on input data 4) GET request to the/jobs/jobid/status endpoint to know the status of the job. If the status is “running”, the status of the job is requested (Step 4 again). If the status is “done”, continue. If the status is “error”, it is managed. 5) GET request to the/jobs/jobid/files/endpoint to get the output list 6) GET request to the/jobs/jobid/files/fileid to download the output.

5 Results

Our analysis was performed on a Windows 10 based ×64 system with an Intel i7-7700 CPU clocked at 3.60 GhZ, with 32 GB of RAM and an NVIDIA GeForce RTX 2080 Ti graphics card. An HTC Vive headset with its controller were used through SteamVR to perform the tagging procedure in VR. This particular VR headset has a total screen resolution of 2,160 × 1,200 pixels (1,080 × 1,200 for each eye). Analysis scripts were coded in Python 3.7. The DIVA Cloud configuration was also tested on a NVIDIA DGX-1 workstation as the remote server performing the computationally challenging tasks. The frame rate of DIVA Cloud is highly dependent on the hardware configuration, image size, interface (Desktop interface in 2D or VR interface), and the user’s movements in VR. Detailed information on the frame rate regarding this study can be found in Supplementary Table S3.

5.1 Metrics

We showed a proof of concept of this approach on various example image stacks, including a CT-scan, MRI sequence and various microscopy images applied to neuronal specimens.

We kept track of different time measurements: the tagging step in VR never exceeded 2 min. Tagging and subsequent model training are performed on a small portion of the data. Inference duration scaled with the size of the 3D image and depended on the computational bandwidth of the cloud infrastructure.

For medical images, we used raw and annotated images in order to compare the one-shot annotation to the expert full tagging. We computed Dice coefficient and Residual Mean Square Errors (RMSE) between our inferred probabilities and the given segmentation. The goal here was to evaluate how fast annotation in VR and quick simple learning can reduce the tagging of new data to a few complementary VR strokes. Performance depends on the nature of the features extracted from the dataset. As features were designed to cover a large variety of patterns and scales, our method can see use in many additional applications and data types. Furthermore, the number and nature of the features can be extended to capture specific properties of image stacks. All measurements are available in Supplementary Tables S1 and S2.

Figure 5 compares the Dice coefficient obtained with different models, along with the corresponding computing time when applied to medical examples images. The RFC and XGB, when stacked with 4 weak learners, reached highest performance with the Dice coefficient. RFC is associated to a shorter computation time, making it an appropriate candidate for efficient analysis.

FIGURE 5
www.frontiersin.org

FIGURE 5. Distribution of Dice coefficient (A) and computation time (B) when applying our annotation procedure to eight different medical examples images. Corresponding raw data is available in Supplementary Table S1.

5.2 Output Probabilities

In order to compare the results of RFC and the strong learner, we show in Figure 6 their application to MRI images showing a patient with breast cancer and a CT-scan of a patient with lung cancer (see Supplementary Figures S1, S2 for more examples). Figure 6B exhibits the tagging step. We decided to tag the lung tumor somewhat less completely than the breast tumor in order to assess the impact of tagging exhaustiveness. Regardless, our pipeline demonstrates qualitative results, identifying the structures of interest with precision using both models (see Figures 6C,D). We note the presence of low probability inferred voxels ranging from dark to light blue for RFC, whereas the strong learner seems to classify in a more binary fashion. In this respect, RFC proves to be more prone to false positive detection. The strong learner was more stringent and, as a result, may induce more false negative errors. We observe this tendency in Figures 6E,F, as RFC predictions exceed ground truth segmentation in volume while the strong learner predictions appeared smaller.

FIGURE 6
www.frontiersin.org

FIGURE 6. Annotation in DIVA on the breast MRI (left panel) and the lung CT-scan (right panel) and tumor (white arrow). (A) Raw data visualized in 3D on DIVA and as an image stack in the bottom right corner. (B) Overlay of the raw image in gray and tags with positive and negative tags in cyan and magenta, respectively. Tagging is performed in VR to quickly annotate which voxels belong to the structure of interest and which do not. (C,D) Overlay of the raw image in gray and output probabilities, respectively for the RFC and the strong learner. (E,F) Overlay of the raw image in gray, output probabilities, and ground truth segmentation in green for RFC and strong learner, respectively. Colorscale for probabilities is indicated between the two panels.

We also tested our accelerated data annotation procedure on microscopy images. Mouse olfactory bulb interneurons were imaged via confocal microscopy, results are shown in Figure 7. These data were considerably noisier than the CT-scan and MRI sequences that were previously utilized. The objective of this analysis was to reconstruct neuronal dendritic arbors. This was achieved by tagging two neuronal branches (see Figure 7B) and then applying our pipeline using the RFC and strong learner (see Figures 7C,D respectively). Almost every neurite and soma were classified using both learners. The RFC yielded high probabilities for inner structures and lower ones for outer structures, allowing isolating tubular structures through proper adaptation of the transfer function.

FIGURE 7
www.frontiersin.org

FIGURE 7. Annotation in DIVA on confocal microscopy images of mouse olfactory bulb interneurons (white arrow). (A) Raw data visualized in 3D on DIVA and as a z-stack in the bottom right corner. (B) Overlay of raw data in gray and tagging data with positive and negative tags respectively in cyan and magenta. (C,D) Overlay of raw data in gray and output probabilities respectively for RFC and strong learner. Colorscale for probabilities is indicated on the right of the image.

The procedure was assessed on a various range of examples: annotation of pancreas (Supplementary Figures S3, S4) and hepatic vessels (Supplementary Figures S5, S6) in CT-scans, annotation of mouse microglia in two-photon fluorescence microscopy (Supplementary Figure S7) and mouse hippocampal neurons in SEBI microscopy (Supplementary Figures S8, S9). With a total size of more than 400 million pixels, this last example confirms that our pipeline is compatible with large datasets.

5.3 Feature Importance

The RFC enables a ranking of the features used during model training through a metric called impurity-based importance. The importance of a feature is calculated as the reduction in its Gini criterion. This reduction in impurity is afterward express in percentages. We show in Figure 8 the eleven most important features for the three examples previously presented, as well as the importance of PIXEL VALUE corresponding simply to the intensity of a given voxel (see Supplementary Figures S10, S11 for the complete features importance for the 56 features used). Interestingly, this feature, generally used in thresholding for crude denoising or segmentation, seems to have almost no importance in the final prediction.

FIGURE 8
www.frontiersin.org

FIGURE 8. Feature importance for the RFC trained on a confocal image of the mouse olfactory bulb interneurons (A), and the MRI of a breast cancer and CT-scan of a lung cancer (B). The top eleven features ranked by impurity-based importance are represented, as well as the PIXEL VALUE feature.

Comparing the RFC impurity-based importance from both neuroscience microscopy and medical image types (CT-scans and MRI), we note a large variety of features that drive the final classification prediction. Indeed, for medical images, it seems that large spatial filters with 3D kernel sizes of 9–11 voxels contribute mainly to the importance, while the typical kernel size seems closer to 5 or 7 for microscopy. It can be explained by the different characteristics of the respective structures of interest (bulkier for the tumors and more tubular for the neuronal branches). While not surprising, an extended version of this procedure with larger sets of features can guide machine learning procedures that attempt to reduce the number of features used for learning.

6 Discussion

Virtual and augmented reality will likely play an increasingly important role in research and medical applications. More specifically, leveraging VR as part of analysis pipelines will be essential in defining algorithms and processes.

This paper focused on performing interaction and analysis in VR. We developed libraries and toolboxes allowing data annotation and analysis where computationally intensive operations are done on servers or the cloud and data fusion within the VR environment. Our approach combines several updates of the DIVA platform and a generalist interface allowing cloud computing from the VR environment. We showed proof-of-concept results on an accelerated image annotation task in which new data requires human-in-the-loop intervention to provide initial results. This provides a direct process in which limited data annotation is sufficient to train a simple statistical learning approach to classify voxels in entire image stacks.

In this work, we demonstrate the usefulness of VR in ameliorating the efficiency of data annotation tasks. The ability to grasp the volume of the data in a stereoscopic 3D viewing context and tag a small portion of it to perform an entire segmentation task is unique and promising. In contrast, the same task applied to 2D slices would require significant exploration and proves to be a tedious process.

We take advantage of the possibility of adjusting the transfer function to visualize the region of interests and their vicinity correctly. Rapidly tagging voxels not belonging to the structure of interest, by a large stroke using the VR controller (e.g., the pink streaks in Figure 6), will be instrumental in accelerating data annotation by easing learning. When compared to ground truths, the inspection of learned results is often a time-consuming task, especially when dealing with 3D data. Through its capacity to fuse different volumes in the representation, DIVA offers a suitable and time-saving environment to perform such a comparison. In turn, it allows exploring the reliability of the learning procedure and assess the quality of the ground truth itself when dealing with ambiguous data. It provides the possibility of re-annotation where anomalies are detected to feed the re-training procedure.

In this paper, we focused on the handcrafted procedure and simple ensemble learning approaches. While some of the data shown were noisy (e.g., Figure 7) or subject to artifacts (e.g., Figures 6A–F), they were relatively unambiguous. Local handcrafted features were sufficient to allow efficient semi-automated annotation. Furthermore, we limited ourselves to cases where data were completely new, and the learner would be mainly used on the explored dataset. However, complex data or the design of reusable learners may require learned features. The current pipeline can perform more complex learning. We showed (see https://github.com/DecBayComp/VoxelLearning) an example where VR annotation was directly used to capture a volume of voxels around the annotated ones and where learning was directly performed on these volumes with learned features. Finally, large-scale deep convolution-based learning can be performed by directly transmitting the full data to the cloud and using DIVA Cloud to link annotation to learning by simply exchanging voxel identifiers.

Our future initiatives will center on two topics. First, we will 1) extend DIVA and DIVA Cloud to run on AR headsets and tablets. AR involves overlaying the visual representation of the data onto the world while not immersing the user in an artificial environment. However, AR often involves representation of lesser quality than VR when used in glasses or headsets and have limited computational resources, especially when used in phones or tablets. The initiative’s core is to reduce the computational burden of full volumetric rendering to allow visualization and interaction in an AR environment. Second, we will 2) extend DIVA Cloud to allow rendering computation on the cloud and stream to the VR headset. This cloud extension will pave the way to large dataset rendering and more computationally involved rendering approaches, such as path tracing.

Data Availability Statement

Publicly available datasets and original contributions were analyzed in this study. This data as well as the entire pipeline can be found here: https://github.com/DecBayComp/VoxelLearning.

Author Contributions

CGu, VM, CGo, TB, GP, MB, and J-BM contributed to the code of DIVA Cloud. VM was the sole developer of the Django interface of DIVA Cloud. MA, KS, and P-ML provided data related to neuroscience, performed data annotation and contributed to the research design. FR and NB provided data and medical expertise regarding visualization and data annotation. They contributed to paper revisions. BH, J-BM, and MB designed and supervised the research. CGu, VM, CGo, BH, MB, and J-BM wrote the paper.

Funding

We acknowledge funding from the Institut Pasteur, the Institut Curie, the Paris Science lettre (PSL) University, the sponsorships of CRPCEN, Gilead Science and foundation EDF, the INCEPTION project (PIA/ANR-16-CONV-0005, OG), the LabEx “Revive” (ANR-10-LABX-73), the programme d’investissement d’avenir supported by L’Agence Nationale de la Recherche ANR-19-P3IA-0001 Institut 3IA Prairie and the support of the AVIRON grant from the Région Ile-de-France (DIM-ELICIT). We acknowledge funding from the Fondation pour la recherche médicale (FRM; DEI20151234398), the Agence National de la recherche (ANR-19-CE42-0003-01), the LabEx CelTisPhyBio (ANR-10-LBX-0038, ANR-10-IDEX-0001-02) and the Institut Curie. We recognize the support of France-BioImaging infrastructure grant ANR-10-INBS-04 (Investments for the future). We acknowledge the financial support of the Agence pour la Recherche sur le Cancer (ARC Foundation), DIM ELICIT and from ITMO Cancer of Aviesan on funds administered by Inserm (grant No. 20CP092-00).

Conflict of Interest

MB and J-BM are cofounders, shareholders and, respectively, Chief Technology Officer (CTO) and Chief Scientific Officer (CSO) of AVATAR MEDICAL SAS, a startup that commercializes software for surgery planning in virtual reality. The DIVA software used in this study is not being commercialized by AVATAR MEDICAL SAS also the company’s technology is based on the same technology. The DIVA software which serves as base for this study is freely available and is reported in El Beheiry, et al. (2020). All developments within this paper are open source. HV was employed by Sanofi R&D.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

We thank Alexandre Baffet (Institut Curie), Thomas Rubin and Jean-Ren Huynh (Collège de France) for the microscopic image datasets provided to generate the figures. We also thank Jebrane Bouaoud (AP-HP) for the medical images provided to generate the figures. We acknowledge Stephane Fournier, the naviscope Inria project lab teams & Thomas Isenberg for fruitful discussions.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fbinf.2021.777101/full#supplementary-material

References

Arganda-Carreras, I., Kaynig, V., Rueden, C., Eliceiri, K. W., Schindelin, J., Cardona, A., et al. (2017). Trainable Weka Segmentation: a Machine Learning Tool for Microscopy Pixel Classification. Bioinformatics 33, 2424–2426. doi:10.1093/bioinformatics/btx180

PubMed Abstract | CrossRef Full Text | Google Scholar

Arivis, A. G. (2021). Image Visualization and Analysis.

Google Scholar

Ayerbe, V. M. C., Morales, M. L. V., Rojas, C. J. L., and Cortés, M. L. A. (2020). Visualization of 3d Models through Virtual Reality in the Planning of Congenital Cardiothoracic Anomalies Correction: An Initial Experience. World J. Pediatr. Congenit. Heart Surg. 11, 627–629. doi:10.1177/2150135120923618

PubMed Abstract | CrossRef Full Text | Google Scholar

Ayoub, A., and Pulijala, Y. (2019). The Application of Virtual Reality and Augmented Reality in Oral & Maxillofacial Surgery. BMC Oral Health 19, 238. doi:10.1186/s12903-019-0937-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Balo, A. R., Wang, M., and Ernst, O. P. (2017). Accessible Virtual Reality of Biomolecular Structural Models Using the Autodesk Molecule Viewer. Nat. Methods 14, 1122–1123. doi:10.1038/nmeth.4506

PubMed Abstract | CrossRef Full Text | Google Scholar

Bartlett, J. D., Lawrence, J. E., Stewart, M. E., Nakano, N., and Khanduja, V. (2018). Does Virtual Reality Simulation Have a Role in Training Trauma and Orthopaedic Surgeons? Bone Jt. J 100-B, 559–565. doi:10.1302/0301-620x.100b5.bjj-2017-1439

CrossRef Full Text | Google Scholar

Belle, M., Godefroy, D., Couly, G., Malone, S. A., Collier, F., Giacobini, P., et al. (2017). Tridimensional Visualization and Analysis of Early Human Development. Cell 169, 161–e12. doi:10.1016/j.cell.2017.03.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Berg, S., Kutra, D., Kroeger, T., Straehle, C. N., Kausler, B. X., Haubold, C., et al. (2019). Ilastik: Interactive Machine Learning for (Bio)image Analysis. Nat. Methods 16, 1226–1232. doi:10.1038/s41592-019-0582-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Berge, L.-P., Aouf, N., Duval, T., and Coppin, G. (2016). “Generation and VR Visualization of 3d point Clouds for Drone Target Validation Assisted by an Operator,” in Proceeding of the 2016 8th Computer Science and Electronic Engineering (CEEC), Colchester, UK, 28-30 Sept. 2016 (IEEE), 66–70. doi:10.1109/ceec.2016.7835890

CrossRef Full Text | Google Scholar

Bernardo, A. (2017). Virtual Reality and Simulation in Neurosurgical Training. World Neurosurg. 106, 1015–1029. doi:10.1016/j.wneu.2017.06.140

PubMed Abstract | CrossRef Full Text | Google Scholar

Betzig, E., Patterson, G. H., Sougrat, R., Lindwasser, O. W., Olenych, S., Bonifacino, J. S., et al. (2006). Imaging Intracellular Fluorescent Proteins at Nanometer Resolution. Science 313, 1642–1645. doi:10.1126/science.1127344

PubMed Abstract | CrossRef Full Text | Google Scholar

Blanc, T., El Beheiry, M., Caporal, C., Masson, J. B., and Hajj, B. (2020). Genuage: Visualize and Analyze Multidimensional Single-Molecule point Cloud Data in Virtual Reality. Nat. Methods 17, 1100–1102. doi:10.1038/s41592-020-0946-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Boedecker, C., Huettl, F., Saalfeld, P., Paschold, M., Kneist, W., Baumgart, J., et al. (2021). Using Virtual 3d-Models in Surgical Planning: Workflow of an Immersive Virtual Reality Application in Liver Surgery. Langenbecks Arch. Surg. 406, 911–915. doi:10.1007/s00423-021-02127-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Bouaoud, J., El Beheiry, M., Jablon, E., Schouman, T., Bertolus, C., Picard, A., et al. (2021). DIVA, a 3d Virtual Reality Platform, Improves Undergraduate Craniofacial Trauma Education. J. Stomatology, Oral Maxillofacial Surg. 122, 367–371. doi:10.1016/j.jormas.2020.09.009

CrossRef Full Text | Google Scholar

Brault, J. B., Khou, C., Basset, J., Coquand, L., Fraisier, V., Frenkiel, M. P., et al. (2016). Comparative Analysis between Flaviviruses Reveals Specific Neural Stem Cell Tropism for Zika Virus in the Mouse Developing Neocortex. EBioMedicine 10, 71–76. doi:10.1016/0003-4916(63)90068-X10.1016/j.ebiom.2016.07.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Cassidy, K. C., Šefčík, J., Raghav, Y., Chang, A., and Durrant, J. D. (2020). ProteinVR: Web-Based Molecular Visualization in Virtual Reality. Plos Comput. Biol. 16, e1007747. doi:10.1371/journal.pcbi.1007747

PubMed Abstract | CrossRef Full Text | Google Scholar

Celery (2021). Distributed Task Queue — Celery 5.1.2 Documentation.

Google Scholar

Chen, E., Luu, W., Chen, R., Rafik, A., Ryu, Y., Zangerl, B., et al. (2020). Virtual Reality Improves Clinical Assessment of the Optic Nerve. Front. Virtual Real. 1, 4. doi:10.3389/frvir.2020.00004

CrossRef Full Text | Google Scholar

Chheang, V., Saalfeld, P., Joeres, F., Boedecker, C., Huber, T., Huettl, F., et al. (2021). A Collaborative Virtual Reality Environment for Liver Surgery Planning. Comput. Graphics 99, 234–246. doi:10.1016/j.cag.2021.07.009

CrossRef Full Text | Google Scholar

Cohen, J. P., Morrison, P., and Dao, L. (2020). COVID-19 Image Data Collection.

Google Scholar

Conrad, R., Ruth, T., Löffler, F., Hadlak, S., Konrad, S., Götze, C., et al. (2020). Efficient Skeleton Editing in a VR Environment Facilitates Accurate Modeling of Highly Branched Mitochondria. Microsc. Microanal 26, 1158–1161. doi:10.1017/s1431927620017158

PubMed Abstract | CrossRef Full Text | Google Scholar

de Mauro, A., Raczkowsky, J., Halatsch, M. E., and Worn, H. (2009). “Virtual Reality Training Embedded in Neurosurgical Microscope,” in In Proceeding of the 2009 IEEE Virtual Reality Conference, Lafayette, LA, USA, 14-18 March 2009 (IEEE), 233–234. doi:10.1109/vr.2009.4811031

CrossRef Full Text | Google Scholar

Dede, C. J., Jacobson, J., and Richards, J. (2017). Introduction: Virtual, Augmented, and Mixed Realities in Education. Singapore: Springer.

Google Scholar

Dekkers, J. F., Alieva, M., Wellens, L. M., Ariese, H. C. R., Jamieson, P. R., Vonk, A. M., et al. (2019). High-resolution 3d Imaging of Fixed and Cleared Organoids. Nat. Protoc. 14, 1756–1771. doi:10.1038/s41596-019-0160-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Django (2016). Django the Web Framework for Perfectionists with Deadlines.

Google Scholar

Djukic, T., Mandic, V., and Filipovic, N. (2013). Virtual Reality Aided Visualization of Fluid Flow Simulations with Application in Medical Education and Diagnostics. Comput. Biol. Med. 43, 2046–2052. doi:10.1016/j.compbiomed.2013.10.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Doutreligne, S., Cragnolini, T., Pasquali, S., Derreumaux, P., and Baaden, M. (2014). “Unitymol: Interactive Scientific Visualization for Integrative Biology” in Proceeding of the 2014 IEEE 4th Symposium on Large Data Analysis and Visualization (LDAV), Paris, France, 9-10 Nov. 2014 (IEEE), 109–110. doi:10.1109/LDAV.2014.7013213

CrossRef Full Text | Google Scholar

Driscoll, M. K., Welf, E. S., Jamieson, A. R., Dean, K. M., Isogai, T., Fiolka, R., et al. (2019). Robust and Automated Detection of Subcellular Morphological Motifs in 3d Microscopy Images. Nat. Methods 16, 1037–1044. doi:10.1038/s41592-019-0539-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Dyer, E., Swartzlander, B. J., and Gugliucci, M. R. (2018). Using Virtual Reality in Medical Education to Teach Empathy. J. Med. Libr. Assoc. 106, 498–500. doi:10.5195/jmla.2018.518

CrossRef Full Text | Google Scholar

El Beheiry, M., Doutreligne, S., Caporal, C., Ostertag, C., Dahan, M., and Masson, J. B. (2019). Virtual Reality: Beyond Visualization. J. Mol. Biol. 431, 1315–1321. doi:10.1016/j.jmb.2019.01.033

CrossRef Full Text | Google Scholar

El Beheiry, M., Godard, C., Caporal, C., Marcon, V., Ostertag, C., Sliti, O., et al. (2020). DIVA: Natural Navigation inside 3d Images Using Virtual Reality. J. Mol. Biol. 432, 4745–4749. doi:10.1016/j.jmb.2020.05.026

CrossRef Full Text | Google Scholar

Elsayed, M., Kadom, N., Ghobadi, C., Strauss, B., Al Dandan, O., Aggarwal, A., et al. (2020). Virtual and Augmented Reality: Potential Applications in Radiology. Acta Radiol. 61, 1258–1265. doi:10.1177/0284185119897362

PubMed Abstract | CrossRef Full Text | Google Scholar

Engel, K., Hadwiger, M., Kniss, J. M., Lefohn, A. E., Salama, C. R., and Weiskopf, D. (2004). “Real-time Volume Graphics,” in Proceedings of the Conference on SIGGRAPH 2004 Course Notes - GRAPH ’04 (Los Angeles, CA: Association for Computing Machinery Press), 29–es. doi:10.1145/1103900.1103929

CrossRef Full Text | Google Scholar

Engel, K., Hadwinger, M., Kniss, J. M., Rezk-Salama, C., and Weiskopf, D. (2017). Real-Time Volume Graphics. Wellesley, Massachusetts, USA: A K Peters, Ltd.

Google Scholar

Esteva, A., Chou, K., Yeung, S., Naik, N., Madani, A., Mottaghi, A., et al. (2021). Deep Learning-Enabled Medical Computer Vision. Npj Digit. Med. 4. doi:10.1038/s41746-020-00376-2

CrossRef Full Text | Google Scholar

Fertleman, C., Aubugeau-Williams, P., Sher, C., Lim, A. N., Lumley, S., Delacroix, S., et al. (2018). A Discussion of Virtual Reality as a New Tool for Training Healthcare Professionals. Front. Public Health 6, 44. doi:10.3389/fpubh.2018.00044

PubMed Abstract | CrossRef Full Text | Google Scholar

Gao, R., Asano, S. M., Upadhyayula, S., Pisarev, I., Milkie, D. E., Liu, T. L., et al. (2019). Cortical Column and Whole-Brain Imaging with Molecular Contrast and Nanoscale Resolution. Science 363, eaau8302. doi:10.1126/science.aau8302

PubMed Abstract | CrossRef Full Text | Google Scholar

Goddard, T. D., Brilliant, A. A., Skillman, T. L., Vergenz, S., Tyrwhitt-Drake, J., Meng, E. C., et al. (2018). Molecular Visualization on the Holodeck. J. Mol. Biol. 430, 3982–3996. doi:10.1016/j.jmb.2018.06.040

PubMed Abstract | CrossRef Full Text | Google Scholar

Golse, N., Petit, A., Lewin, M., Vibert, E., and Cotin, S. (2020). Augmented Reality during Open Liver Surgery Using a Markerless Non-rigid Registration System. J. Gastrointest. Surg. 25, 662–671. doi:10.1007/s11605-020-04519-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Guerriero, L., Quero, G., Diana, M., Soler, L., Agnus, V., Marescaux, J., et al. (2018). Virtual Reality Exploration and Planning for Precision Colorectal Surgery. Dis. Colon Rectum 61, 719–723. doi:10.1097/DCR.0000000000001077

PubMed Abstract | CrossRef Full Text | Google Scholar

Gunther, U., Pietzsch, T., Gupta, A., Harrington, K. I. S., Tomancak, P., Gumhold, S., et al. (2019). “Scenery: Flexible Virtual Reality Visualization on the Java VM,” in Proceeding of the 2019 IEEE Visualization Conference (VIS), Vancouver, BC, Canada, 20-25 Oct. 2019 (IEEE), 1–5. doi:10.1109/visual.2019.8933605

CrossRef Full Text | Google Scholar

Hattab, G., Hatzipanayioti, A., Klimova, A., Pfeiffer, M., Klausing, P., Breucha, M., et al. (2021). Investigating the Utility of VR for Spatial Understanding in Surgical Planning: Evaluation of Head-Mounted to Desktop Display. Sci. Rep. 11, 13440. doi:10.1038/s41598-021-92536-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Hoffman, D. P., Shtengel, G., Xu, C. S., Campbell, K. R., Freeman, M., Wang, L., et al. (2020). Correlative Three-Dimensional Super-resolution and Block-Face Electron Microscopy of Whole Vitreously Frozen Cells. Science 367, eaaz5357. doi:10.1126/science.aaz5357

PubMed Abstract | CrossRef Full Text | Google Scholar

Hörl, D., Rojas Rusak, F., Preusser, F., Tillberg, P., Randel, N., Chhetri, R. K., et al. (2019). BigStitcher: Reconstructing High-Resolution Image Datasets of Cleared and Expanded Samples. Nat. Methods 16, 870–874. doi:10.1038/s41592-019-0501-0

PubMed Abstract | CrossRef Full Text | Google Scholar

ImmersiveTouch (2021). Medical VR Solutions.

Google Scholar

Johnston, A. P. R., Rae, J., Ariotti, N., Bailey, B., Lilja, A., Webb, R., et al. (2018). Journey to the centre of the Cell: Virtual Reality Immersion into Scientific Data. Traffic 19, 105–110. doi:10.1111/tra.12538

PubMed Abstract | CrossRef Full Text | Google Scholar

Laas, E., El Beheiry, M., Masson, J. B., and Malhaire, C. (2021). Partial Breast Resection for Multifocal Lower Quadrant Breast Tumour Using Virtual Reality. BMJ Case Rep. 14, e241608. doi:10.1136/bcr-2021-241608

CrossRef Full Text | Google Scholar

Lang, H., and Huber, T. (2020). Virtual and Augmented Reality in Liver Surgery. Ann. Surg. 271, e8. doi:10.1097/sla.0000000000003601

PubMed Abstract | CrossRef Full Text | Google Scholar

Lau, I., Gupta, A., and Sun, Z. (2021). Clinical Value of Virtual Reality versus 3d Printing in Congenital Heart Disease. Biomolecules 11, 884. doi:10.3390/biom11060884

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, C., and Wong, G. K. C. (2019). Virtual Reality and Augmented Reality in the Management of Intracranial Tumors: A Review. J. Clin. Neurosci. 62, 14–20. doi:10.1016/j.jocn.2018.12.036

CrossRef Full Text | Google Scholar

Liu, W., Lai, B., Wang, C., Bian, X., Yang, W., Xia, Y., et al. (2020). “Learning to Match 2d Images and 3d LiDAR point Clouds for Outdoor Augmented Reality,” in Proceeding of the 2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW), Atlanta, GA, USA, 22-26 March 2020 (IEEE), 654–655. doi:10.1109/vrw50115.2020.00178

CrossRef Full Text | Google Scholar

Lobachev, O., Berthold, M., Pfeffer, H., Guthe, M., and Steiniger, B. S. (2021). Inspection of Histological 3d Reconstructions in Virtual Reality. Front. Virtual Real. 2. doi:10.3389/frvir.2021.628449

CrossRef Full Text | Google Scholar

Lohre, R., Warner, J. J. P., Athwal, G. S., and Goel, D. P. (2020). The Evolution of Virtual Reality in Shoulder and Elbow Surgery. JSES Int. 4, 215–223. doi:10.1016/j.jseint.2020.02.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Manzo, C., and Garcia-Parajo, M. F. (2015). A Review of Progress in Single Particle Tracking: from Methods to Biophysical Insights. Rep. Prog. Phys. 78, 124601. doi:10.1088/0034-4885/78/12/124601

PubMed Abstract | CrossRef Full Text | Google Scholar

Martirosov, S., Hořejší, P., Kopeček, P., Bureš, M., and Šimon, M. (2021). The Effect of Training in Virtual Reality on the Precision of Hand Movements. Appl. Sci. 11, 8064. doi:10.3390/app11178064

CrossRef Full Text | Google Scholar

Masson, J. B., Laurent, F., Cardona, A., Barré, C., Skatchkovsky, N., Zlatic, M., et al. (2020). Identifying Neural Substrates of Competitive Interactions and Sequence Transitions during Mechanosensory Responses in drosophila. Plos Genet. 16, e1008589. doi:10.1371/journal.pgen.1008589

PubMed Abstract | CrossRef Full Text | Google Scholar

Matthews, D. (2018). Virtual-reality Applications Give Science a New Dimension. Nature 557, 127–128. doi:10.1038/d41586-018-04997-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Mostafa, Y. (2012). Learning from Data : A Short Course (United States: AMLBook.Com).

Google Scholar

Naviscope (2021). Image-guided Navigation and VIsualization of Large Data Sets in Live Cell Imaging and microSCOPy.

Google Scholar

O'Connor, M., Deeks, H. M., Dawn, E., Metatla, O., Roudaut, A., Sutton, M., et al. (2018). Sampling Molecular Conformations and Dynamics in a Multiuser Virtual Reality Framework. Sci. Adv. 4, eaat2731. doi:10.1126/sciadv.aat2731

PubMed Abstract | CrossRef Full Text | Google Scholar

Ong, C. S., Krishnan, A., Huang, C. Y., Spevak, P., Vricella, L., Hibino, N., et al. (2018). Role of Virtual Reality in Congenital Heart Disease. Congenit. Heart Dis. 13, 357–361. doi:10.1111/chd.12587

PubMed Abstract | CrossRef Full Text | Google Scholar

Pastel, S., Chen, C.-H., Martin, L., Naujoks, M., Petri, K., and Witte, K. (2021). Comparison of Gaze Accuracy and Precision in Real-World and Virtual Reality. Virtual Reality 25, 175–189. doi:10.1007/s10055-020-00449-3

CrossRef Full Text | Google Scholar

Patel, B. N., Rosenberg, L., Willcox, G., Baltaxe, D., Lyons, M., Irvin, J., et al. (2019). Human-machine Partnership with Artificial Intelligence for Chest Radiograph Diagnosis. NPJ Digit Med. 2, 111. doi:10.1038/s41746-019-0189-7

CrossRef Full Text | Google Scholar

Pavez, E., Chou, P. A., de Queiroz, R. L., and Ortega, A. (2018). Dynamic Polygon Clouds: Representation and Compression for VR/AR. APSIPA Trans. Signal Inf. Process. 7, 1–22. doi:10.1017/atsip.2018.15

CrossRef Full Text | Google Scholar

Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., et al. (2011). Scikit-learn: Machine Learning in Python. J. Machine Learn. Res. 12, 2825–2830.

Google Scholar

Peng, H., Ruan, Z., Long, F., Simpson, J. H., and Myers, E. W. (2010). V3d Enables Real-Time 3d Visualization and Quantitative Analysis of Large-Scale Biological Image Data Sets. Nat. Biotechnol. 28, 348–353. doi:10.1038/nbt.1612

PubMed Abstract | CrossRef Full Text | Google Scholar

Pfandler, M., Lazarovici, M., Stefan, P., Wucherer, P., and Weigl, M. (2017). Virtual Reality-Based Simulators for Spine Surgery: a Systematic Review. Spine J. 17, 1352–1363. doi:10.1016/j.spinee.2017.05.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Pfeiffer, M., Kenngott, H., Preukschas, A., Huber, M., Bettscheider, L., Müller-Stich, B., et al. (2018). IMHOTEP: Virtual Reality Framework for Surgical Applications. Int. J. Comput. Assist. Radiol. Surg. 13, 741–748. doi:10.1007/s11548-018-1730-x

CrossRef Full Text | Google Scholar

Pinter, C., Lasso, A., Choueib, S., Asselin, M., Fillion-Robin, J. C., Vimort, J. B., et al. (2020). SlicerVR for Medical Intervention Training and Planning in Immersive Virtual Reality. IEEE Trans. Med. Robot Bionics 2, 108–117. doi:10.1109/tmrb.2020.2983199

PubMed Abstract | CrossRef Full Text | Google Scholar

PrecisionOS (2021). Virtual Reality (VR) Surgery Training & Simulators.

Google Scholar

Qi, C. R., Yi, L., Su, H., and Guibas, L. J. (2016). PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. arXiv preprint. arXiv:1706.02413.

Google Scholar

Quero, G., Lapergola, A., Soler, L., Shahbaz, M., Hostettler, A., Collins, T., et al. (2019). Virtual and Augmented Reality in Oncologic Liver Surgery. Surg. Oncol. Clin. N. Am. 28, 31–44. doi:10.1016/j.soc.2018.08.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Raimondi, F., Vida, V., Godard, C., Bertelli, F., Reffo, E., Boddaert, N., et al. (2021). Fast-track Virtual Reality for Cardiac Imaging in Congenital Heart Disease. J. Card. Surg. 36, 2598–2602. doi:10.1111/jocs.15508

CrossRef Full Text | Google Scholar

Ramirez, P. Z., Paternesi, C., Luigi, L. D., Lella, L., Gregorio, D. D., and Stefano, L. D. (2020). “Shooting Labels: 3d Semantic Labeling by Virtual Reality,” in Proceeding of the 2020 IEEE International Conference on Artificial Intelligence and Virtual Reality, Utrecht, Netherlands, 14-18 Dec. 2020 (IEEE), 99–106. doi:10.1109/AIVR50618.2020.00027

CrossRef Full Text | Google Scholar

Reitinger, B., Bornik, A., Beichel, R., and Schmalstieg, D. (2006). Liver Surgery Planning Using Virtual Reality. IEEE Comput. Graph Appl. 26, 36–47. doi:10.1109/mcg.2006.131

PubMed Abstract | CrossRef Full Text | Google Scholar

Ruiz, F., Neiva-Vaz, C., Vazquez, M.-P., Masson, J.-B., El Beheiry, M., Pannier, S., et al. (2021). Management of Ischiopagus Twin Separation with a Focus on W-S Incision Design. J. Pediatr. Surg. Case Rep. 64, 101747. doi:10.1016/j.epsc.2020.101747

CrossRef Full Text | Google Scholar

Sadeghi, A. H., Bakhuis, W., Van Schaagen, F., Oei, F. B. S., Bekkers, J. A., Maat, A. P. W. M., et al. (2020). Immersive 3d Virtual Reality Imaging in Planning Minimally Invasive and Complex Adult Cardiac Surgery. Eur. Heart J. - Digital Health 1, 62–70. doi:10.1093/ehjdh/ztaa011

CrossRef Full Text | Google Scholar

Safaryan, K., and Mehta, M. R. (2021). Enhanced Hippocampal Theta Rhythmicity and Emergence of Eta Oscillation in Virtual Reality. Nat. Neurosci. 24, 1065–1070. doi:10.1038/s41593-021-00871-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Salvatore, S. D., Vadalà, G., Oggiano, L., Russo, F., Ambrosio, L., and Costici, P. F. (2021). Virtual Reality in Preoperative Planning of Adolescent Idiopathic Scoliosis Surgery Using Google Cardboard. Neurospine 18, 199–205. doi:10.14245/ns.2040574.287

PubMed Abstract | CrossRef Full Text | Google Scholar

Sellés, T. S. (2013). GaiaVR.

Google Scholar

Seymour, N. E. (2008). Vr to or: A Review of the Evidence that Virtual Reality Simulation Improves Operating Room Performance. World J. Surg. 32, 182–188. doi:10.1007/s00268-007-9307-9

CrossRef Full Text | Google Scholar

Shao, X., Yuan, Q., Qian, D., Ye, Z., Chen, G., le Zhuang, K., et al. (2020). Virtual Reality Technology for Teaching Neurosurgery of Skull Base Tumor. BMC Med. Educ. 20, 3. doi:10.1186/s12909-019-1911-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Shattuck, D. W. (2018). Multiuser Virtual Reality Environment for Visualising Neuroimaging Data. Healthc. Technol. Lett. 5, 183–188. doi:10.1049/htl.2018.5077

PubMed Abstract | CrossRef Full Text | Google Scholar

Sommer, C., Straehle, C., Köthe, U., and Hamprecht, F. A. (2011). “Ilastik: Interactive Learning and Segmentation Toolkit,” in Proceeding of the 2011 IEEE International Symposium on Biomedical Imaging: From Nano to Macro, Chicago, IL, USA, 30 March-2 April 2011 (IEEE), 230–233. doi:10.1109/ISBI.2011.5872394

CrossRef Full Text | Google Scholar

Spark, A., Kitching, A., Esteban-Ferrer, D., Handa, A., Carr, A. R., Needham, L. M., et al. (2020). vLUME: 3d Virtual Reality for Single-Molecule Localization Microscopy. Nat. Methods 17, 1097–1099. doi:10.1038/s41592-020-0962-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Stefani, C., Lacy-Hulbert, A., and Skillman, T. (2018). Confocalvr: Immersive Visualization for Confocal Microscopy. J. Mol. Biol. 430, 4028–4035. doi:10.1016/j.jmb.2018.06.035

PubMed Abstract | CrossRef Full Text | Google Scholar

Stets, J. D., Sun, Y., Corning, W., and Greenwald, S. W. (2017). Visualization and Labeling of point Clouds in Virtual Reality. SIGGRAPH Asia 2017 Posters. New York, NY: Association for Computing Machinery, 2. doi:10.1145/3145690.3145729

CrossRef Full Text | Google Scholar

Sun, G. J., Sailor, K. A., Mahmood, Q. A., Chavali, N., Christian, K. M., Song, H., et al. (2013). Seamless Reconstruction of Intact Adult-Born Neurons by Serial End-Block Imaging Reveals Complex Axonal Guidance and Development in the Adult hippocampus. J. Neurosci. 33, 11400–11411. doi:10.1523/JNEUROSCI.1374-13.2013

CrossRef Full Text | Google Scholar

SurgicalTheater (2021). Experience 360°XR ANATOMY.

Google Scholar

syGlass (2021). Scientific Virtual Reality.

Google Scholar

Theart, R. P., Loos, B., and Niesler, T. R. (2017). Virtual Reality Assisted Microscopy Data Visualization and Colocalization Analysis. BMC Bioinformatics 18, 64. doi:10.1186/s12859-016-1446-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Thomsen, A. S., Bach-Holm, D., Kjærbo, H., Højgaard-Olsen, K., Subhi, Y., Saleh, G. M., et al. (2017). Operating Room Performance Improves after Proficiency-Based Virtual Reality Cataract Surgery Training. Ophthalmology 124, 524–531. doi:10.1016/j.ophtha.2016.11.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Tomikawa, M., Hong, J., Shiotani, S., Tokunaga, E., Konishi, K., Ieiri, S., et al. (2010). Real-time 3-dimensional Virtual Reality Navigation System with Open MRI for Breast-Conserving Surgery. J. Am. Coll. Surg. 210, 927–933. doi:10.1016/j.jamcollsurg.2010.01.032

CrossRef Full Text | Google Scholar

UnityTechnologies (2021). Unity.

Google Scholar

Uppot, R. N., Laguna, B., McCarthy, C. J., De Novi, G., Phelps, A., Siegel, E., et al. (2019). Implementing Virtual and Augmented Reality Tools for Radiology Education and Training, Communication, and Clinical Care. Radiology 291, 570–580. doi:10.1148/radiol.2019182210

PubMed Abstract | CrossRef Full Text | Google Scholar

Uruthiralingam, U., and Rea, P. M. (2020). Augmented and Virtual Reality in Anatomical Education - A Systematic Review. Adv. Exp. Med. Biol. 1235, 89–101. doi:10.1007/978-3-030-37639-0_5

PubMed Abstract | CrossRef Full Text | Google Scholar

Usher, W., Klacansky, P., Federer, F., Bremer, P. T., Knoll, A., Yarch, J., et al. (2017). A Virtual Reality Visualization Tool for Neuron Tracing. IEEE Trans. Vis. Comput. Graph 24, 994–1003. doi:10.1109/TVCG.2017.2744079

PubMed Abstract | CrossRef Full Text | Google Scholar

Venkatesan, M., Mohan, H., Ryan, J. R., Schürch, C. M., Nolan, G. P., Frakes, D. H., et al. (2021). Virtual and Augmented Reality for Biomedical Applications. Cell Rep Med 2, 100348. doi:10.1016/j.xcrm.2021.100348

PubMed Abstract | CrossRef Full Text | Google Scholar

Verhey, J. T., Haglin, J. M., Verhey, E. M., and Hartigan, D. E. (2020). Virtual, Augmented, and Mixed Reality Applications in Orthopedic Surgery. Int. J. Med. Robot 16, e2067. doi:10.1002/rcs.2067

CrossRef Full Text | Google Scholar

Wake, N., Nussbaum, J. E., Elias, M. I., Nikas, C. V., and Bjurlin, M. A. (2020). 3d Printing, Augmented Reality, and Virtual Reality for the Assessment and Management of Kidney and Prostate Cancer: A Systematic Review. Urology 143, 20–32. doi:10.1016/j.urology.2020.03.066

PubMed Abstract | CrossRef Full Text | Google Scholar

Walbron, P., Thomazeau, H., and Sirveaux, F. (2019). „Virtual reality simulation" in der Orthopädie und Unfallchirurgie in Frankreich. Unfallchirurg 122, 439–443. doi:10.1007/s00113-019-0649-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, S. S., Zhang, S. M., and Jing, J. J. (2012). Stereoscopic Virtual Reality Models for Planning Tumor Resection in the Sellar Region. BMC Neurol. 12, 146. doi:10.1186/1471-2377-12-146

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, Y., Li, Q., Liu, L., Zhou, Z., Ruan, Z., Kong, L., et al. (2019). Teravr Empowers Precise Reconstruction of Complete 3-d Neuronal Morphology in the Whole Brain. Nat. Commun. 10, 3474. doi:10.1038/s41467-019-11443-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Wirth, F., Quehl, J., Ota, J., and Stiller, C. (2019). “PointAtMe: Efficient 3d point Cloud Labeling in Virtual Reality,” in Proceeding of the 2019 IEEE Intelligent Vehicles Symposium (IV), Paris, France, 9-12 June 2019 (IEEE), 1693–1698. doi:10.1109/ivs.2019.8814115

CrossRef Full Text | Google Scholar

Wisotzky, E. L., Rosenthal, J.-C., Eisert, P., Hilsmann, A., Schmid, F., Bauer, M., et al. (2019). “Interactive and Multimodal-Based Augmented Reality for Remote Assistance Using a Digital Surgical Microscope,” in Proceeding of the 2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR), Osaka, Japan, 23-27 March 2019 (IEEE), 1477–1484. doi:10.1109/vr.2019.8797682

CrossRef Full Text | Google Scholar

Yoo, J. S., Patel, D. S., Hrynewycz, N. M., Brundage, T. S., and Singh, K. (2019). The Utility of Virtual Reality and Augmented Reality in Spine Surgery. Ann. Transl Med. 7, S171. doi:10.21037/atm.2019.06.38

PubMed Abstract | CrossRef Full Text | Google Scholar

Zheng, Z., Lauritzen, J. S., Perlman, E., Robinson, C. G., Nichols, M., Milkie, D., et al. (2018). A Complete Electron Microscopy Volume of the Brain of Adult drosophila Melanogaster. Cell 174, 730–e22. doi:10.1016/j.cell.2018.06.019

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: virtual reality, cloud computation, one-shot learning, inference, human-in-the-loop, MRI, CT-scan

Citation: Guérinot C, Marcon V, Godard C, Blanc T, Verdier H, Planchon G, Raimondi F, Boddaert N, Alonso M, Sailor K, Lledo P-M, Hajj B, El Beheiry M and Masson J-B (2022) New Approach to Accelerated Image Annotation by Leveraging Virtual Reality and Cloud Computing. Front. Bioinform. 1:777101. doi: 10.3389/fbinf.2021.777101

Received: 14 September 2021; Accepted: 15 December 2021;
Published: 31 January 2022.

Edited by:

Stephen Taylor, University of Oxford, United Kingdom

Reviewed by:

Ondrej Strnad, King Abdullah University of Science and Technology, Saudi Arabia
William C. Ray, Nationwide Children’s Hospital, United States

Copyright © 2022 Guérinot, Marcon, Godard, Blanc, Verdier, Planchon, Raimondi, Boddaert, Alonso, Sailor, Lledo, Hajj, El Beheiry and Masson. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Mohamed El Beheiry, mohamed.el-beheiry@pasteur.fr; Jean-Baptiste Masson, jbmasson@pasteur.fr

These authors have contributed equally to this work and share first authorship

These authors have contributes equally to this work and share last authorship

Download