Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 10 September 2020
Sec. Performance Science

Distinguishing Intermediate and Novice Surgeons by Eye Movements

\r\nGonca Gokce Menekse Dalveren,*Gonca Gokce Menekse Dalveren1,2*Nergiz Ercil CagiltayNergiz Ercil Cagiltay3
  • 1Department of Computer Science, Norwegian University of Science and Technology, Gjøvik, Norway
  • 2Department of Information Systems Engineering, Atılım University, Ankara, Turkey
  • 3Department of Software Engineering, Atilim University, Ankara, Turkey

Surgical skill-level assessment is key to collecting the required feedback and adapting the educational programs accordingly. Currently, these assessments for the minimal invasive surgery programs are primarily based on subjective methods, and there is no consensus on skill level classifications. One of the most detailed of these classifications categorize skill levels as beginner, novice, intermediate, sub-expert, and expert. To properly integrate skill assessment into minimal invasive surgical education programs and provide skill-based training alternatives, it is necessary to classify the skill levels in as detailed a way as possible and identify the differences between all skill levels in an objective manner. Yet, despite the existence of very encouraging results in the literature, most of the studies have been conducted to better understand the differences between novice and expert surgical skill levels leaving out the other crucial skill levels between them. Additionally, there are very limited studies by considering the eye-movement behaviors of surgical residents. To this end, the present study attempted to distinguish novice- and intermediate-level surgical residents based on their eye movements. The eye-movement data was recorded from 23 volunteer surgical residents while they were performing four computer-based simulated surgical tasks under different hand conditions. The data was analyzed using logistic regression to estimate the skill levels of both groups. The best results of the estimation revealing a 91.3% recognition rate of predicting novice and intermediate surgical residents on one scenario were selected from four under the dominant hand condition. These results show that the eye-movements can be potentially used to identify surgeons with intermediate and novice skills. However, the results also indicate that the order in which the scenarios are provided, and the design of the scenario, the tasks, and their appropriateness with the skill levels of the participants are all critical factors to be considered in improving the estimation ratio, and hence require thorough assessment for future research.

Introduction

Surgical skill-level assessment is a critical process to ensure competence and prevent clinical errors while developing effective instructional methods for training. It is commonly acknowledged that the skill levels of surgeons can vary depending on numerous elements (Uhrich et al., 2002; Silvennoinen et al., 2009; Chandra et al., 2010; De Blacam et al., 2012). Currently, the assessment of surgical residents’ skill levels is based on subjective approaches and conventional methods, and as such, are prone to bias and questionable rationality (Resnick et al., 1991; Wanzel et al., 2002; Richstone et al., 2010; Meneghetti et al., 2012). Hence, the traditional assessment methods have many shortcomings, such as subjective evaluation, need for expert surgeons and requiring standardized methods of assessment (Moorthy et al., 2003; Ahmidi et al., 2012; Zheng et al., 2015). Several studies have reported that more effective skill-level assessment techniques and training evaluation systems could improve skill-based training, and accordingly patient healthcare (Reiley et al., 2011; D’Angelo et al., 2015; Zheng et al., 2015). Such assessment is also necessary to adapt the content and level of training programs to the surgical residents’ needs and experience, providing sufficient feedback for both trainees and educators, and improving the curriculum (De Blacam et al., 2012; Cundy et al., 2015; Cagiltay et al., 2017; Mizota et al., 2020). On the other hand, simulation environments allow trainees to develop knowledge and necessary skills in a safe and effective environment to apply in real-world applications (Earle, 2006; Van Sickle et al., 2006; Carter et al., 2010; Cristancho et al., 2012; Bilgic et al., 2019). Thus, these virtual and interactive settings in computer-based systems are risk-free (Gallagher and Satava, 2002; Maran and Glavin, 2003; Earle, 2006; Van Sickle et al., 2006; Loukas et al., 2011; Seagull, 2012). Accordingly, several training scenarios have been developed for surgical-skill improvement and assessment requirements for surgery purposes (McNatt and Smith, 2001; Gallagher and Satava, 2002; MacDonald et al., 2003; Jones, 2007; Curtis et al., 2012; Feldman et al., 2012; Hsu et al., 2015). Oropesa et al. (2011) report that adaptation of newly developed tools and techniques to surgical training programs has the possibility to realize an essential role for the assessment of surgical skills. However, there are currently very limited studies considering the precise assessment of these skill levels and they have mainly used the performance metrics for the assessment purposes. For instance, in previous studies, surgeons’ experience levels were assessed for each task in accordance to time, errors, and economy of movement for each hand (McNatt and Smith, 2001; Adrales et al., 2003; Arora et al., 2010; Meneghetti et al., 2012; Perrenot et al., 2012; Mathis et al., 2020). An earlier study objectively estimated the skill levels in a virtual surgical training environment using performance data from surgical residents, such as task completion time, distance, and success (Topalli and Cagiltay, 2019). Another study used electromyography (EMG) activity and muscular discomfort scores to assess the surgical skill levels of surgeons (Uhrich et al., 2002). Oostema et al. (2008) quantified the laparoscopic operative experience for each user using time, path length, and smoothness. Similarly, McDougall et al. (2006) attempted to distinguish surgeons who were laparoscopically naive and those with laparoscopic experience regardless of the degree of previous related surgical experience (McDougall et al., 2006). Oropesa et al. (2013) used motion metrics for assessing laparoscopic psychomotor skills from video analysis.

New technologies offer advantages through recording the eye movements (Atkins et al., 2013; Zheng et al., 2015; Kruger and Doherty, 2016; Gegenfurtner et al., 2017; Jian and Ko, 2017) of surgeons and analyzing the obtained data to provide a cost-effective, automated, and objective basis for assessing their skill levels (Ahmidi et al., 2012; Tien et al., 2014). In this respect, eye-tracking provides objective metrics about human behavior (Yarbus, 1967; Gegenfurtner and Seppänen, 2013; Dogusoy-Taylan and Cagiltay, 2014; Piccardi et al., 2016; Tsai et al., 2016; Bröhl et al., 2017; Jian and Ko, 2017; Tricoche et al., 2020), and these systems have many beneficial properties, making it easy to record and analyze eye-movement data (Gegenfurtner and Seppänen, 2013; Tien et al., 2015; Zheng et al., 2015; Kruger and Doherty, 2016; Gegenfurtner et al., 2017). Thus, eye-tracking is used for assessing and understanding the differences between skill levels in the medical domain (Stuijfzand et al., 2016; Gegenfurtner et al., 2017; McLaughlin et al., 2017; Fichtel et al., 2019). The differences in performances are related to the information-processing capabilities of the left and right hemispheres of the brain, implying that when visual control is required, the dominant hand is likely to perform better than the non-dominant hand and both hands (Hoffmann, 1997). For instance, Vickers (1995) reported the difference in eye movements between expert and novice sports players in foul shooting, in which expert players were seen not to follow the whole shooting process with their eyes but novice players used their eyes to adjust their shots before releasing the ball. In another different though related study, Kasarskis et al. (2001) investigated the differences between the eye movements of expert and novice pilots in a landing simulation and showed that the fixation times of novices were longer than those of the experts because the latter assembled the necessary information more rapidly than the former. Furthermore, the study of Khan et al. (2012) reports significant differences between surgeons from different skill levels on their gaze patterns while watching surgical videos. Another study shows that according to the results of a meta-analysis fixation duration of non-experts is longer than the experts (Gegenfurtner et al., 2011). Understanding these differences, classifying different skill-leveled surgeons appropriately and objectively is another critical task for surgical education programs (McNatt and Smith, 2001; Uhrich et al., 2002; Perrenot et al., 2012; Doleck et al., 2016). By objectively classifying the experience levels of participants, adaptive systems can be developed for their training (Perrenot et al., 2012), and objective standards can be generated (Oostema et al., 2008) as an evaluation criterion for student admission to such programs. However, in the literature, there is no study reporting the precise estimations of the surgeons’ skill levels, especially in relation to their endoscopic surgical skills and eye movement behaviors.

Additionally, there is no consensus on the classification of the skill levels of minimal invasive surgery procedures. Some studies classify it as novice and expert (Aggarwal et al., 2006; Fichtel et al., 2019), where others consider it as novice, intermediate, and expert (Grantcharov et al., 2003; Schreuder et al., 2009; Shetty et al., 2012). Silvennoinen et al. (2009) presented the most detailed classification of the expertise levels of minimally invasive surgery in the following five stages: a beginner who has simply non-specialist knowledge, a novice who has started to develop initial knowledge in the domain, an intermediate who has intensified his/her knowledge, a sub-expert who has general knowledge but incomplete knowledge of a specialized domain, and an expert who has deep knowledge of a specialized domain (Silvennoinen et al., 2009). Earlier studies were mainly conducted with novices and experts for determining their skill levels (Vickers, 1995; Rosser et al., 1998; Kasarskis et al., 2001; Richstone et al., 2010). Even though there have been other efforts to classify the experience levels of surgeons, these were mostly to differentiate novices and experts using a box trainer from their hand movements (Ahmidi et al., 2010; Horeman et al., 2011, 2013; van Empel et al., 2012). Hence, there are very few studies classifying the novice and intermediate endoscopic surgical skill levels based on eye movements with one earlier study (Chmarra et al., 2010) reporting 74.2% accuracy. Although there are several studies referring to simulation based training environments for endoscopic surgery procedures, they are very limited in classifying the endoscopic surgery skill levels in a detail manner and at a higher level of accuracy. Furthermore, there are very limited studies that focus on better understanding the possible contributions of eye-movement analyses on these assessment procedures.

The main aim of this study was to provide insights into novice and intermediate surgical skill level assessment techniques using eye-tacking technology in virtual reality settings. First, it attempts to show a possible relationship between the eye-movements and surgical skill levels for intermediates and novices. Afterward, it attempts to classify these skill levels through their eye-movement behaviors. The purpose behind this initiative was to integrate skill assessment into the curriculum of surgical education programs to provide appropriate feedback and understand the differences between these skill levels. In particular, compared to the gap between novice and expert groups being higher (Silvennoinen et al., 2009), differentiating novice and intermediate skill level groups is particularly challenging. Additionally, to the best of our knowledge, no study has been found that classified these groups based on eye-movement events. For instance, the study of Khan et al. (2012) reports differences on gaze patterns of surgical residents while watching surgical videos. According to the results of this study it was reported that novice surgeons’ eye-gaze patterns generally roam from key areas of the operational field, whereas expert surgeons’ eye-gaze patterns focused on the task-relevant areas (Khan et al., 2012). Also, the study of Gegenfurtner et al. (2011) shows differences on fixation durations between experts and non-experts. According to this study, experts had shorter fixation durations compared to the non-experts (Gegenfurtner et al., 2011). The results of this earlier studies provide evidences indicating differences between novice and intermediate surgical residences’ eye movements while performing surgical tasks. However, these studies consider the differences between experts and non-experts or novices. Hence, earlier studies are very limited considering eye movement behavioral differences of novice and intermediate groups. Therefore, this study was conducted with novices and intermediates and the main hypothesis of the study is that there is a significant difference between intermediate and novice surgical residents’ eye movements considering number of fixations, fixation durations and number of saccades while they are performing surgical tasks under different hand conditions (dominant-hand, non-dominant hand, and both hand). The second hypothesis is that the novice and intermediate surgical residents can be classified by using their eye movements considering number of fixations while they are performing surgical tasks under different hand conditions (dominant-hand, non-dominant hand, and both hand). Hence, the main contribution of this study is classifying the novice- and intermediate-group surgeons through their eye movement events to predict their skill levels accordingly.

For this purpose, in this study, four surgical scenarios were applied and performed in a virtual reality environment under different hand conditions (dominant hand, non-dominant hand, and both hands) using haptic devices controlled by surgical residents. Their eye-movement data was collected while they performed the given tasks in these scenarios. The data was classified into number of fixations, fixation duration and number of saccades events with an open-source eye-movement classification algorithm, Binocular Individual Threshold (BIT) (van der Lans et al., 2011), a velocity-based algorithm for identifying fixations and saccades. BIT is effective since it is a parameter-free fixation-identification mechanism which automatically identifies task- and individual-specific velocity thresholds (van der Lans et al., 2011; Menekse Dalveren and Cagiltay, 2019). BIT classifies fixations and saccades from the data of both eyes based on individual-specific thresholds. This algorithm has advantages over other similar algorithms such as containing binocular viewing and being independent of machine and sampling frequency (van der Lans et al., 2011). Therefore, BIT algorithm can classify the data from varying eye-tracking devices with various sampling frequencies and different level of sensitivities (van der Lans et al., 2011).

Materials and Methods

In this study, the eye movements of surgical residents were investigated to distinguish their surgical skill levels. Recorded raw eye data was classified and analyzed for differentiating the intermediate and novice surgical residents. This study was carried out in accordance with the protocol which was approved by the Human Research Ethics Board of Atılım University. All subjects gave verbal and written informed consent in accordance with the Declaration of Helsinki.

Participants

Twenty-three volunteer participants from two departments were recruited for this study: 13 from neurosurgery and 10 from ear-nose-throat (ENT) surgery. Since it is difficult to access surgeons in specialized fields to volunteer for such initiatives, this number of participants is usually considered as acceptable. The majority of the earlier studies were also conducted with limited number of participants, such as 14 surgeons (Wilson et al., 2010), 22 surgeons (Cope et al., 2015), 9 neurosurgeons (Eivazi et al., 2017), 12 students (Uemura et al., 2016), and 19 surgical residents (Hsu et al., 2015). Of the 23 participants in the current study, 14 (three females) were novices with an average age of 27.71 (SD = 6.96), with no one having previously performed an endoscopic surgery on their own. On average, the 14 participants had observed 9.57 (SD = 13.51) and assisted in 3.57 (SD = 10.64) surgical procedures. The remaining nine were intermediates with an average age of 29.33 (SD = 1.50) years, who had observed 48.33 (SD = 31.62) and assisted in 32.00 (SD = 24.19) surgical operations, while performing 16.56 (SD = 16.60) operations on average on their own.

Simulated Scenarios

The simulation scenarios were developed to address the development of required skills for endoscopic skull-base surgery operations which are mainly conducted by ENT and neurosurgery experts. Skills, such as depth perception, 2D and 3D environmental transformations, left-right hand coordination, and hand-eye coordination were aimed to be addressed through the designed scenarios, which were based on surgical experts’ guidance from the neurosurgery and ENT departments of the medical school.

Virtually simulated environments resemble the real world with their visualization and interaction properties (Perrenot et al., 2012; Zhang et al., 2017). In this study, four scenarios were developed based on surgical skill development requirements for endoscopic surgery purposes to collect the participants’ eye movement data.

Scenario 1

In this scenario, the participants are expected to use an operational instrument through a haptic device. The scenario is designed to improve depth perception, camera control, 2D-3D conversion, and effective tool usage. With the aid of this tool, the participants have to catch a red ball, which appears randomly in different locations in a room (Figure 1A). The ball must be placed close to the cube; then, the color of the ball turns green. The participants must grab this ball and place it on the green cube. The cube also appears at random positions in the room. This task is repeated 10 times.

FIGURE 1
www.frontiersin.org

Figure 1. Scenarios [(A) scenario 1, (B) scenario 2, (C) scenario 3, and (D) scenario 4].

Scenario 2

A red ball appears randomly in one of the boxes as shown in Figure 1B. The red ball will explode if the approach is at the right angle using the haptic device. This process is repeated 10 times with the aim of developing depth perception and improving the participant’s ability to approach a certain point from the correct angle.

Scenario 3

This scenario is a simulated surgical model, in which the participant feels as if they were actually undertaking surgery. As depicted in Figure 1C, the model shows the inside of a human nose. The participants are expected to remove the 10 objects located at different places within the model using the surgical tool. They can move the tool within the model using the haptic device and feel the tissue as the device gives force feedback upon collision with any surface. By using the surgical tool in the most accurate way the participant can complete the operation by carefully removing all the objects from the model.

Scenario 4

This scenario is a simulated surgical model and designed like inside of a human nose with similar texture, simulating the field vision of a surgeon during an actual operation. The main task in this scenario is to move an object (the white ball) smoothly on a specified path (Figure 1D). The task is completed when the participant reaches the end of the specified path. The important aspect in this scenario is using the haptic device at a proper angle to keep the object moving on the path.

Procedure

The participants were informed about the purpose, procedures, and duration of the experiment. The eye-movement data of the surgical residents were collected using an eye-tracking device while they performed the surgical tasks. For the eye data recordings, the Eye Tribe tracker was positioned under the monitor (Figure 2) to provide the information regarding the distance of the user to the monitor and the visibility of the eyes. Therefore, the participants were asked to adjust their seating position to a distance of 70 cm in front of the tracker. The recording frequency was set to 60 Hz and a 9-point calibration process was performed.

FIGURE 2
www.frontiersin.org

Figure 2. Experiment design.

The scenarios were ordered from easy to difficult levels with surgical experts’ guidance from the neurosurgery and ENT departments of the medical school. All the participants performed the scenarios in the same order of 1, 2, 3, and 4 to reduce the chance of negative effects on learning. Each scenario has 10 tasks with a fixed period of time as 20 s. Therefore, each scenario takes 200 s (20 × 10). For added objectivity and reduced order effect, 12 of the participants performed the tasks using their dominant hand, and the remainder with their non-dominant hand. In all cases, the tasks were last performed under the both-hands condition. Each scenario was performed by surgeons using Geomagic Touch haptic devices. All scenarios took place in a simulated environment using a surgical instrument under virtually lit conditions in dominant and non-dominant hand conditions, whereas in the both-hands condition, the participants had to use the light source and the surgical instrument in a coordinated fashion. Therefore, in the both-hands condition, the participants were required to use two haptic devices. They used the first haptic device to hold the surgical instrument with their dominant hand, and they held the second haptic device with their non-dominant hand as a light source to illuminate the operating area as shown in Figure 3. Since the Eye Tribe eye tracker can only record the eye movement data in x and y coordinates, there is a need to classify this raw data into eye movement events, such as fixations; a slow-period eye movement with reduced dispersion and velocity. The classification of raw eye data was performed with BIT, which is an open-source eye movement classification algorithm capable of categorizing raw eye data into fixations and saccades independently from the eye tracker and without specifying the threshold values (van der Lans et al., 2011). In addition, the classified eye movement data was analyzed using statistical methods to predict the skill levels of surgeons based on their eye-movement behaviors.

FIGURE 3
www.frontiersin.org

Figure 3. Both-hands condition.

Analysis

The analysis of all the data classified by the BIT algorithm was performed using SPSS for the Windows software package (version 23; IBM Corporation, New York, United States) at 95% confidence level. Since the sample size was 23 and the normality assumptions were violated, the non-parametric test technique of Mann-Whitney was used (McCrum-Gardner, 2008) to identify differences between the intermediate and novice surgeons in terms of eye-movement data. This technique is used to compare two independent groups in relation to a quantitative variable (McCrum-Gardner, 2008). A logistic regression analysis was conducted to predict the skill levels of the surgical residents.

Results

According to the Mann-Whitney test for the novices and intermediates, there was a statistically significant difference in the number of fixations between the two groups (Figure 4) for Scenario 1 under the dominant hand condition. In other scenarios and conditions, no significant difference was found for number of fixations, fixation durations and number of saccades. The scatter plot for number of fixations of Scenario 1 under dominant hand condition was given in Figure 5. In Scenario 1, under the dominant hand condition, the novice surgeons fixated more than the intermediate surgeons according to the results of the BIT algorithm (U = 16, p < 0.05, z = −2.980, r = −0.62). This result reveals that there is a potential difference between the novice and intermediate surgical residents with the experience level having an impact on the number of fixations and based on the r value (−0.62) the effect size shows that the difference between intermediate and novice surgeons can be considered as high (Rosenthal et al., 1994). Also, another statistical power analysis was performed to evaluate the effect sizes based on data of this study and the Eta squared (η2) and d Cohen values were found as 0.38 and 1.57, respectively (Lenhard and Lenhard, 2016).

FIGURE 4
www.frontiersin.org

Figure 4. Number of fixations of novice and intermediate surgeons.

FIGURE 5
www.frontiersin.org

Figure 5. Scatter of the data for number of fixations.

The logistic regression analysis ascertained the effects of the number of fixations, fixation duration and the dominant hand on the prediction of the participants’ skill level. According to the results, in the dominant-hand condition for Scenario 1, the logistic regression model was statistically significant, x2(3) = 15.661, p = 0.001. The model explained 66.9% (Negelkerke R2) of the variance in the skill level, and correctly classified 91.3% of all the cases. For Scenario 2, the logistic regression model was not statistically significant, x2(3) = 6.445, p = 0.092. The model explained 33.1% (Negelkerke R2) of the variance in the skill level, while accurately classifying 69.6% of the cases. For Scenario 3, the logistic regression model was not statistically significant, x2(3) = 5.449, p = 0.142. The model explained 28.6% (Negelkerke R2) of the variance in the skill level with 65.2% of the cases correctly classified. Lastly, in Scenario 4, the model was not statistically significant, x2(3) = 4.941, p = 0.176, explaining 26.2% (Negelkerke R2) of the variance in the skill level and correctly classifying 56.5% of the cases.

According to the results, in the non-dominant hand condition for Scenario 1, the logistic regression model was statistically significant, x2(3) = 9.605, p = 0.022. The model explained 46.3% (Negelkerke R2) of the variance in the skill level and correctly classified 65.2% of all cases. For Scenario 2, the model was also not statistically significant, x2(3) = 4.830, p = 0.185, explaining 25.7% (Negelkerke R2) of the variance in the skill level and correctly classifying 56.5% of the cases. For Scenario 3, the logistic regression model was statistically significant, x2(3) = 10.375, p = 0.016, explaining 49.2% (Negelkerke R2) of the variance in the skill level and correctly classifying 73.9% of all cases. For Scenario 4, the model was not statistically significant, x2(3) = 5.388, p = 0.146 since it explained 28.3% (Negelkerke R2) of the variance in the skill level and correctly classified 56.5% of the cases.

Finally, in the both-hand condition for Scenario 1 the logistic regression model was statistically significant, x2(3) = 14.289, p = 0.003, explaining 62.7% (Negelkerke R2) of the variance in the skill level and correctly classifying 73.9% of all cases. For Scenario 2, the model was not statistically significant, x2(3) = 7.440, p = 0.059 as it 37.5% (Negelkerke R2) of the variance and classified 65.2% of the cases. For Scenario 3, the model was statistically significant, x2(3) = 8.670, p = 0.034 and explained 42.6% (Negelkerke R2) of the variance in the skill level and correctly classified 65.2% of all cases. For Scenario 4, the logistic regression model was again not statistically significant, x2(3) = 8.031, p = 0.45 as it explained 39.9% (Negelkerke R2) of the variance in the skill level and correctly classified 60.9% of the cases.

Discussion

The results of this study show that it is possible to objectively evaluate the skill levels of novice and intermediate surgeons using eye-movement events, such as number of fixations. This result supports the findings of earlier studies (Gallagher and Satava, 2002; Andreatta et al., 2008; Richstone et al., 2010; Gegenfurtner et al., 2011; Khan et al., 2012). For instance, studies indicating that surgical skill levels can be objectively evaluated by eye-tracking metrics through virtually simulated and live environments (Andreatta et al., 2008; Richstone et al., 2010). Additionally, the findings are in parallel with the results of another earlier study showing that experienced laparoscopic surgeons performed the tasks significantly faster, with fewer errors and more economy in the movement of instruments, and greater consistency in performance (Gallagher and Satava, 2002). Although this is more challenging than comparatively evaluating the skill levels of novice and expert surgeons, this study shows that it is possible to classify intermediate and novice level surgical residents through their eye movement behaviors with a high accuracy. Hence, the main contribution of this current study is showing the eye movement behavioral differences of novice and intermediate surgeons and classifying these groups with a high accuracy (91.3%) which are not reported in the earlier studies. This assessment process could be used to accurately monitor the acquisition of skills throughout training programs at a much earlier stage.

However, the results indicate that in different scenarios, the accuracy of estimations may change (Table 1). Iqbal et al. (2004) reported that the difficulty level of the tasks demanded longer processing times and brought higher subjective ratings of mental workload. In other words, the difficulty level of scenarios needs to be designed in parallel with the skill levels of the trainees. When the tasks in the scenario are too easy or too hard in relation to the trainees’ skill levels, then it could be hard to assess their levels with these tasks. Another study stated that the fidelity level was an important factor affecting mental workload (Munshi et al., 2015). Accordingly, the fidelity levels of the scenarios could be another consideration in designing training scenarios according to a specific skill level. For a better integration of these simulation-based training environments into traditional surgical education programs, further studies need to be conducted to standardize these training scenarios. Therefore, the order of the scenarios and the design of the tasks defined for each scenario may also have an effect on these different findings, necessitating additional research to gain a more thorough understanding of the affective factors on the accuracy of estimations.

TABLE 1
www.frontiersin.org

Table 1. Logistic regression results for all conditions and scenarios.

Conclusion

The use of eye-movement events is an objective approach which does not require the time and expenditure required to employ expert evaluators. The results of this study reveal that by appropriately designing scenarios addressing both skills under specific conditions (as in the case of the dominant hand condition in Scenario 1 in this study) it may be possible to distinguish the levels of intermediate and novice surgeons through their eye behaviors, up to 91.3%. The highest ratio of the estimation was achieved in Scenario 1 under the dominant hand condition, which implies several possibilities. First, the content and design of the assessment scenarios should be in line with the skill levels of the participants. Second, the design of the scenarios and tasks that require specific skills and their appropriateness with the skill levels of the participants is another critical factor in improving the estimation ratio. In other words, for this group of participants, the design of Scenario 1 was found to be more appropriate for skill assessment purposes. Third, the order of the scenarios implemented in the research procedure is also of significance; for example, as Scenario 1 was implemented first, the accuracy of the assessment regarding the other scenarios might have declined due to the learning affect.

In our earlier study (Topalli and Cagiltay, 2019), the best estimation accuracy was 86% using features, such as time, distance, camera distance, success, catch time, error time, error distance, and deviation count under dominant hand, non-dominant hand and both-hands conditions. This study showed better accuracy when the participants’ eye-movement events were analyzed. Accordingly, this study provides evidence that the eye-movement behaviors of the surgical residents are very important for assessing their skill levels.

Based on the results of this study, eye movements can provide insights into the skill levels of surgical residents and their current education program. In light of this information, it might be beneficial to properly design and organize simulation-based instructional systems for surgical training programs integrating eye movement behavioral analysis into their current methods. Additionally, the progress levels of trainees can be continuously and objectively assessed, and accordingly guidance can be given during training programs.

In conclusion, performance evaluations through computer-based simulation environments by considering the eye movements of surgeons can be improved. In the near future, data obtained from the eye movements of surgical residents can potentially positively support the current evaluation methods by adding to the degree of objectivity. Future studies considering the eye movement data of surgical residents synchronized with their performance data, as well as hand movement data can provide even better results on these estimations and potentially improve the level of prediction. Additionally, by analyzing the performance data (such as task accuracy, task duration and number of errors) in correlation with the eye movements of the trainees’ deeper insights about the possible relationships between eye movements and surgical skill levels can be gained. Furthermore, future studies are required in consideration of the higher level of expertise in the endoscopic surgery field.

Data Availability Statement

The datasets generated for this study are available on request to the corresponding author.

Ethics Statement

The studies involving human participants were reviewed and approved by the Ethics Committee of the Atılım University. The participants provided their written informed consent to participate in this study.

Author Contributions

NC designed the experiment. GM analyzed the data. GM and NC wrote the article. Both authors contributed to the article and approved the submitted version.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We would like to thank the Hacettepe University Medical School for their valuable support throughout the research.

References

Adrales, G. L., Park, A. E., Chu, U. B., Witzke, D. B., Donnelly, M. B., Hoskins, J. D., et al. (2003). A valid method of laparoscopic simulation training and competence assessment1, 2. J. Surg. Res. 114, 156–162. doi: 10.1016/s0022-4804(03)00315-9

CrossRef Full Text | Google Scholar

Aggarwal, R., Grantcharov, T., Moorthy, K., Hance, J., and Darzi, A. (2006). A competency-based virtual reality training curriculum for the acquisition of laparoscopic psychomotor skill. Am. J. Surg. 191, 128–133. doi: 10.1016/j.amjsurg.2005.10.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Ahmidi, N., Hager, G. D., Ishii, L., Fichtinger, G., Gallia, G. L., and Ishii, M. (2012). “Surgical task and skill classification from eye tracking and tool motion in minimally invasive surgery,” in Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, eds T. Jiang, N. Navab, J. P. W. Pluim, and M. A. Viergever (Berlin: Springer), 295–302. doi: 10.1007/978-3-642-15711-0_37

CrossRef Full Text | Google Scholar

Ahmidi, N., Ishii, M., Fichtinger, G., Gallia, G. L., and Hager, G. D. (2010). An objective and automated method for assessing surgical skill in endoscopic sinus surgery using eye-tracking and tool-motion data. Int. Forum Allergy Rhinol. 2, 507–515. doi: 10.1002/alr.21053

PubMed Abstract | CrossRef Full Text | Google Scholar

Andreatta, P. B., Woodrum, D. T., Gauger, P. G., and Minter, R. M. (2008). LapMentor metrics possess limited construct validity. Simul. Healthc. 3, 16–25. doi: 10.1097/sih.0b013e31816366b9

PubMed Abstract | CrossRef Full Text | Google Scholar

Arora, S., Sevdalis, N., Nestel, D., Woloshynowych, M., Darzi, A., and Kneebone, R. (2010). The impact of stress on surgical performance: a systematic review of the literature. Surgery 147, 318–330, 330.e1–330.e6.

Google Scholar

Atkins, M. S., Tien, G., Khan, R. S., Meneghetti, A., and Zheng, B. (2013). What do surgeons see: capturing and synchronizing eye gaze for surgery applications. Surg. Innov. 20, 241–248. doi: 10.1177/1553350612449075

PubMed Abstract | CrossRef Full Text | Google Scholar

Bilgic, E., Alyafi, M., Hada, T., Landry, T., Fried, G. M., and Vassiliou, M. C. (2019). Simulation platforms to assess laparoscopic suturing skills: a scoping review. Surg. Endosc. 33, 2742–2762. doi: 10.1007/s00464-019-06821-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Bröhl, C., Theis, S., Rasche, P., Wille, M., Mertens, A., and Schlick, C. M. (2017). Neuroergonomic analysis of perihand space: effects of hand proximity on eye-tracking measures and performance in a visual search task. Behav. Inf. Technol. 36, 737–744. doi: 10.1080/0144929x.2016.1278561

CrossRef Full Text | Google Scholar

Cagiltay, N. E., Ozcelik, E., Sengul, G., and Berker, M. (2017). Construct and face validity of the educational computer-based environment (ECE) assessment scenarios for basic endoneurosurgery skills. Surg. Endosc. 31, 4485–4495. doi: 10.1007/s00464-017-5502-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Carter, Y. M., Wilson, B. M., Hall, E., and Marshall, M. B. (2010). Multipurpose simulator for technical skill development in thoracic surgery. J. Surg. Res. 163, 186–191. doi: 10.1016/j.jss.2010.04.051

PubMed Abstract | CrossRef Full Text | Google Scholar

Chandra, V., Nehra, D., Parent, R., Woo, R., Reyes, R., Hernandez-Boussard, T., et al. (2010). A comparison of laparoscopic and robotic assisted suturing performance by experts and novices. Surgery 147, 830–839. doi: 10.1016/j.surg.2009.11.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Chmarra, M. K., Klein, S., De Winter, J. C., Jansen, F.-W., and Dankelman, J. (2010). Objective classification of residents based on their psychomotor laparoscopic skills. Surg. Endosc. 24, 1031–1039. doi: 10.1007/s00464-009-0721-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Cope, A. C., Mavroveli, S., Bezemer, J., Hanna, G. B., and Kneebone, R. (2015). Making meaning from sensory cues: a qualitative investigation of postgraduate learning in the operating room. Acad. Med. 90, 1125–1131. doi: 10.1097/acm.0000000000000740

PubMed Abstract | CrossRef Full Text | Google Scholar

Cristancho, S., Moussa, F., and Dubrowski, A. (2012). Simulation-augmented training program for off-pump coronary artery bypass surgery: developing and validating performance assessments. Surgery 151, 785–795. doi: 10.1016/j.surg.2012.03.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Cundy, T. P., Thangaraj, E., Rafii-Tari, H., Payne, C. J., Azzie, G., Sodergren, M. H., et al. (2015). Force-sensing enhanced simulation environment (ForSense) for laparoscopic surgery training and assessment. Surgery 157, 723–731. doi: 10.1016/j.surg.2014.10.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Curtis, M. T., Diazgranados, D., and Feldman, M. (2012). Judicious use of simulation technology in continuing medical education. J. Contin. Educ. Health Prof. 32, 255–260. doi: 10.1002/chp.21153

PubMed Abstract | CrossRef Full Text | Google Scholar

D’Angelo, A.-L. D., Law, K. E., Cohen, E. R., Greenberg, J. A., Kwan, C., Greenberg, C., et al. (2015). The use of error analysis to assess resident performance. Surgery 158, 1408–1414. doi: 10.1016/j.surg.2015.04.010

PubMed Abstract | CrossRef Full Text | Google Scholar

De Blacam, C., O’keeffe, D. A., Nugent, E., Doherty, E., and Traynor, O. (2012). Are residents accurate in their assessments of their own surgical skills? Am. J. Surg. 204, 724–731. doi: 10.1016/j.amjsurg.2012.03.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Dogusoy-Taylan, B., and Cagiltay, K. (2014). Cognitive analysis of experts’ and novices’ concept mapping processes: an eye tracking study. Comput. Hum. Behav. 36, 82–93. doi: 10.1016/j.chb.2014.03.036

CrossRef Full Text | Google Scholar

Doleck, T., Jarrell, A., Poitras, E. G., Chaouachi, M., and Lajoie, S. P. (2016). A tale of three cases: examining accuracy, efficiency, and process differences in diagnosing virtual patient cases. Australas. J. Educ. Technol. 36, 61–76.

Google Scholar

Earle, D. (2006). Surgical training and simulation laboratory at Baystate Medical Center. Surg. Innov. 13, 53–60. doi: 10.1177/155335060601300109

PubMed Abstract | CrossRef Full Text | Google Scholar

Eivazi, S., Hafez, A., Fuhl, W., Afkari, H., Kasneci, E., Lehecka, M., et al. (2017). Optimal eye movement strategies: a comparison of neurosurgeons gaze patterns when using a surgical microscope. Acta Neurochir. 159, 959–966. doi: 10.1007/s00701-017-3185-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Feldman, M., Lazzara, E. H., Vanderbilt, A. A., and Diazgranados, D. (2012). Rater training to support high-stakes simulation-based assessments. J. Contin. Educ. Health Prof. 32, 279–286. doi: 10.1002/chp.21156

PubMed Abstract | CrossRef Full Text | Google Scholar

Fichtel, E., Lau, N., Park, J., Parker, S. H., Ponnala, S., Fitzgibbons, S., et al. (2019). Eye tracking in surgical education: gaze-based dynamic area of interest can discriminate adverse events and expertise. Surg. Endosc. 33, 2249–2256. doi: 10.1007/s00464-018-6513-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Gallagher, A. G., and Satava, R. (2002). Virtual reality as a metric for the assessment of laparoscopic psychomotor skills. Surg. Endosc. 16, 1746–1752. doi: 10.1007/s00464-001-8215-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Gegenfurtner, A., Lehtinen, E., Jarodzka, H., and Säljö, R. (2017). Effects of eye movement modeling examples on adaptive expertise in medical image diagnosis. Comput. Educ. 113, 212–225. doi: 10.1016/j.compedu.2017.06.001

CrossRef Full Text | Google Scholar

Gegenfurtner, A., Lehtinen, E., and Säljö, R. (2011). Expertise differences in the comprehension of visualizations: a meta-analysis of eye-tracking research in professional domains. Educ. Psychol. Rev. 23, 523–552. doi: 10.1007/s10648-011-9174-7

CrossRef Full Text | Google Scholar

Gegenfurtner, A., and Seppänen, M. (2013). Transfer of expertise: an eye tracking and think aloud study using dynamic medical visualizations. Comput. Educ. 63, 393–403. doi: 10.1016/j.compedu.2012.12.021

CrossRef Full Text | Google Scholar

Grantcharov, T. P., Bardram, L., Funch-Jensen, P., and Rosenberg, J. (2003). Learning curves and impact of previous operative experience on performance on a virtual reality simulator to test laparoscopic surgical skills. Am. J. Surg. 185, 146–149. doi: 10.1016/s0002-9610(02)01213-8

CrossRef Full Text | Google Scholar

Hoffmann, E. R. (1997). Movement time of right-and left-handers using their preferred and non-preferred hands. Int. J. Ind. Ergon. 19, 49–57. doi: 10.1016/0169-8141(95)00092-5

CrossRef Full Text | Google Scholar

Horeman, T., Dankelman, J., Jansen, F. W., and Van Den Dobbelsteen, J. J. (2013). Assessment of laparoscopic skills based on force and motion parameters. IEEE Trans. Biomed. Eng. 61, 805–813. doi: 10.1109/tbme.2013.2290052

PubMed Abstract | CrossRef Full Text | Google Scholar

Horeman, T., Rodrigues, S. P., Jansen, F. W., Dankelman, J., and Van Den Dobbelsteen, J. J. (2011). Force parameters for skills assessment in laparoscopy. IEEE Trans. Haptics 5, 312–322. doi: 10.1109/toh.2011.60

PubMed Abstract | CrossRef Full Text | Google Scholar

Hsu, J. L., Korndorffer, J. R. Jr., and Brown, K. M. (2015). Design of vessel ligation simulator for deliberate practice. J. Surg. Res. 197, 231–235. doi: 10.1016/j.jss.2015.02.068

PubMed Abstract | CrossRef Full Text | Google Scholar

Iqbal, S. T., Zheng, X. S., and Bailey, B. P. (2004). “Task-evoked pupillary response to mental workload in human-computer interaction,” in Proceedings of the CHI’04 Extended Abstracts on Human Factors in Computing Systems, The Hague, 1477–1480.

Google Scholar

Jian, Y.-C., and Ko, H.-W. (2017). Influences of text difficulty and reading ability on learning illustrated science texts for children: an eye movement study. Comput. Educ. 113, 263–279. doi: 10.1016/j.compedu.2017.06.002

CrossRef Full Text | Google Scholar

Jones, D. B. (2007). Video trainers, simulation and virtual reality: a new paradigm for surgical training. Asian J. Surg. 30, 6–12. doi: 10.1016/s1015-9584(09)60121-4

CrossRef Full Text | Google Scholar

Kasarskis, P., Stehwien, J., Hickox, J., Aretz, A., and Wickens, C. (2001). “Comparison of expert and novice scan behaviors during VFR flight,” in Proceedings of the 11th International Symposium on Aviation Psychology, Columbus, OH, 1–6.

Google Scholar

Khan, R. S., Tien, G., Atkins, M. S., Zheng, B., Panton, O. N., and Meneghetti, A. T. (2012). Analysis of eye gaze: do novice surgeons look at the same location as expert surgeons during a laparoscopic operation? Surg. Endosc. 26, 3536–3540. doi: 10.1007/s00464-012-2400-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Kruger, J.-L., and Doherty, S. (2016). Measuring cognitive load in the presence of educational video: towards a multimodal methodology. Australas. J. Educ. Technol. 32, 19–31.

Google Scholar

Lenhard, W., and Lenhard, A. (2016). Calculation of Effect Sizes. Dettelbach: Psychometric.

Google Scholar

Loukas, C., Nikiteas, N., Kanakis, M., and Georgiou, E. (2011). Deconstructing laparoscopic competence in a virtual reality simulation environment. Surgery 149, 750–760. doi: 10.1016/j.surg.2010.11.012

PubMed Abstract | CrossRef Full Text | Google Scholar

MacDonald, J., Williams, R. G., and Rogers, D. A. (2003). Self-assessment in simulation-based surgical skills training. Am. J. Surg. 185, 319–322. doi: 10.1016/s0002-9610(02)01420-4

CrossRef Full Text | Google Scholar

Maran, N. J., and Glavin, R. J. (2003). Low-to high-fidelity simulation–a continuum of medical education? Med. Educ. 37, 22–28. doi: 10.1046/j.1365-2923.37.s1.9.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Mathis, R., Watanabe, Y., Ghaderi, I., and Nepomnayshy, D. (2020). What are the skills that represent expert-level laparoscopic suturing? A Delphi Study. Surg. Endosc. 34, 1318–1323. doi: 10.1007/s00464-019-06904-w

PubMed Abstract | CrossRef Full Text | Google Scholar

McCrum-Gardner, E. (2008). Which is the correct statistical test to use? Br. J. Oral Maxillofac. Surg. 46, 38–41. doi: 10.1016/j.bjoms.2007.09.002

PubMed Abstract | CrossRef Full Text | Google Scholar

McDougall, E. M., Corica, F. A., Boker, J. R., Sala, L. G., Stoliar, G., Borin, J. F., et al. (2006). Construct validity testing of a laparoscopic surgical simulator. J. Am. Coll. Surg. 202, 779–787. doi: 10.1016/j.jamcollsurg.2006.01.004

PubMed Abstract | CrossRef Full Text | Google Scholar

McLaughlin, L., Bond, R., Hughes, C., Mcconnell, J., and Mcfadden, S. (2017). Computing eye gaze metrics for the automatic assessment of radiographer performance during X-ray image interpretation. Int. J. Med. Inf. 105, 11–21. doi: 10.1016/j.ijmedinf.2017.03.001

PubMed Abstract | CrossRef Full Text | Google Scholar

McNatt, S., and Smith, C. (2001). A computer-based laparoscopic skills assessment device differentiates experienced from novice laparoscopic surgeons. Surg. Endosc. 15, 1085–1089. doi: 10.1007/s004640080022

PubMed Abstract | CrossRef Full Text | Google Scholar

Meneghetti, A. T., Pachev, G., Zheng, B., Panton, O. N., and Qayumi, K. (2012). Objective assessment of laparoscopic skills: dual-task approach. Surg. Innov. 19, 452–459. doi: 10.1177/1553350611430673

PubMed Abstract | CrossRef Full Text | Google Scholar

Menekse Dalveren, G. G., and Cagiltay, N. E. (2019). Evaluation of ten open-source eye-movement classification algorithms in simulated surgical scenarios. IEEE Access 7, 161794–161804. doi: 10.1109/access.2019.2951506

CrossRef Full Text | Google Scholar

Mizota, T., Anton, N. E., Huffman, E. M., Guzman, M. J., Lane, F., Choi, J. N., et al. (2020). Development of a fundamentals of endoscopic surgery proficiency-based skills curriculum for general surgery residents. Surg. Endosc. 34, 771–778. doi: 10.1007/s00464-019-06827-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Moorthy, K., Munz, Y., Sarker, S. K., and Darzi, A. (2003). Objective assessment of technical skills in surgery. BMJ 327, 1032–1037. doi: 10.1136/bmj.327.7422.1032

PubMed Abstract | CrossRef Full Text | Google Scholar

Munshi, F., Lababidi, H., and Alyousef, S. (2015). Low-versus high-fidelity simulations in teaching and assessing clinical skills. J. Taibah Univ. Med. Sci. 10, 12–15. doi: 10.1016/j.jtumed.2015.01.008

CrossRef Full Text | Google Scholar

Oostema, J. A., Abdel, M. P., and Gould, J. C. (2008). Time-efficient laparoscopic skills assessment using an augmented-reality simulator. Surg. Endosc. 22, 2621–2624. doi: 10.1007/s00464-008-9844-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Oropesa, I., Sánchez-González, P., Chmarra, M. K., Lamata, P., Fernández, A., Sánchez-Margallo, J. A., et al. (2013). EVA: laparoscopic instrument tracking based on endoscopic video analysis for psychomotor skills assessment. Surg. Endosc. 27, 1029–1039. doi: 10.1007/s00464-012-2513-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Oropesa, I., Sánchez-González, P., Lamata, P., Chmarra, M. K., Pagador, J. B., Sánchez-Margallo, J. A., et al. (2011). Methods and tools for objective assessment of psychomotor skills in laparoscopic surgery. J. Surg. Res. 171, e81–e95.

Google Scholar

Perrenot, C., Perez, M., Tran, N., Jehl, J.-P., Felblinger, J., Bresler, L., et al. (2012). The virtual reality simulator dV-Trainer® is a valid assessment tool for robotic surgical skills. Surg. Endosc. 26, 2587–2593. doi: 10.1007/s00464-012-2237-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Piccardi, L., De Luca, M., Nori, R., Palermo, L., Iachini, F., and Guariglia, C. (2016). Navigational style influences eye movement pattern during exploration and learning of an environmental map. Front. Behav. Neurosci. 10:140. doi: 10.3389/fnbeh.2016.00140

PubMed Abstract | CrossRef Full Text | Google Scholar

Reiley, C. E., Lin, H. C., Yuh, D. D., and Hager, G. D. (2011). Review of methods for objective surgical skill evaluation. Surg. Endosc. 25, 356–366. doi: 10.1007/s00464-010-1190-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Resnick, R., Taylor, B., and Maudsley, R. (1991). In-training evaluation—it’s more than just a form. Ann. R. Coll. Phys. Surg. Can. 24, 415–420.

Google Scholar

Richstone, L., Schwartz, M. J., Seideman, C., Cadeddu, J., Marshall, S., and Kavoussi, L. R. (2010). Eye metrics as an objective assessment of surgical skill. Ann. Surg. 252, 177–182. doi: 10.1097/sla.0b013e3181e464fb

PubMed Abstract | CrossRef Full Text | Google Scholar

Rosenthal, R., Cooper, H., and Hedges, L. (1994). Parametric measures of effect size. Handb. Res. Synth. 621, 231–244.

Google Scholar

Rosser, J. C. Jr., Rosser, L. E., and Savalgi, R. S. (1998). Objective evaluation of a laparoscopic surgical skill program for residents and senior surgeons. Arch. Surg. 133, 657–661. doi: 10.1001/archsurg.133.6.657

PubMed Abstract | CrossRef Full Text | Google Scholar

Schreuder, H. W., Van Dongen, K. W., Roeleveld, S. J., Schijven, M. P., and Broeders, I. A. (2009). Face and construct validity of virtual reality simulation of laparoscopic gynecologic surgery. Am. J. Obstet. Gynecol. 200, 540.e1–540.e8.

Google Scholar

Seagull, F. J. (2012). Human factors tools for improving simulation activities in continuing medical education. J. Contin. Educ. Health Prof. 32, 261–268. doi: 10.1002/chp.21154

PubMed Abstract | CrossRef Full Text | Google Scholar

Shetty, S., Panait, L., Baranoski, J., Dudrick, S. J., Bell, R. L., Roberts, K. E., et al. (2012). Construct and face validity of a virtual reality–based camera navigation curriculum. J. Surg. Res. 177, 191–195. doi: 10.1016/j.jss.2012.05.086

PubMed Abstract | CrossRef Full Text | Google Scholar

Silvennoinen, M., Mecklin, J.-P., Saariluoma, P., and Antikainen, T. (2009). Expertise and skill in minimally invasive surgery. Scand. J. Surg. 98, 209–213. doi: 10.1177/145749690909800403

PubMed Abstract | CrossRef Full Text | Google Scholar

Stuijfzand, B. G., Van Der Schaaf, M. F., Kirschner, F. C., Ravesloot, C. J., Van Der Gijp, A., and Vincken, K. L. (2016). Medical students’ cognitive load in volumetric image interpretation: insights from human-computer interaction and eye movements. Comput. Hum. Behav. 62, 394–403. doi: 10.1016/j.chb.2016.04.015

CrossRef Full Text | Google Scholar

Tien, T., Pucher, P. H., Sodergren, M. H., Sriskandarajah, K., Yang, G.-Z., and Darzi, A. (2014). Eye tracking for skills assessment and training: a systematic review. J. Surg. Res. 191, 169–178. doi: 10.1016/j.jss.2014.04.032

PubMed Abstract | CrossRef Full Text | Google Scholar

Tien, T., Pucher, P. H., Sodergren, M. H., Sriskandarajah, K., Yang, G.-Z., and Darzi, A. (2015). Differences in gaze behaviour of expert and junior surgeons performing open inguinal hernia repair. Surg. Endosc. 29, 405–413. doi: 10.1007/s00464-014-3683-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Topalli, D., and Cagiltay, N. E. (2019). Classification of intermediate and novice surgeons’ skill assessment through performance metrics. Surg. Innov. 26, 621–629. doi: 10.1177/1553350619853112

PubMed Abstract | CrossRef Full Text | Google Scholar

Tricoche, L., Ferrand-Verdejo, J., Pélisson, D., and Meunier, M. (2020). Peer presence effects on eye movements and attentional performance. Front. Behav. Neurosci. 13:280. doi: 10.3389/fnbeh.2019.00280

PubMed Abstract | CrossRef Full Text | Google Scholar

Tsai, M.-J., Huang, L.-J., Hou, H.-T., Hsu, C.-Y., and Chiou, G.-L. (2016). Visual behavior, flow and achievement in game-based learning. Comput. Educ. 98, 115–129. doi: 10.1016/j.compedu.2016.03.011

CrossRef Full Text | Google Scholar

Uemura, M., Jannin, P., Yamashita, M., Tomikawa, M., Akahoshi, T., Obata, S., et al. (2016). Procedural surgical skill assessment in laparoscopic training environments. Int. J. Comput. Assist. Radiol. Surg. 11, 543–552. doi: 10.1007/s11548-015-1274-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Uhrich, M., Underwood, R., Standeven, J., Soper, N., and Engsberg, J. (2002). Assessment of fatigue, monitor placement, and surgical experience during simulated laparoscopic surgery. Surg. Endosc. 16, 635–639. doi: 10.1007/s00464-001-8151-5

PubMed Abstract | CrossRef Full Text | Google Scholar

van der Lans, R., Wedel, M., and Pieters, R. (2011). Defining eye-fixation sequences across individuals and tasks: the Binocular-Individual Threshold (BIT) algorithm. Behav. Res. Methods 43, 239–257. doi: 10.3758/s13428-010-0031-2

PubMed Abstract | CrossRef Full Text | Google Scholar

van Empel, P. J., Van Rijssen, L. B., Commandeur, J. P., Verdam, M. G., Huirne, J. A., Scheele, F., et al. (2012). Validation of a new box trainer-related tracking device: the TrEndo. Surg. Endosc. 26, 2346–2352. doi: 10.1007/s00464-012-2187-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Van Sickle, K. R., Ritter, E. M., and Smith, C. D. (2006). The pretrained novice: using simulation-based training to improve learning in the operating room. Surg. Innov. 13, 198–204. doi: 10.1177/1553350606293370

PubMed Abstract | CrossRef Full Text | Google Scholar

Vickers, J. (1995). Gaze control in basketball foul shooting. Stud. Vis. Inf. Process. 6, 527–541. doi: 10.1016/s0926-907x(05)80044-3

CrossRef Full Text | Google Scholar

Wanzel, K. R., Ward, M., and Reznick, R. K. (2002). Teaching the surgical craft: from selection to certification. Curr. Probl. Surg. 39, 583–659. doi: 10.1067/mog.2002.123481

PubMed Abstract | CrossRef Full Text | Google Scholar

Wilson, M., Mcgrath, J., Vine, S., Brewer, J., Defriend, D., and Masters, R. (2010). Psychomotor control in a virtual laparoscopic surgery training environment: gaze control parameters differentiate novices from experts. Surg. Endosc. 24, 2458–2464. doi: 10.1007/s00464-010-0986-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Yarbus A. L. (ed.). (1967). “Eye movements during perception of complex objects,” in Eye Movements and Vision, (Boston, MA: Springer), 171–211. doi: 10.1007/978-1-4899-5379-7_8

CrossRef Full Text | Google Scholar

Zhang, X., Jiang, S., Ordóñez De Pablos, P., Lytras, M. D., and Sun, Y. (2017). How virtual reality affects perceived learning effectiveness: a task–technology fit perspective. Behav. Inf. Technol. 36, 548–556. doi: 10.1080/0144929x.2016.1268647

CrossRef Full Text | Google Scholar

Zheng, B., Jiang, X., and Atkins, M. S. (2015). Detection of changes in surgical difficulty evidence from pupil responses. Surg. Innov. 22, 629–635. doi: 10.1177/1553350615573582

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: eye movement events, eye movement classification, eye tracking, surgical skill assessment, surgical training

Citation: Menekse Dalveren GG and Cagiltay NE (2020) Distinguishing Intermediate and Novice Surgeons by Eye Movements. Front. Psychol. 11:542752. doi: 10.3389/fpsyg.2020.542752

Received: 16 March 2020; Accepted: 17 August 2020;
Published: 10 September 2020.

Edited by:

Silvia Riva, St Mary’s University, Twickenham, United Kingdom

Reviewed by:

Paola Iannello, Catholic University of the Sacred Heart, Italy
Bin Zheng, University of Alberta, Canada
Trevor Jeremy Crawford, Lancaster University, United Kingdom

Copyright © 2020 Menekse Dalveren and Cagiltay. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Gonca Gokce Menekse Dalveren, goncagmenekse@gmail.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.