- 1Michael G. DeGroote School of Medicine, McMaster University, Hamilton, ON, Canada
- 2Queen's School of Medicine, Queen's University, Kingston, ON, Canada
- 3Department of Ophthalmology and Vision Sciences, University of Toronto, Toronto, ON, Canada
- 4Donald K. Johnson Eye Institute, Krembil Research Institute, University Health Network, Toronto, ON, Canada
Introduction: Glaucoma is a leading cause of irreversible blindness, and its rising global prevalence has led to a significant increase in glaucoma surgeries. However, predicting postoperative outcomes remains challenging due to the complex interplay of patient factors, surgical techniques, and postoperative care. Artificial intelligence (AI) has emerged as a promising tool for enhancing predictive accuracy in clinical decision-making.
Methods: This systematic review was conducted to evaluate the current evidence on the use of AI to predict surgical outcomes in glaucoma patients. A comprehensive search of Medline, Embase, Web of Science, and Scopus was performed. Studies were included if they applied AI models to glaucoma surgery outcome prediction.
Results: Six studies met inclusion criteria, collectively analyzing 4,630 surgeries. A variety of algorithms were applied, including random forests, support vector machines, and neural networks. Overall, AI models consistently outperformed traditional statistical approaches, with the best-performing model achieving an accuracy of 87.5%. Key predictors of outcomes included demographic factors (e.g., age), systemic health indicators (e.g., smoking status and body mass index), and ophthalmic parameters (e.g., baseline intraocular pressure, central corneal thickness, mitomycin C use).
Discussion: While AI models demonstrated superior performance to traditional statistical approaches, the lack of external validation and standardized surgical success definitions limit their clinical applicability. This review highlights both the promise and the current limitations of artificial intelligence in glaucoma surgery outcome prediction, emphasizing the need for prospective, multicenter studies, publicly available datasets, and standardized evaluation metrics to enhance the generalizability and clinical utility of future models.
Systematic review registration: https://www.crd.york.ac.uk/PROSPERO/view/CRD42024621758, identifier: CRD42024621758.
Introduction
Glaucoma remains a leading cause of irreversible blindness worldwide, with 60.5 million individuals affected in 2010 and global prevalence expected to reach 111.8 million by 2040 (Flaxman et al., 2017). As the global burden of disease grows, the number of glaucoma surgeries continues to rise. In the United States alone, the annual volume of glaucoma surgeries significantly increased by 176.7%, rising from 80,151 procedures in 2011 to 221,602 procedures in 2021 (Jayaram et al., 2024).
Although surgery plays a critical role in glaucoma management, postoperative outcomes are notoriously variable, influenced by a complex interplay of patient factors, surgical technique, and postoperative care. Complications such as hypotony maculopathy and bleb-related endophthalmitis pose significant risks to vision and quality of life, while also placing substantial burdens on healthcare systems due to prolonged follow-up, additional interventions, and higher costs (Vijaya et al., 2011; Stokes et al., 2022). This variability has led researchers to develop predictive models aimed at identifying patients at higher risk of complications, traditionally using logistic regression and other conventional statistical techniques (Lavin et al., 1992; Jampel et al., 2001; Lehmann et al., 2000; Parrish et al., 2001).
However, these traditional models are inherently limited by their reliance on linear assumptions and a small set of predefined variables, meaning they often struggle to capture the complex, nonlinear relationships that may drive surgical success or failure. In contrast, artificial intelligence (AI) models offer far more flexible and adaptive approaches. AI models can capture complex, nonlinear relationships in surgical outcomes that traditional statistical methods often overlook (Hashimoto et al., 2018). In addition, they can continuously learn and evolve, incorporating new data over time to improve their predictive accuracy and remain clinically relevant (Huang et al., 2018).
AI has already demonstrated considerable success in other areas of ophthalmology, including automated detection of diabetic retinopathy, prediction of visual field progression, and detection of glaucoma itself using fundus photography and optical coherence tomography (OCT) (Lee and Lee, 2020; Giannini et al., 2019). Its application to surgical outcome prediction, however, represents a more recent development; one that holds great potential to support personalized surgical planning and targeted postoperative care.
The successful translation of these predictive models into routine surgical care depends not only on their accuracy but also on their clinical interpretability, generalizability across patient populations, and external validation. To better understand the current state of this evolving field, we conducted a systematic review to comprehensively evaluate the scientific evidence on AI-driven prediction of postoperative outcomes in glaucoma surgery. The review aims to assess the quality, performance, and limitations of current models, with the ultimate goal of identifying opportunities for improvement and priorities for future research.
Methods
Study design
This systematic review was conducted to evaluate the application of AI in predicting postoperative outcomes following glaucoma surgery. To ensure methodological transparency and comprehensive reporting, we adhered to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines (Moher et al., 2009).
Search strategy
A comprehensive search of the literature was performed across four major databases: Medline, Embase, Web of Science, and Scopus. The search covered the period from inception to November 20, 2024. The search strategy incorporated terms related to AI, machine learning, deep learning, glaucoma surgery, trabeculectomy, and surgical outcomes. Publication date and language were not restricted. The complete search strings for each database are available in Supplementary Table 1.
Eligibility criteria
Studies were eligible for inclusion if they met the following criteria: they were human studies with a cohort, cross-sectional, observational, or randomized controlled trial design; they examined the use of AI to predict or assess postoperative outcomes following glaucoma surgery; and they were published in English. Studies were excluded if they were systematic reviews, scoping reviews, case reports, letters to the editor, abstracts, or conference posters. Animal studies, studies without relevant outcomes or outcome data, and studies published in languages other than English were also excluded.
Screening and study selection
Two independent reviewers (ZK and LK) screened the titles and abstracts of all citations for eligibility based on the predefined inclusion and exclusion criteria. Full-text review was performed for all studies that met initial screening criteria. Any disagreements during either stage of screening were resolved through consensus discussion between the two reviewers. If consensus could not be reached, a third reviewer served as an arbiter. After full-text review, six studies remained and were selected for final inclusion and data extraction. The study selection process is summarized via a PRISMA flow chart in Supplementary Figure 1.
Data extraction
Data extraction was conducted using a standardized spreadsheet developed a priori in Google Sheets (available in Supplementary Table 2). Two independent reviewers (ZK and LK) extracted data from each of the six included studies following a full-text review. All extracted data were compared between reviewers, and discrepancies were resolved through discussion. If necessary, a third reviewer (MB) was consulted to resolve any conflicts.
The extracted data included study authorship, year of publication, study objective, study design, setting (single-center or multicenter), sample size, mean age, and sex distribution of the study population. Data extraction also captured each study's inclusion and exclusion criteria, data source, type of surgical intervention, and definitions of surgical success and failure. Additionally, data on the outcomes predicted by the AI models, the types of algorithms employed, the methods used to develop and validate the models, and the performance of each model were collected. The specific input parameters or features used for prediction were recorded along with any reported feature importance rankings. Key conclusions from each study, any author-reported limitations, and all reported evaluation metrics, including accuracy, area under the receiver operating characteristic curve (AUROC), area under the precision-recall curve (AUPRC), anomaly correlation coefficient (ACC), Matthews correlation coefficient (MCC), F1 score, precision, negative predictive value (NPV), sensitivity, and specificity, were also extracted.
To ensure consistent interpretation of AUROC scores across studies, we applied the classification system proposed by Hosmer and Lemeshow (2000). AUROC values below 0.5 were categorized as “bad,” values from 0.5 to 0.7 as “poor,” values from 0.7 to 0.8 as “acceptable,” values from 0.8 to 0.9 as “excellent,” and values exceeding 0.9 as “outstanding.”
Quality assessment
The methodological quality and risk of bias for all included studies were evaluated using the Joanna Briggs Institute (JBI) Critical Appraisal Checklist for Cohort Studies (Moola et al., 2020). Each study was independently appraised by two reviewers (ZK and LK). Any discrepancies in the initial assessments were discussed until consensus was reached. If necessary, a third reviewer (MB) was consulted.
Data synthesis
A meta-analysis was not performed due to substantial heterogeneity across the included studies, including differences in the types of algorithms and models used, variability in how data and outcomes were presented, and inconsistency in the evaluation metrics applied. Instead, we conducted a narrative synthesis, with a focus on summarizing each study's methodological approach, algorithms, input features, and reported predictive performance.
Study outcomes
This systematic review primarily examined how effectively AI algorithms predict postoperative outcomes in glaucoma surgery. Secondary outcomes included characterization of the specific algorithms and models used, the types of input data and features incorporated into model development, approaches to model training and validation, and the degree to which performance varied across different study designs, definitions of surgical success, and model architectures.
Results
Study characteristics
Our search yielded 91 studies; 39 studies underwent title and abstract screening after duplicate removal; 8 articles were selected for full-text review; 2 articles were removed, leaving a total of 6 studies. Our included studies collectively analyzed 4,630 glaucoma surgeries involving 3,523 unique patients. The median sample size across studies was 218.5 (IQR: 153 to 1,540), ranging from 102 (Barry and Wang, 2024) to 2,398 surgeries (Agnifili et al., 2023), with a median follow-up duration of 1 year (IQR: 0), spanning from 3 months (Barry and Wang, 2024) to 5 years (Birla et al., 2024). In terms of study design, four studies were retrospective cohort studies and two employed prospective cohort designs. Five studies relied on single-center data, with only one drawing from a multicenter database (Agnifili et al., 2023).
In terms of patient demographics, female patients comprised 48.9% (n = 1,723) of the total cohort. The average age across 4 studies reporting means was 68.0 years (pooled SD: 15.9), while 2 studies reported median ages of 74 (Agnifili et al., 2023) and 26 years Birla et al., (2024), respectively. In terms of racial representation across studies, 53.6% of the total sample were White, 19.1% were Asian, 8.9% were non-White Hispanic, 8.6% were unknown or unreported by the authors, 4.2% were Black, and 5.6% were of another race. None of the included studies reported the socioeconomic characteristics of their respective patient populations.
Most studies (n = 4) included patients with various subtypes of glaucoma, with three including both primary and secondary glaucoma. Birla et al. (2024) was the only study to focus exclusively on just one subtype [juvenile open-angle glaucoma (JOAG)]. Barry and Wang (2024) and Lin et al. (2024) did not report the specific diagnoses of 100% and 28% of their respective samples. Overall, four of the studies included patients with primary open-angle glaucoma (POAG), three included patients with pseudoexfoliative glaucoma (PXG), two included patients with JOAG, two included patients with primary angle-closure glaucoma (PACG), one included patients with pigment dispersion glaucoma (PDG), and one included patients with uveitic glaucoma (UG) and neovascular glaucoma (NVG). Additionally, the types of procedures varied among studies. Most (n = 4) focused exclusively on trabeculectomy, one focused on Ahmed valve implantation, and another took a broader approach and encompassed trabeculectomy, ExPress shunts, tube shunts, minimally invasive glaucoma surgery (MIGS), and cyclophotocoagulation (CPC).
The heterogeneity of patient populations, diagnoses, and procedures contributed to differences in success criteria and model features across the studies. Details of the included studies can be found in Table 1.
Algorithms and models
Across the six included studies, over 20 distinct algorithms were tested. The most frequently used algorithms included: random forests (n = 3), support vector machine (SVM, n = 3), decision trees (n = 2), and neural networks (n = 2). Other tested models included extreme gradient boosting (XGBoost), k-nearest neighbors, Bio-Clinical BERT, and Text Transformers, among others (Table 2). This wide range of models highlights both the exploratory nature of the field and the absence of a universally agreed-upon approach for surgical outcome prediction in glaucoma. Overall, the best performing models across studies were random forests (n = 2), decision trees (n = 2), XGBoost (n = 1), and a transformer multilayer neural network architecture (n = 1) (Table 3).
Methodological quality and validation
All included studies were appraised using the JBI Critical Appraisal Checklist for Cohort Studies and were determined to be of high methodological quality (Table 4). Five of the six studies performed internal validation, with most using k-fold cross-validation (k = 3 to 10). External validation, a more rigorous test of generalizability, was reported in only one study (Birla et al., 2024), which tested its model on a separate set of five unrelated patients. However, a validation cohort of this size is insufficient to assess real-world performance across diverse patient populations and overestimates the model's predictive reliability.
Management of class imbalance
Class imbalance, where one outcome (typically surgical failure) is far less common than the other, was directly addressed in three studies: Banna et al. (2022) applied up-sampling to equalize success and failure cases; Barry and Wang (2024) adjusted for imbalance through weighted loss functions, classification threshold optimization, and by tuning for both accuracy and F1 score; and Lin et al. (2024) fine-tuned models using macro-averaged AUROC and F1 score, both of which account for imbalanced data. The lack of standardization in addressing class imbalance (and not addressing it) contributed to variability in reported performance across studies and limits comparability.
Surgical success criteria and outcome definitions
Surgical success was defined differently across the six studies, though most definitions centered on postoperative intraocular pressure (IOP) reduction. Some studies defined success as achieving IOP below an absolute threshold (e.g., <18 mmHg), while others required a percentage reduction from baseline. Beyond IOP, some studies expanded their definition of success to include factors such as avoidance of additional glaucoma surgery or reduced need for medication. Two studies also incorporated visual outcomes, specifically loss of light perception. Of note, the timing of success evaluation ranged from 3 months to 5 years postoperatively, with most studies requiring at least two consecutive follow-up visits before outcomes were classified. These inconsistencies in outcome definitions complicate direct performance comparisons and may have influenced the variability in reported AI model accuracy across studies.
Model performance metrics
All six studies reported AUROC scores to evaluate their models' ability to distinguish between surgical success and failure. Five studies also reported accuracy, sensitivity, and specificity, and four reported precision scores. Additional metrics, including AUPRC (n = 3), F1 score (n = 3), MCC (n = 1), and NPV (n = 1), were reported in some cases.
The highest-performing model was developed by Birla et al. (2024), achieving an accuracy of 87.5% and an AUROC of 0.926. This model used a tree-based algorithm trained on both preoperative and intraoperative variables, including age at diagnosis, baseline intraocular pressure (IOP), duration of preoperative medical treatment, Tenon's thickness, intraoperative mitomycin C (MMC) administration, and scleral fistulation technique. Despite being trained on a relatively small dataset (n = 218), its performance suggests that effective feature selection may be more critical than dataset size alone.
In contrast, Lin et al. (2024) developed the least accurate model, a long short-term memory (LSTM)-based deep learning model trained exclusively on free-text operative notes, which achieved an accuracy of only 40.9%. This low performance is partially explained by the multi-class nature of the prediction task, which categorized outcomes into surgical success, failure due to elevated IOP, or failure due to low IOP. The poor accuracy highlights the challenges of using unstructured text data alone for predictive modeling in glaucoma surgery. However, when the LSTM model incorporated structured electronic health record (EHR) data, accuracy improved to 63.5%, reinforcing the importance of structured data for AI-based surgical outcome prediction.
Sensitivity and specificity varied widely across studies, with three studies reporting higher specificity than sensitivity, while two reported the opposite. Sensitivity ranged from 0.152 (Gaussian Naïve Bayes in Barry and Wang, 2024) to 1.000 (multiple models in Lee et al., 2024), while specificity ranged from 0.089 (decision tree in Barry and Wang, 2024) to 0.926 (Gaussian Naïve Bayes in Barry and Wang, 2024). Such extreme variability in performance metrics likely reflects differences in study populations, feature selection, dataset size, and model architecture. Models trained on single-center datasets generally reported higher accuracy, but this likely reflects overfitting to site-specific data rather than true predictive superiority. The lack of robust external validation in most studies raises concerns about their generalizability in real-world clinical practice.
Discussion
Overview of findings
To our knowledge, this is the first systematic review to evaluate the role of AI in predicting postoperative outcomes in glaucoma surgery. Our findings indicate that AI models hold promise for forecasting surgical outcomes, though performance varied widely across studies. The best-performing models achieved accuracy rates as high as 87.5% (Birla et al., 2024), while the weakest model yielded an accuracy of just 40.9% when predicting multi-class outcomes (Lin et al., 2024). Notably, tree-based algorithms such as random forests consistently ranked among the top-performing models, aligning with broader findings in surgical AI research where tree-based methods have demonstrated robust performance (Hassan et al., 2022).
Despite differences in modeling approaches, certain features emerged as consistently important for predictive accuracy. These included demographic factors (e.g., age), systemic health indicators [e.g., body mass index (BMI), smoking history, comorbid conditions], and ophthalmic parameters [e.g., central corneal thickness (CCT), preoperative intraocular pressure (IOP), preoperative visual acuity (VA), and active ocular medications]. The predominance of structured clinical data suggests that natural language processing (NLP) remains an underutilized tool in glaucoma surgery prediction, with only one study (Lin et al., 2024) incorporating unstructured operative notes. However, significant challenges remain, including the lack of standardized definitions for surgical success and limited external validation, both of which hinder the generalizability and clinical applicability of these models.
This review highlights both the promise and current limitations of AI in glaucoma surgery outcome prediction. While AI models offer a powerful tool for personalized surgical planning and risk stratification, their translation into clinical practice will require addressing key methodological and practical barriers, as explored in the following sections.
Algorithm performance and key determinants of success
Comparing model performance across the included studies was challenging, largely due to variations in dataset size, validation protocols, and performance metrics. Nevertheless, several general observations emerged. First, model performance did not necessarily improve with larger sample sizes. For instance, Birla et al. (2024) achieved high accuracy despite training on just over 200 samples, which may reflect effective feature selection, a well-curated dataset, or possible overfitting. The limited sample size underscores the need for caution when drawing conclusions about model performance. In addition, models that included a broader range of features, particularly systemic health data, often exhibited superior performance (Banna et al., 2022; Lee et al., 2024; Lin et al., 2024).
A more meaningful comparison arises from studies that directly evaluated multiple models under uniform validation conditions. Random forest and XGBoost typically outperformed other algorithms (Banna et al., 2022; Barry and Wang, 2024; Lee et al., 2024; Lin et al., 2024). Barry and Wang (2024) reported that random forest models surpassed both XGBoost and logistic regression across nearly all validation metrics, highlighting the value of capturing nonlinear relationships in clinical data. Although sensitivity, specificity, and precision varied significantly across models, no single metric consistently distinguished a best model, suggesting that clinical priorities and data availability should inform model selection.
Random forests excelled particularly well in small-to-medium-sized datasets and multiclass prediction tasks, often outperforming more complex architectures when training data were limited (Banna et al., 2022; Barry and Wang, 2024; Birla et al., 2024). These ensemble-based models are also known to mitigate overfitting (Banna et al., 2022; Merali et al., 2019) and have shown successful applications in electronic medical record analyses (Merali et al., 2019; Tseng et al., 2020). Indeed, tree-based models have frequently outperformed deep learning methods in tabular medical datasets, particularly when the sample size is modest (Grinsztajn et al., 2022; Shwartz-Ziv and Armon, 2022). However, these findings may not extrapolate to scenarios involving substantially larger datasets or high-dimensional feature modalities such as imaging or textual data, where deep learning often has an advantage (Shwartz-Ziv and Armon, 2022).
Deep learning models, including neural networks, Transformers, and LSTMs, tended to underperform in these studies, likely due to insufficient training samples. For example, Barry and Wang (2024) reported that a neural network trained on over 2,000 cases performed comparably to their top tree-based model, whereas Banna et al. (2022) found that a neural network trained on fewer samples fared worse. In addition, tree-based approaches are inherently less suited to unstructured data (e.g., images or text), where deep learning methods are typically favored. Lin et al. (2024) explored Transformers and LSTMs for analyzing surgeon operative notes, but achieved suboptimal accuracy (<50%). This may reflect either limited training data or an insufficiently discriminative textual feature set. Nonetheless, performance improved when structured EHR data were combined with unstructured text, implying that multimodal approaches may be more effective than relying exclusively on text-based inputs.
Lastly, no studies investigated regression-based approaches to predict continuous variables such as postoperative IOP, a strategy that could potentially yield more nuanced clinical insights. Whether a study adopts a classification or regression framework ultimately depends on clinical priorities and the nature of the available data. While classification models are useful for binary or multiclass outcomes (e.g., success vs. failure), regression models might better capture gradations in postoperative results, thus supporting more personalized surgical planning.
The importance of feature selection
Despite inconsistent reporting of feature importance rankings across studies, we were still able to conclude that feature selection plays a critical role in model performance, a finding that aligns with prior research on AI algorithms (Pudjihartono et al., 2022; Wang et al., 2024). Across our included studies, a range of preoperative, intraoperative, and postoperative variables were found to be of high predictive value. Among preoperative variables, age, baseline IOP, CCT, and active topical medications emerged as key discriminative factors. Age was the most consistently highly ranked variable, though its association with surgical success or failure varied across studies, underscoring its biological relevance even if its predictive direction remains inconsistent. Baseline IOP, a well-established clinical predictor, was heavily relied upon by AI models, with higher preoperative IOP often associated with greater postoperative reduction. This reinforces its importance in preoperative risk stratification and surgical planning. Less traditional factors, such as smoking history and body mass index (BMI), also demonstrated strong discriminative power, suggesting that a comprehensive preoperative assessment, incorporating both established and AI-identified risk factors, could enhance patient selection and surgical planning.
While most models relied on preoperative clinical data, such as age, baseline IOP, CCT, and past ocular surgeries, the inclusion of intraoperative and early postoperative factors further enhanced predictive accuracy. For example, MMC administration, surgical technique, and concurrent cataract extraction were notable predictors in studies that included intraoperative variables. These findings highlight the importance of capturing real-time surgical and recovery data to improve predictive power. Postoperatively, early IOP reduction and visual acuity trends were consistently identified as critical predictors of long-term outcomes. The reliance on early postoperative IOP aligns with clinical evidence that early IOP trends strongly predict surgical success or failure (Esfandiari et al., 2017; Rong et al., 2013). This finding reinforces the importance of close postoperative monitoring and timely interventions, such as early suture lysis or medication adjustments, to mitigate risks of failure. By identifying patients at higher risk based on early IOP trajectories, AI-assisted decision support could enable more proactive and personalized management strategies, ultimately improving long-term outcomes.
Although incorporating intraoperative and early postoperative variables enhances model performance, their inclusion may not always be clinically practical. In real-world applications, if the primary objective of these models is to assess a patient's suitability for surgery, these features would not be available at the time of preoperative evaluation. However, predictive models may also serve a role in immediate postoperative management, where these features could provide valuable insights. Thus, while their inclusion may not be essential for preoperative decision-making, it remains clinically relevant in models designed to support early postoperative monitoring and intervention. Overall, feature selection plays a critical role in model performance, with both traditional and AI-identified risk factors contributing to predictive accuracy.
Implications for surgical practice
AI-driven predictions in glaucoma surgery have the potential to transform clinical practice by enabling personalized surgical planning, targeted postoperative monitoring, and proactive management strategies. By identifying patients at higher risk of surgical failure, such as those with thin corneas, low baseline IOP, or poor conjunctival health, AI models can guide surgeons toward alternative procedures such as tube shunts or adjunctive therapies such as MMC to optimize outcomes. Additionally, insights into intraoperative factors like scleral flap design and Tenon's thickness can help refine surgical techniques, improving wound healing and bleb formation.
Postoperatively, AI models can enhance patient care by flagging high-risk individuals for closer monitoring. For example, patients with poor early IOP control or declining visual acuity may benefit from timely interventions such as early suture lysis, bleb needling, or medication adjustments. This proactive approach can reduce the need for reoperations and improve long-term outcomes. Furthermore, AI-driven risk stratification can support patient counseling, helping clinicians set realistic expectations and tailor treatment plans based on individual risk profiles.
To maximize clinical utility, AI models may be integrated into EHRs to provide real-time decision support during preoperative planning and postoperative care. Specific use scenarios include preoperative decision support to assess surgical candidacy and guide procedure selection, identification of high-risk patients who might benefit from increased monitoring or adjunctive treatments, and adaptive follow-up planning to allocate resources more effectively based on individualized risk stratification. Automated alerts for high-risk patients can prompt clinicians to take proactive measures, while benchmarking surgical outcomes across institutions can identify best practices and areas for improvement. By leveraging AI-driven predictions, clinicians can enhance surgical outcomes, improve patient satisfaction, and reduce the burden of glaucoma-related vision loss.
Limitations
Despite offering valuable insights into the emerging role of AI for predicting surgical outcomes in glaucoma, this review has several notable limitations. First, the small number of included studies, combined with their heterogeneous designs, datasets, and outcome definition, limits the ability to draw definitive conclusions or compare model performance across different clinical settings. Most of the studies were single-center and retrospective, raising concerns about potential overfitting to local patient populations; moreover, nearly all lacked robust external validation, making it difficult to assess generalizability to broader populations. The limited sample sizes in certain studies further amplify the risk of overfitting, especially when complex models, such as deep learning architectures, are applied. In addition, many of the datasets in included studies were imbalanced with respect to key demographic variables such as ethnicity and age, which not only undermines model performance in underrepresented groups but also raises broader concerns in respect to potential biases and inequities. The use of imbalanced data in automated decision-making systems may inadvertently perpetuate or exacerbate existing disparities in healthcare, and as AI models begin to influence clinical decision-making, these ethical implications must be addressed. This underscores the need for transparent model development, demographic diversity in training datasets, and ongoing bias monitoring in real-world implementation.
Second, definitions of “surgical success” and the selection of predictive features varied markedly among the included studies, contributing to inconsistencies in reported performance metrics. Although this variety reflects real-world differences in surgical practice and patient populations, it also underscores the need for standardized outcome definitions and comprehensive multicenter datasets. Additionally, while class imbalance was addressed in some studies, no uniform approach was used, which may obscure the true performance of AI models in predicting rare outcomes. Finally, the exclusion of non-English articles and conference abstracts may have led to publication bias, as relevant unpublished or non–peer-reviewed data were not incorporated. Together, these limitations underscore the need for larger, prospective, and externally validated studies with uniform definitions and standardized data collection to better establish the clinical utility of AI in glaucoma surgery.
Future directions and testable theories
A primary goal for advancing AI in glaucoma surgery is to build comprehensive, high-quality datasets that capture preoperative, intraoperative, and postoperative variables. These data should include clinically relevant predictors, such as baseline IOP, CCT, and early postoperative IOP trends, while also integrating operative notes and imaging modalities like OCT. For example, Birla et al. (2024) demonstrated that a model incorporating intraoperative factors such as mitomycin C administration, Tenon's thickness, and scleral fistulation technique achieved an AUROC of 0.926, suggesting that inclusion of surgical detail can meaningfully enhance predictive accuracy. Similarly, Banna et al. (2022) and Lee et al. (2024) identified systemic health factors such as BMI, smoking history, and medication use as influential, underscoring the need to capture a broad set of patient-level features. Standardizing success and failure definitions across studies will further enable meaningful comparisons and consistent benchmarking of new models. In the included studies, definitions varied significantly, ranging from absolute IOP thresholds to percent reductions and avoidance of additional interventions, introducing variability in performance metrics. Adopting consistent use of the term “success” is critical due to its influence on model training. Thus, prospective clinical trials should adopt uniform, clinically relevant composite endpoints, such as those recommended by the American Academy of Ophthalmology's Glaucoma Preferred Practice Pattern® Committee (Gedde et al., 2025) to support model generalizability. Furthermore, approaches to missing data and class imbalance, such as those methods proven effective in our review [e.g., up-sampling (Banna et al., 2022), weighted loss functions (Barry and Wang, 2024), and macro-averaged metrics (Lin et al., 2024)], should be explicitly reported to mitigate potential biases. In addition, large, multicenter datasets should be made publicly available to foster transparency and external validation while maintaining diligent de-identification protocols for patients' health information. The improved performance seen in single-center datasets in this review may reflect overfitting rather than generalizability, highlighting the importance of robust external testing in future work.
In parallel, there is a need to systematically compare deep learning models to more traditional machine learning methods (e.g., random forests, XGBoost) on large, multimodal datasets that include unstructured text and imaging. While random forests consistently outperformed deep models in most studies reviewed, such as those by Barry and Wang (2024) and Banna et al. (2022), accuracy improved when structured and unstructured data were combined, as demonstrated by Lin et al. (2024), where the model's performance increased from 40.9% to 63.5%. These findings should guide future trials exploring multimodal model development. Furthermore, rigorous validation, particularly external validation on independent, multicenter cohorts, is essential to establish real-world generalizability. Clear reporting of metrics such as AUROC, precision, and F1 score can clarify trade-offs between false positives and negatives, especially in imbalanced datasets. Moving forward, prospective clinical trials should evaluate whether AI-guided risk predictions improve surgeon decision-making and patient outcomes beyond standard care. For instance, models identifying patients at risk based on early postoperative IOP trends, demonstrated in multiple studies as a key predictor, could be tested for their ability to prompt earlier interventions such as suture lysis or medication adjustments. Moreover, real-time AI integration into clinical workflows (e.g., EHR-based alerts) represents a valuable testable theory: does instantaneous feedback to surgeons enhance postoperative results? In the future, this may be assessed through EHR-integrated systems that automatically redact personal health information from model inputs, audit all data accesses, and require clinician confirmation before displaying predictions to ensure patient privacy is maintained. We suggest that future studies attempt to evaluate the impact of AI predictive models on real outcomes in glaucoma patients, such as through RCTs where AI guides treatment decisions.
Another critical challenge is model interpretability. Strategies like SHapley Additive exPlanations (SHAP) can offer transparent insights into feature importance and help clinicians understand model outputs, potentially improving user trust and accelerating adoption. As the field evolves, continuous learning mechanisms may be incorporated to keep models updated with emerging surgical techniques and shifting patient demographics. Researchers could also explore regression-based approaches for predicting continuous outcomes such as exact postoperative IOP, which might provide richer, more personalized insights than binary classifications. Examining the clinical impact of these more granular predictions, especially in dynamic healthcare environments, remains an important direction for future investigations.
In addition to methodological advancements, several practical barriers must be addressed before AI can be deployed for clinical use. These include the need for intuitive, user-friendly interfaces that seamlessly integrate into clinical workflows, interoperability with EHRs, and adequate training for clinical staff to interpret and use AI-generated insights. Without addressing these operational factors, even the most highly accurate models may struggle to gain traction in real-world practice among ophthalmologists. In addition, patient privacy is essential, and federated learning approaches, which enable collaborative model training across institutions without requiring patient-level data sharing, may serve to enhance data privacy and regulatory compliance. “Finally, before AI-based predictive models can be approved for use as medical devices, they must meet regulatory requirements set by authorities such as the United States Food and Drug Administration (FDA) and obtain Conformité Européenne (CE) marking in Europe. For FDA approval, criteria include prospective clinical validation, transparency and interpretability of the model's decision-making process, external validation and demonstration of generalizability, robust risk analysis, a clearly defined intended use (e.g., preoperative decision support, identification of high-risk patients for intensified care, or adaptive follow-up planning), seamless integration into clinical workflows, and adherence to privacy and cybersecurity standards (U.S. Food Drug Administration, 2021, 2022). Similarly, CE marking in Europe requires formal clinical evaluation demonstrating real-world clinical benefit, submission of comprehensive technical documentation, appropriate risk classification, and evidence of generalizability and bias mitigation across diverse patient populations (European Parliament Council of the European Union, 2017).
To accelerate external validation while promoting equitable, generalizable AI development, we encourage glaucoma centers, research institutions, and developers to partner in creating multi-institutional data-sharing consortia. Such collaboration is critical as we look toward building diverse, high-quality datasets and in turn, advancing the responsible translation of predictive models into clinical care.
Conclusion
The application of AI to predicting postoperative outcomes in glaucoma surgery is a promising but immature field. While the reviewed studies demonstrate clear potential, this potential remains limited by small datasets, inconsistent feature selection, and a lack of external validation. To fully realize the benefits of AI, future research must focus on standardization, external validation, prospective evaluation, and clinical impact assessment. Prospective trials evaluating AI integration into real-world surgical decision-making and postoperative care are also essential. Thus, although it is evident that AI already demonstrates potential in this field, without shared standards and robust validation, it remains at the “proof of concept” level. Addressing these methodological gaps will enable AI to fulfill its promise of enhancing personalized glaucoma surgery and improving patient outcomes.
Data availability statement
The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author.
Author contributions
ZK: Data curation, Formal analysis, Investigation, Methodology, Project administration, Visualization, Writing – original draft, Writing – review & editing. LK: Formal analysis, Investigation, Writing – original draft, Writing – review & editing. JB: Formal analysis, Writing – original draft, Writing – review & editing. MB: Conceptualization, Project administration, Supervision, Visualization, Writing – review & editing. DM: Conceptualization, Writing – review & editing.
Funding
The author(s) declare that no financial support was received for the research and/or publication of this article.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Gen AI was used in the creation of this manuscript.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fdata.2025.1605018/full#supplementary-material
References
Agnifili, L., Figus, M., Porreca, A., Brescia, L., Sacchi, M., Covello, G., et al. (2023). A machine learning approach to predict the glaucoma filtration surgery outcome. Sci. Rep. 13:18157. doi: 10.1038/s41598-023-44659-6
Banna, H. U., Zanabli, A., McMillan, B., Lehmann, M., Gupta, S., Gerbo, M., et al. (2022). Evaluation of machine learning algorithms for trabeculectomy outcome prediction in patients with glaucoma. Sci. Rep. 12:2473. doi: 10.1038/s41598-022-06438-7
Barry, S., and Wang, S. Y. (2024). Predicting glaucoma surgical outcomes using neural networks and machine learning on electronic health records. Transl. Vis. Sci. Technol. 13:15. doi: 10.1167/tvst.13.6.15
Birla, S., Varshney, T., Singh, A., Sharma, A., Panigrahi, A., Gupta, S., et al. (2024). Machine learning-assisted prediction of trabeculectomy outcomes among patients of juvenile glaucoma by using 5-year follow-up data. Indian J. Ophthalmol. 72:987. doi: 10.4103/IJO.IJO_2009_23
Esfandiari, H., Pakravan, M., Loewen, N. A., and Yaseri, M. (2017). Predictive value of early postoperative IOP and bleb morphology in Mitomycin-C augmented trabeculectomy. F1000Res 6:837. doi: 10.12688/f1000research.12904.1
European Parliament and Council of the European Union (2017). Regulation (EU) 2017/745 on Medical Devices. Official Journal of the European Union. Available online at: https://eur-lex.europa.eu/eli/reg/2017/745/oj (Accessed May 30, 2025).
Flaxman, S. R., Bourne, R. A., Resnikoff, S., Ackland, P., Braithwaite, T., Cicinelli, M. V., et al. (2017). Global causes of blindness and distance vision impairment 1990–2020: a systematic review and meta-analysis. Lancet Glob. Health 5, e1221–e1234. doi: 10.1016/S2214-109X(17)30393-5
Gedde, S. J., Vinod, K., Bowden, E. C., et al. (2025). Special commentary: reporting clinical endpoints in studies of minimally invasive glaucoma surgery. Ophthalmology 132, 141–153. doi: 10.1016/j.ophtha.2024.07.030
Giannini, H. M., Ginestra, J. C., Chivers, C., Draugelis, M., Hanish, A., Schweickert, W. D., et al. (2019). A machine learning algorithm to predict severe sepsis and septic shock: development, implementation, and impact on clinical practice. Crit. Care Med. 47, 1485–1492. doi: 10.1097/CCM.0000000000003891
Grinsztajn, L., Oyallon, E., and Varoquaux, G. (2022). Why do tree-based models still outperform deep learning on tabular data? arXiv [Preprint]. arXiv:2207.08815. Available online at: https://arxiv.org/abs/2207.08815 (Accessed March 15, 2025).
Hashimoto, D. A., Rosman, G., Rus, D., and Meireles, O. R. (2018). Artificial intelligence in surgery: promises and perils. Ann. Surg. 268, 70–76. doi: 10.1097/SLA.0000000000002693
Hassan, A. M., Rajesh, A., Asaad, M., Nelson, J. A., Coert, J. H., Mehrara, B. J., et al. (2022). A surgeon's guide to artificial intelligence-driven predictive models. Am. Surg. 89:11–19. doi: 10.1177/00031348221103648
Hosmer, D. W., and Lemeshow, S. (2000). Applied Logistic Regression, 2nd Edn. New York, NY: John Wiley & Sons.
Huang, C., Clayton, E. A., Matyunina, L. V., Walker, A., McDonald, L. D., Benigno, B., et al. (2018). Machine learning predicts individual cancer patient responses to therapeutic drugs with high accuracy. Sci. Rep. 8:16444. doi: 10.1038/s41598-018-34753-5
Jampel, H. D., Quigley, H. A., Kerrigan-Baumrind, L. A., Melia, B. M., Friedman, D., and Barron, Y. (2001). Risk factors for late-onset infection following glaucoma filtration surgery. Arch. Ophthalmol. 119, 1001–1008. doi: 10.1001/archopht.119.7.1001
Jayaram, R. H., Maheshwari, A., Sarrafpour, S., Liu, J., and Teng, C. (2024). Trends for glaucoma surgeries in the United States from 2011 and 2021 – a national study. Invest. Ophthalmol. Vis. Sci. 65:4641.
Lavin, M. J., Franks, W. A., Wormald, R. P. L., and Hitchings, R. A. (1992). Clinical risk factors for failure in glaucoma tube surgery: a comparison of three tube designs. Arch. Ophthalmol. 110, 480–485. doi: 10.1001/archopht.1992.01080160058030
Lee, C. S., and Lee, A. Y. (2020). Clinical applications of continual learning machine learning. Lancet Digit. Health 2, e279–e281. doi: 10.1016/S2589-7500(20)30116-6
Lee, S. Y., Lee, D. Y., and Ahn, J. (2024). Evaluation of machine learning approach for surgical results of Ahmed valve implantation in patients with glaucoma. BMC Ophthalmol. 24:248. doi: 10.1186/s12886-024-03510-w
Lehmann, O. J., Bunce, C., Matheson, M. M., Graham, A., and Watson, P. G. (2000). Risk factors for development of post-trabeculectomy endophthalmitis. Br. J. Ophthalmol. 84, 1349–1353. doi: 10.1136/bjo.84.12.1349
Lin, W. C., Chen, A., Song, X., Weiskopf, N. G., Chiang, M. F., and Hribar, M. R. (2024). Prediction of multiclass surgical outcomes in glaucoma using multimodal deep learning based on free-text operative notes and structured EHR data. J. Am. Med. Inform. Assoc. 31, 456–464. doi: 10.1093/jamia/ocad213
Merali, Z. G., Witiw, C. D., Badhiwala, J. H., Wilson, J. R., and Fehlings, M. G. (2019). Using a machine learning approach to predict outcome after surgery for degenerative cervical myelopathy. PLoS ONE 14:e0215133. doi: 10.1371/journal.pone.0215133
Moher, D., Liberati, A., Tetzlaff, J., Altman, D. G., and The PRISMA Group. (2009). Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. PLoS Med. 6:e1000097. doi: 10.1371/journal.pmed.1000097
Moola, S., Munn, Z., Tufanaru, C., Aromataris, E., Sears, K., Sfetcu, R., et al. (2020). “Chapter 7: systematic reviews of etiology and risk,” in JBI Manual for Evidence Synthesis, Eds. E. Aromataris and Z. Munn (Adelaide: JBI).
Parrish, R. K. 2nd, Schiffman, J. C., Feuer, W. J., and Heuer, D. K. (2001). Prognosis and risk factors for early postoperative wound leaks after trabeculectomy with and without 5-fluorouracil. Am. J. Ophthalmol. 132, 633–640. doi: 10.1016/s0002-9394(01)01178-3
Pudjihartono, N., Fadason, T., Kempa-Liehr, A. W., and O'Sullivan, J. M. (2022). A review of feature selection methods for machine learning-based disease risk prediction. Front. Bioinform. 2:927312. doi: 10.3389/fbinf.2022.927312
Rong, S. S., Feng, M. Y., Wang, N., Meng, H., Thomas, R., Fan, S., et al. (2013). Can early postoperative intraocular pressure predict success following mitomycin-C augmented trabeculectomy in primary angle-closure glaucoma. Eye 27, 403–409. doi: 10.1038/eye.2012.291
Shwartz-Ziv, R., and Armon, A. (2022). Tabular data: deep learning is not all you need. Inf. Fusion 81, 84–90. doi: 10.1016/j.inffus.2021.11.011
Stokes, S. M., Scaife, C. L., Brooke, B. S., Glasgow, R. E., Mulvihill, S. J., Finlayson, S. R. G., et al. (2022). Hospital costs following surgical complications: a value-driven outcomes analysis of cost savings due to complication prevention. Ann. Surg. 275, e375-e381. doi: 10.1097/SLA.0000000000004243
Tseng, P.-Y., Chen, Y.-T., Wang, C.-H., Chiu, K.-M., Peng, Y.-S., Hsu, S.-P., et al. (2020). Prediction of the development of acute kidney injury following cardiac surgery by machine learning. Crit. Care 24:478. doi: 10.1186/s13054-020-03179-9
U.S. Food and Drug Administration (2021). Artificial Intelligence/Machine Learning (AI/ML)-Based Software as a Medical Device (SaMD) Action Plan. Available online at: https://www.fda.gov/media/145022/download (Accessed May 30, 2025).
U.S. Food and Drug Administration (2022). Clinical Decision Support Software: Guidance for Industry and Food and Drug Administration Staff. Available online at: https://www.fda.gov/media/109618/download (Accessed May 30, 2025).
Vijaya, L., Manish, P., Ronnie, G., and Shantha, B. (2011). Management of complications in glaucoma surgery. Indian J. Ophthalmol. 59 (Suppl. 1), S131–S140. doi: 10.4103/0301-4738.73689
Keywords: glaucoma, trabeculectomy, glaucoma surgery, artificial intelligence, machine learning, deep learning, surgical outcomes, outcome prediction
Citation: Kailani Z, Kim L, Bierbrier J, Balas M and Mathew DJ (2025) Artificial intelligence for surgical outcome prediction in glaucoma: a systematic review. Front. Big Data 8:1605018. doi: 10.3389/fdata.2025.1605018
Received: 07 April 2025; Accepted: 22 July 2025;
Published: 08 August 2025.
Edited by:
Weihua Yang, Southern Medical University, ChinaReviewed by:
Abhishek Appaji, BMS College of Engineering, IndiaThiago Gonçalves dos Santos Martins, Federal University of São Paulo, Brazil
Copyright © 2025 Kailani, Kim, Bierbrier, Balas and Mathew. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: David J. Mathew, ZGF2aWQubWF0aGV3QG1haWwudXRvcm9udG8uY2E=