Skip to main content

ORIGINAL RESEARCH article

Front. Neurol., 13 October 2022
Sec. Applied Neuroimaging
This article is part of the Research Topic Insights in Applied Neuroimaging: 2021 View all 12 articles

Inter-rater agreement between humans and computer in quantitative assessment of computed tomography after cardiac arrest

  • 1Department of Neurology With Experimental Neurology, Freie Universität Berlin and Humboldt-Universität zu Berlin, Charité—Universitätsmedizin Berlin, Berlin, Germany
  • 2BIH Charité Junior Digital Clinician Scientist Program, Berlin Institute of Health at Charité—Universitätsmedizin Berlin, BIH Biomedical Innovation Academy, Berlin, Germany
  • 3Department of Neuroradiology, Freie Universität Berlin and Humboldt-Universität zu Berlin, Charité—Universitätsmedizin Berlin, Berlin, Germany
  • 4Department of Nephrology and Intensive Care Medicine—Circulatory Arrest Center Berlin, Corporate Member of Freie Universität Berlin and Humboldt-Universität zu Berlin, Charité—Universitätsmedizin Berlin, Berlin, Germany

Background: Head computed tomography (CT) is used to predict neurological outcome after cardiac arrest (CA). The current reference standard includes quantitative image analysis by a neuroradiologist to determine the Gray-White-Matter Ratio (GWR) which is calculated via the manual measurement of radiodensity in different brain regions. Recently, automated analysis methods have been introduced. There is limited data on the Inter-rater agreement of both methods.

Methods: Three blinded human raters (neuroradiologist, neurologist, student) with different levels of clinical experience retrospectively assessed the Gray-White-Matter Ratio (GWR) in head CTs of 95 CA patients. GWR was also quantified by a recently published computer algorithm that uses coregistration with standardized brain spaces to identify regions of interest (ROIs). We calculated intraclass correlation (ICC) for inter-rater agreement between human and computer raters as well as area under the curve (AUC) and sensitivity/specificity for poor outcome prognostication.

Results: Inter-rater agreement on GWR was very good (ICC 0.82–0.84) between all three human raters across different levels of expertise and between the computer algorithm and neuroradiologist (ICC 0.83; 95% CI 0.78–0.88). Despite high overall agreement, we observed considerable, clinically relevant deviations of GWR measurements (up to 0.24) in individual patients. In our cohort, at a GWR threshold of 1.10, this did not lead to any false poor neurological outcome prediction.

Conclusion: Human and computer raters demonstrated high overall agreement in GWR determination in head CTs after CA. The clinically relevant deviations of GWR measurement in individual patients underscore the necessity of additional qualitative evaluation and integration of head CT findings into a multimodal approach to prognostication of neurological outcome after CA.

Introduction

Hypoxic-ischemic encephalopathy (HIE) remains a major cause of death and disability following cardiac arrest (1). Many CA survivors remain with disabling neurological symptoms ranging from cognitive and movement disorders to severe impairments of consciousness (2). Reliable neurological outcome prediction after CA is a challenging task. When deciding on continuation or withdrawal of life sustaining therapies, an accurate prognosis is crucial. Currently, a combination of several prognostic investigations is recommended: repeated neurological examination, electroencephalography (EEG), somatosensory evoked potentials (SSEP), serum neuron-specific enolase (NSE) and brain imaging (3).

Despite efforts for standardization, there is heterogeneity in prognostic performance between raters and centers for most of the tests used. Inter-rater agreement on EEG interpretation is fair to substantial (4) and good on SSEP interpretation for the prediction of poor outcome in comatose patients early after CA (5).

Head CTs are obtained early after CA to rule out intracranial causes for the arrest and later after CA (typically within 3–5 days) to assess the degree of HIE. The current reference standard for assessing head CT imaging is interpretation by a neuroradiologist (“qualitative analysis”). Typical characteristics of HIE include global brain edema frequently visible as sulcal effacement, loss of discrimination between gray and white matter, pseudo-subarachnoid hemorrhage or hypodense cortex and gray matter basal ganglia structures (“reversal sign”) (6). International guidelines recommend using imaging findings of “diffuse and extensive anoxic injury” for prediction of poor neurological outcome without providing a clear definition of this finding (3, 7, 8). Studies on the inter-rater reliability of qualitative assessment of head CT have consequently found poor to moderate agreement on the presence and/or severity of HIE between different raters, sites, and specialties (9, 10).

Quantitative analysis could aid in reducing Inter-rater variability. Determination of the gray-white matter ratio (GWR) by manual placement of small regions of interest (ROI) in different gray and white matter target regions has been evaluated in many mostly retrospective, single center studies (11, 12). Protocols for GWR determination are not fully standardized and differ relevantly between centers (13). Our group recently developed and successfully tested an algorithm for automated GWR assessment that uses linear and non-linear co-registration with MRI-based standard atlases to determine ROIs (14).

Limited data is available on the Inter-rater agreement in GWR determination, especially with respect to different levels of expertise. Moreover, there are no studies that investigate the agreement between conventional human raters and automated assessment by a computer algorithm.

Therefore, this study aims at the following:

1. To assess inter-rater agreement of GWR assessment between 3 human raters with different levels of expertise.

2. To assess the agreement between human raters and automated computer GWR assessment.

3. To identify potential sources of Inter-rater variability.

4. To evaluate the impact of Inter-rater variability on prognostic performance.

Methods

Patients

The study was approved by the local ethics committee. We retrospectively included 353 patients from a previously published cohort of our CA database from the circulatory arrest center of a large academic hospital (15). One hundred eleven patients received native head CTs within the first 7 days after CA. Three patients had technically compromised data files and could not be reanalyzed. Thirteen patients with additional cerebral pathologies such as older ischemic lesions, hemorrhage or severe motion artifacts in imaging were excluded (Supplementary Figure 1). Patients were treated with targeted temperature management (TTM, 33°C for 24 h) according to the guidelines (16). Clinicians were blinded to the results of quantitative CT analysis but had access to the radiologic report. Neurological outcome was assessed by the treating physicians at hospital discharge using the cerebral performance category (CPC) scale and dichotomized in “good” (CPC 1–3) and “poor” outcome (CPC 4–5) for statistical analysis (17).

GWR determination

All raters were blinded to clinical information except for the overall context of CA and blinded to each other's results. Images were rated by a board-certified neuroradiologist with 13 years of clinical expertise, a resident neurologist with 3 years of experience in post-cardiac arrest care and a final year medical student who underwent a short, 3-h training session to assess GWR beforehand but had no clinical experience in the field. Using Horos (Version 3.1.2, The Horos Project, https://horosproject.org), each rater was asked to bilaterally place 10 mm2 ROIs into the putamen and posterior limb of the internal capsule (PLIC).

For automated analysis, we used a previously published algorithm, modified to use circular ROIs instead of atlas maps (14). In summary, CTs are first co-registered in a linear and non-linear mode to a standardized CT template in an MRI-based standard space using the FNIRT and FLIRT functions from FSL (Version 5.0.9, Analysis Group, FMRIB, Oxford, United Kingdom) (18). The reverse transformation fields are then used to automatically place predefined standardized 10 mm2 circular ROIs into the center of target regions in the individual CT spaces. Mean hounsfield units (HU) in the ROIs were measured and GWR was calculated as HU PutamenHU PLIC.

Statistical analyses

For statistical analysis, we used RStudio (Version 1.4.1653, RStudio, Boston, MA) with the pROC-package (19) for receiver operating characteristics (ROC), the ggpubr-package for data visualization and the psych-package (20) for ICC calculation. Intraclass correlation (ICC) (ICC 3, two-way mixed, single measure model) was conducted to measure overall and pairwise Inter-rater agreement on the radiodensity (HU) and GWR between the reference standard neuroradiologist and the other three raters. We also calculated ICC (ICC3k, Two-way mixed, average measure) between a mean of three human raters and the computer.

Intraclass correlation values < 0.2 were considered as poor agreement, between 0.21 and 0.4 as fair agreement, between 0.41 and 0.6 as moderate agreement, 0.61 and 0.80 as good agreement, and values > 0.80 as very good agreement (21). For comparison between the raters, we used the Wilcoxon test for the rating results and the DeLong's test for the AUCs. “Severe HIE”, was assumed if GWR was <1.10 (22). Statistical significance was defined as p<0.05. Confidence intervals (CIs) for neurological outcome prediction were calculated with 95% CIs using the Wilson score method.

Results

Patients

Ninety five patients with native head CTs were eligible for analysis. In the predominantly male (66.3%) study cohort (Table 1), the majority had an OHCA (83.2%) and a non-shockable rhythm (69.5%). Half of the patients had a primary cardiac cause of arrest, the other half either a respiratory (35%) or other cause (intoxication, metabolic, and unknown). Median length of ICU stay was 9 days [IQR 4–23]. At ICU discharge, 56 (59%) patients had a poor neurological outcome (CPC 4–5), 9 of which in unresponsive wakefulness syndrome (UWS; CPC 4), the other 47 dead (CPC 5). CTs were acquired at a median of 4 h (IQR, 1–19.5) after cardiac arrest.

TABLE 1
www.frontiersin.org

Table 1. Baseline data of 95 CA patients.

Inter-rater agreement in GWR determination

Figure 1 illustrates ROI placement in 3 patients with different extent of post hypoxic brain damage. For the measurement of putamen radiodensity, ICC for agreement with the study neuroradiologist (reference standard) was very good for all three raters (student 0.83, neurologist 0.92, computer 0.92). The agreement for PLIC was lower, although still at good levels (student 0.66, neurologist 0.71, computer 0.72; Figure 2). The student had the lowest correlation with the neuroradiologist in both regions (Putamen ICC 0.84, CI 0.78–0.88; PLIC 0.66 0.56–0.75) with differences ranging up to a maximum of 4.4 HU for PLIC. Moreover, the computers' agreement with the average of all human raters was very good (ICC Putamen 0.96, CI 0.95–0.97; PLIC 0.98, CI 0.84–0.92; Figure 3).

FIGURE 1
www.frontiersin.org

Figure 1. ROI Placement during GWR assessment in Putamen and Posterior Limb of the internal Capsule (PLIC) in CTs of three patients after cardiac arrest (A–C). All four raters (Neuroradiologist in red, Neurologist in blue, Student in yellow, and Computer in cyan) and GWRs displayed.

FIGURE 2
www.frontiersin.org

Figure 2. Intraclass correlation (ICC) (95% CI) for the pairwise agreement of three different raters with the study neuroradiologist in GWR and ROI assessment (Putamen and PLIC) in Hounsfield Units (HU) in CTs after cardiac arrest.

FIGURE 3
www.frontiersin.org

Figure 3. Intraclass correlation (ICC) (95%-CI) for the combined agreement of three human raters (mean of student, neurologist and neuroradiologist) with the computer algorithm in GWR and ROI assessment (Putamen and PLIC) in Hounsfield Units (HU) in CTs after cardiac arrest.

The raters' overall agreement with the study neuroradiologist on the GWR was very good (overall ICC 0.83; CIs 0.78–0.87), with 95%-CIs ranging into good agreement (Figure 2). Furthermore, the agreement between the computer and the average of all three human raters was very good (0.93; CI 0.91–0.96, Figure 3). However, when comparing individual GWR values between computer and neuroradiologist (ICC 0.84; CI 0.78–0.88), differences ranged from a minimum of 0.003 to a maximum of 0.24 with disagreement across the pre-defined cut-off for severe HIE (GWR < 1.10) in eight poor neurological outcome patients (Figure 4, CT image examples in Supplementary material). Similarly, the neurologist disagreed in six poor neurological outcome cases and the student nine cases with the neuroradiologist (Supplementary Figure 2). No disagreement with the neuroradiologist across the cutoff was observed in any good neurological outcome patient for any rater.

FIGURE 4
www.frontiersin.org

Figure 4. Pairwise GWR values for n = 95 patients rated by computer and neuroradiologist. Blue pairs representing patients with good outcome (CPC 1–3), red pairs poor outcome (CPC 4–5). Triangles for manual rater (neuroradiologist), dots for computer. Blue dotted line at GWR-cutoff = 1.10.

Prognostic variability

Median GWRs were significantly lower in patients with poor neurological outcome than in patients with good neurological outcome, regardless of the rater (Supplementary Table 1). Predictive performance for poor neurological outcome prediction by GWR, quantified by the area under the ROC curve (AUC) was equally good for neuroradiologist and computer (AUC 0.80; CI 0.71–0.89) and lower for the neurologist (0.74; CI 0.65–0.84) and student (0.78; CI 0.69–0.87) (Figure 5). The differences in AUC between all raters were not statistically significant. Sensitivity at GWR < 1.10, a 100% specificity cutoff for all raters ranged between 18% for the computer and was highest at 29% for the neuroradiologist (Supplementary Table 2).

FIGURE 5
www.frontiersin.org

Figure 5. Receiver operating characteristics (ROC) for poor outcome prediction after CA using GWR rated by (A) 3 different human raters (B) Neuroradiologist vs. Computer. Sensitivity and Specificity in %.

Discussion

Our main findings are:

1. The Inter-rater agreement on the Gray-White Matter Ratio measured in head CTs after CA was very good between 3 human raters across different levels of expertise.

2. The Inter-rater agreement on GWR measured in head CTs was very good between human raters and a computer algorithm.

3. Inter-rater agreement was lower for ROIs placed in the posterior limb of the internal capsule than for those placed in the putamen.

4. Despite high overall agreement, considerable deviations in GWR between raters were observed in individual patients. In our cohort, at a GWR threshold of 1.10, this did not lead to any false poor neurological outcome predictions.

In this study investigating the Inter-rater variability of quantitative head CT assessment in patients after cardiac arrest, Inter-rater agreement was considerably better than in studies investigating descriptive, qualitative assessment (9, 10). Previous evidence on the subject had been inconclusive with one study reporting moderate agreement between three human raters and another study reporting good agreement between four human raters (23, 24). We provide the first data for the Inter-rater agreement between human raters and an atlas-based computer algorithm.

We observed the highest variability in the HU measurements of the PLIC-ROI, which mainly accounted for the variability of GWR values. ROI-placement in the Putamen was more consistent. This is illustrated in our visual examples and might be due to a less evident visual delineation of internal capsule in contrast to a clear structural border of the basal ganglia (Figure 1). Another reason could be the increasing difficulty with visually identifying anatomical regions when gray-white matter differentiation is lost due to brain edema (Figure 1A). We also provide evidence that this effect could partly be solved by automated delineation: we observed the highest agreement for the PLIC-ROI between the computer and the neuroradiologist. Considerable GWR deviations or disagreements across the cutoff for HIE occurred both in CTs with extensive as well as subtle pathological changes (Supplementary Figures 3, 4). Three main sources for variability can be observed in these images: First, raters seem to have a preference as to where to place ROIs within the target structures. Future studies with manual placement should therefore standardize ROI placement i.e., by defining the position within the anatomical region and the ROIs relation to other adjacent landmarks. Second, due to small structural focal hypo- or hyperdensitites (e.g., vessels and small lacunar defects) HU values can also vary if ROIs are placed correctly within the anatomic region. We therefore suggest that raters consider the HU value during ROI placement to be in an area representative of the regions radiodensity. Both problems can be solved if instead of circular ROIs, the whole anatomical region is delineated and its radiodensity averaged, an approach that has been successfully used in automated CT quantification by our group an others (14, 25, 26). Third, there are cases where ROIs are misplaced by the rater. Therefore, the results of manual and/or automated ROI placements should be visually re-inspected after qualitative and quantitative analysis are completed to identify cases of misplacement.

Although overall agreement on the GWR values as measure of the degree of edema was very good, we observed considerable variability in individual patients in both neurological outcome groups, the extent of which has not been previously reported. In some cases, this also affected whether a patient was above or below our predefined cut-off for “severe HIE” (GWR < 1.10). In our cohort this exclusively occurred in poor neurological outcome patients, no good neurological outcome patient was misclassified as below the cut-off. Thus, it influenced the sensitivity while retaining the high specificity for poor neurological outcome prediction. It is still possible, however, that the degree of variability observed in some patients of our study might occur in good neurological outcome patients in another cohort. We therefore advise caution for solely relying on the GWR values as prognostic information derived from CT after CA, especially when the values are close to the cut-off. We suggest that CT interpretation should be based on an SOP that integrates both qualitative and quantitative analysis before concluding on the absence or presence of HIE. This information should then always be put into the context of the results of other diagnostic modalities (EEG, SEP, and serum biomarkers).

Despite a tendency toward more consistent ROI placement in experienced human raters, we observed no significant difference in overall prognostic performance between all 4 raters, even though a clinically unexperienced graduate student who was pre-trained for the task was included. The results contrast with the considerable Inter-rater variability of experienced clinicians in qualitative assessment (9) and underscores that CT quantification reduces Inter-rater variability. We therefore recommend a protocol-based training for physicians doing GWR-assessment in future studies and in the clinical routine. Using a standardized automated method could eliminate the problem of Inter-rater variability overall by assisting the rating physician through visually delineating structures or recommending ROIs.

In this study, GWR performed similar in poor neurological outcome prediction as compared to previous studies by our group and others (13, 14, 22). We account the lower sensitivity of 18–28% in this study to the design that did not stratify between early CTs obtained within 24 h and late CTs performed later than 24 h after CA. The majority of CTs included in our study were early CTs. In our cohort, GWR < 1.10 was a safe and reliable cutoff for poor neurological outcome prediction at 100% specificity, regardless of the rater. Using a higher cutoff in our cohort improved sensitivity considerably for some raters without trade-offs in specificity, for instance up to 48% for the neuroradiologist or 42% for the computer (Figure 5) but would have further increased the risk of misclassification of patients when applying to another cohort (Supplementary Figure 2).

Limitations

There are potential additional sources of rater-independent variability such as the type of CT scanner, acquisition parameters or post-processing software (27). We did not examine intra-rater agreement as additional source of variability. Because of the low number of raters, the three human raters' performance is not representative of that of their group (neuroradiologist, neurologist, and student). CT images were used as part of a multimodal approach to decide on continuation or withdrawal of life-sustaining therapy (WLST). Thus, we cannot exclude self-fulfilling prophecy (28). Prognostication was always based on careful consideration of multimodal diagnostics and a considerable observation period. Because neurological outcome was assessed at hospital discharge, we cannot exclude later improvement in patients assigned to the poor neurological outcome group. We therefore assigned CPC 3 (severe neurological deficit) to the good neurological outcome group to prevent overly pessimistic prognosis. As this was a single-center study, the subject should be further studied on larger, prospective cohorts in different clinical settings.

Conclusion

Inter-rater agreement on quantitative head CT analysis after CA was very good in between human raters with different levels of expertise and a computer algorithm. As we observed considerable Inter-rater variability in a few individual patients, we advise caution for solely relying on GWR values as prognostic information derived from CT after CA. The results underscore the need for strategies to further standardize quantitative head CT analysis and for multimodal prognostication in general. Inter-rater variability should be investigated and considered in all future studies of CT quantification.

Data availability statement

The datasets presented in this article are not readily available because of ethical and privacy restrictions. Requests to access the datasets should be directed to the corresponding author.

Ethics statement

The studies involving human participants were reviewed and approved by Ethikkommission der Charité—Universitätsmedizin Berlin. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements.

Author contributions

MK undertook literature research, data collection, manual and automated image analysis, statistical analysis, and drafted the manuscript. ZC rated images and performed statistical analysis. MS rated images. CP, MS, CS, CG, and CL critically reviewed the manuscript. All authors contributed to the article and approved the submitted version.

Funding

MK is participant in the BIH Charité Junior Digital Clinician Scientist Program funded by the Charité—Universitätsmedizin Berlin, and the Berlin Institute of Health at Charité (BIH). CL is participant in the BIH Clinical Fellow Program funded by the Charité—Universitätsmedizin Berlin, and the Berlin Institute of Health at Charité (BIH). Both received a grant from the Laerdal Foundation for research in post cardiac care.

Conflict of interest

Author CL reported institutional fees for lecturing from BD Bard, Zoll, and Bristol Meyer Squibb outside the submitted work. Author CS reported receiving personal fees for consultancy from Philips, BD Bard, Zoll, Rhinocill, and Sedana Medical outside the submitted work.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fneur.2022.990208/full#supplementary-material

References

1. Virani SS, Alonso A, Benjamin EJ, Bittencourt MS, Callaway CW, Carson AP, et al. Heart disease and stroke statistics-2020 update: A report from the american heart association. Circulation. (2020) 141:e139–596. doi: 10.1161/CIR.0000000000000746

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Sandroni C, Cronberg T, Sekhon M. Brain injury after cardiac arrest: pathophysiology, treatment, and prognosis. Intensive Care Med. (2021) 47:1393–414. doi: 10.1007/s00134-021-06548-2

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Nolan JP, Sandroni C, Bottiger BW, Cariou A, Cronberg T, Friberg H, et al. European Resuscitation Council and European Society of Intensive Care Medicine guidelines 2021: post-resuscitation care. Intensive Care Med. (2021) 47:369–421. doi: 10.1007/s00134-021-06368-4

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Westhall E, Rosen I, Rossetti AO, van Rootselaar AF, Wesenberg Kjaer T, Friberg H, et al. Inter-rater variability of EEG interpretation in comatose cardiac arrest patients. Clin Neurophysiol. (2015) 126:2397–404. doi: 10.1016/j.clinph.2015.03.017

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Pfeifer R, Weitzel S, Gunther A, Berrouschot J, Fischer M, Isenmann S, et al. Investigation of the inter-observer variability effect on the prognostic value of somatosensory evoked potentials of the median nerve (SSEP) in cardiac arrest survivors using an SSEP classification. Resuscitation. (2013) 84:1375–81. doi: 10.1016/j.resuscitation.2013.05.016

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Greer DM, Wu O. Neuroimaging in Cardiac Arrest Prognostication. Semin Neurol. (2017) 37:66–74. doi: 10.1055/s-0036-1594253

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Bender A. S1-Leitlinie Hypoxisch-ischämische Enzephalopathie im Erwachsenenalter. Deutsche Gesellschaft für Neurologie. In: Leitlinien für Diagnostik und Therapie in der Neurologie. (2018)

Google Scholar

8. Panchal AR, Bartos JA, Cabanas JG, Donnino MW, Drennan IR, Hirsch KG, et al. Part 3: Adult basic and advanced life support: 2020 american heart association guidelines for cardiopulmonary resuscitation and emergency cardiovascular care. Circulation. (2020) 142:S366–S468. doi: 10.1161/CIR.0000000000000918

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Caraganis A, Mulder M, Kempainen RR, Brown RZ, Oswood M, Hoffman B, et al. Interobserver variability in the recognition of hypoxic-ischemic brain injury on computed tomography soon after out-of-hospital cardiac arrest. Neurocrit Care. (2020) 33:414–21. doi: 10.1007/s12028-019-00900-7

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Beekman R, Maciel CB, Ormseth CH, Zhou SE, Galluzzo D, Miyares LC, et al. Early head CT in post-cardiac arrest patients: A helpful tool or contributor to self-fulfilling prophecy? Resuscitation. (2021) 165:68–76. doi: 10.1016/j.resuscitation.2021.06.004

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Lopez Soto C, Dragoi L, Heyn CC, Kramer A, Pinto R, Adhikari NKJ, et al. Imaging for neuroprognostication after cardiac arrest: systematic review and meta-analysis. Neurocrit Care. (2019) 32:206–216. doi: 10.1007/s12028-019-00842-0

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Zhou F, Wang H, Jian M, Wang Z, He Y, Duan H, et al. Gray-white matter ratio at the level of the basal ganglia as a predictor of neurologic outcomes in cardiac arrest survivors: a literature review. Front Med. (2022) 9:847089. doi: 10.3389/fmed.2022.847089

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Na MK, Kim W, Lim TH, Jang B, Cho Y, Choi KS, et al. Gray matter to white matter ratio for predicting neurological outcomes in patients treated with target temperature management after cardiac arrest: A systematic review and meta-analysis. Resuscitation. (2018) 132:21–8. doi: 10.1016/j.resuscitation.2018.08.024

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Kenda M, Scheel M, Kemmling A, Aalberts N, Guettler C, Streitberger KJ, et al. Automated assessment of brain CT after cardiac arrest-an observational derivation/validation cohort study. Crit Care Med. (2021) 49:e1212–22. doi: 10.1097/CCM.0000000000005198

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Scheel M, Storm C, Gentsch A, Nee J, Luckenbach F, Ploner CJ, et al. The prognostic value of gray-white-matter ratio in cardiac arrest patients treated with hypothermia. Scand J Trauma Resusc Emerg Med. (2013) 21:23. doi: 10.1186/1757-7241-21-23

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Nolan JP, Soar J, Cariou A, Cronberg T, Moulaert VR, Deakin CD, et al. European Resuscitation Council and European Society of Intensive Care Medicine Guidelines for Post-resuscitation Care 2015: Section 5 of the European Resuscitation Council Guidelines for Resuscitation 2015. Resuscitation. (2015) 95:202–22. doi: 10.1016/j.resuscitation.2015.07.018

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Phelps R, Dumas F, Maynard C, Silver J, Rea T. Cerebral Performance Category and long-term prognosis following out-of-hospital cardiac arrest. Crit Care Med. (2013) 41:1252–7. doi: 10.1097/CCM.0b013e31827ca975

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Jenkinson M, Beckmann CF, Behrens TE, Woolrich MW, Smith SM. Fsl. Neuroimage. (2012) 62:782–90. doi: 10.1016/j.neuroimage.2011.09.015

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Robin X, Turck N, Hainard A, Tiberti N, Lisacek F, Sanchez JC, et al. pROC: an open-source package for R and S+ to analyze and compare ROC curves. BMC Bioinformatics. (2011) 12:77. doi: 10.1186/1471-2105-12-77

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Revelle W. Psych: Procedures for Psychological, Psychometric, and Personality Research. Evanston, Illinois: Northwestern University. R package version 2.1.9(R package version 2.1.9). (2021).

Google Scholar

21. Altman DG. Practical Statistics for Medical Research. 1st ed. London, New York: Chapman and Hall. (1991). p. 611.

Google Scholar

22. Streitberger KJ, Endisch C, Ploner CJ, Stevens R, Scheel M, Kenda M, et al. Timing of brain computed tomography and accuracy of outcome prediction after cardiac arrest. Resuscitation. (2019) 145:8–14. doi: 10.1016/j.resuscitation.2019.09.025

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Kirsch K, Heymel S, Gunther A, Vahl K, Schmidt T, Michalski D, et al. Prognostication of neurologic outcome using gray-white-matter-ratio in comatose patients after cardiac arrest. BMC Neurol. (2021) 21:456. doi: 10.1186/s12883-021-02480-6

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Cristia C, Ho ML, Levy S, Andersen LW, Perman SM, Giberson T, et al. The association between a quantitative computed tomography (CT) measurement of cerebral edema and outcomes in post-cardiac arrest-a validation study. Resuscitation. (2014) 85:1348–53. doi: 10.1016/j.resuscitation.2014.05.022

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Wu O, Batista LM, Lima FO, Vangel MG, Furie KL, Greer DM. Predicting clinical outcome in comatose cardiac arrest patients using early noncontrast computed tomography. Stroke. (2011) 42:985–92. doi: 10.1161/STROKEAHA.110.594879

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Hanning U, Bernhard Sporns P, Lebiedz P, Niederstadt T, Zoubi T, Schmidt R, et al. Automated assessment of early hypoxic brain edema in non-enhanced CT predicts outcome in patients after cardiac arrest. Resuscitation. (2016) 104:91–4. doi: 10.1016/j.resuscitation.2016.03.018

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Oh JH, Choi SP, Wee JH, Park JH. Inter-scanner variability in Hounsfield unit measured by CT of the brain and effect on gray-to-white matter ratio. Am J Emerg Med. (2019) 37:680–4. doi: 10.1016/j.ajem.2018.07.016

PubMed Abstract | CrossRef Full Text | Google Scholar

28. Geocadin RG, Peberdy MA, Lazar RM. Poor survival after cardiac arrest resuscitation: a self-fulfilling prophecy or biologic destiny? * Crit Care Med. (2012) 40:979–80. doi: 10.1097/CCM.0b013e3182410146

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: cardiac arrest (CA), neuroprognostication, computed tomography, automated image analysis, resuscitation, inter-rater agreement, brain imaging

Citation: Kenda M, Cheng Z, Guettler C, Storm C, Ploner CJ, Leithner C and Scheel M (2022) Inter-rater agreement between humans and computer in quantitative assessment of computed tomography after cardiac arrest. Front. Neurol. 13:990208. doi: 10.3389/fneur.2022.990208

Received: 09 July 2022; Accepted: 12 August 2022;
Published: 13 October 2022.

Edited by:

Jan Kassubek, University of Ulm, Germany

Reviewed by:

Nico Sollmann, University of California, San Francisco, United States
Kornelia Kreiser, Technical University of Munich, Germany

Copyright © 2022 Kenda, Cheng, Guettler, Storm, Ploner, Leithner and Scheel. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Martin Kenda, martin.kenda@charite.de

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.