Skip to main content

GENERAL COMMENTARY article

Front. Oncol., 18 August 2015
Sec. Neuro-Oncology and Neurosurgical Oncology

Commentary: “Neuropsychological Assessment of Individuals with Brain Tumor: Comparison of Approaches Used in the Classification of Impairment”

  • 1Department of Neuro-Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX, USA
  • 2Behavioural Sciences Unit proudly supported by Kids with Cancer Foundation, Kids Cancer Centre, Sydney Children’s Hospital and School of Women’s and Children’s Health, Faculty of Medicine, University of New South Wales, Sydney, NSW, Australia

A commentary on

Neuropsychological assessment of individuals with brain tumor: comparison of approaches used in the classification of impairment
by Dwan TM, Ownsworth T, Chambers S, Walker DG, Shum DHK. Front Oncol (2015) 5:56. doi: 10.3389/fonc.2015.00056

Neurocognitive functioning (NCF) has been increasingly recognized as an important contributor to quality of life in patients with cerebral neoplasms. Accordingly, studies of NCF in this population have grown tremendously over the past few decades, and performance-based neuropsychological measures assessing diverse NCF domains are now routinely included in brain tumor clinical trials (1). However, little consensus exists regarding the prevalence of NCF impairment in this population, with Dwan et al. noting that rates reported in the literature vary from around 13% to over 90% (2). Numerous factors contribute to the widely differing impairment rates across studies, including tumor characteristics (histology, lesion size, and location) (3), patient demographics (socioeconomic status and age) (4), and treatments received (surgical resection and/or chemoradiation) (5).

The article by Dwan and colleagues highlights method variance as another important contributor to the variability in rates of NCF impairment reported. The authors adeptly point to a substantial lack of consistency in impairment classification schemas throughout the literature. Utilizing a few of the more commonly employed classification strategies in a sample of patients with brain tumors, they also demonstrate that classification of impairment based upon intra-individual comparison to estimated premorbid intellectual function differs from classification using published norms or matched controls. The authors then conclude that “this supports the need to interpret individuals’ neuropsychological test results in the context of their estimated premorbid IQ.” While we agree with Dwan et al. that consideration of premorbid functioning is important when interpreting neuropsychological test results, we believe that the approach utilized in their report may be prone to misclassification for reasons discussed below, and we briefly suggest an alternative that is potentially more statistically sound and robust.

In the analyses under discussion, the authors classify a test score as “impaired” if it falls at least one SD below the patient’s own premorbid intellectual function estimate, as determined by performance on a word-reading task. This method is economical and demands minimal statistical sophistication, making it attractive for use in clinical and research settings alike. However, as the authors briefly acknowledge, the approach risks elevating false positive (Type I) error rates for individuals at the upper tail of the premorbid intellectual distribution, as well as false negative errors (Type II) for those at the lower end of the distribution. Importantly, risk of error tends to increase with the number of tests included in a battery, which are numerous in routine neuropsychological practice, and when comparisons are made at the individual test level rather than domain level. A related issue involves differences between correlations among tests. Since tests of NCF vary in sensitivity, ceiling/floor levels, and domains of specificity, uniform correlations between measures of premorbid intellectual function and other tests are unlikely. This is problematic since it cannot be assumed that low-average estimated premorbid intellectual function entails that visual memory, processing speed, executive function, and performances within other domains should also be low average.

Fortunately, various methods exist to address aspects of these concerns. While a comprehensive delineation of the numerous techniques is beyond the scope of this paper, one popular method involves calculation of regression-based norms from healthy control data (6). With this approach, various variables can be used to build equations predicting expected performances for each test from estimates of premorbid intellectual function, age, education, and other factors known to influence NCF. With this method, the predicted (or expected) score can then be directly compared to the observed performance for an individual patient using inferential statistical analyses.

While this method has its own limitations and is by no means the only viable alternative, it has been shown to improve test sensitivity (6). Additionally, simulation studies show that variations of the technique have low classification error rates, even when control samples are relatively small (7). Some may object that these analyses are complicated and impractical. However, user friendly and freely available computer programs exist to facilitate the use of such methods (8, 9). These approaches can even be employed in studies lacking a healthy control group by substituting summary data gleaned from test manuals or previously published reports. Accordingly, consideration of including these and other novel analytical techniques into future studies regarding classification methodology is warranted, which may facilitate more precise prevalence estimates of NCF in patients with cerebral tumors.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The Associate Editor Kerrie Leanne McDonald and the Reviewer Eng-Siew Koh declare that, despite being affiliated to the same institution as author Joanna E. Fardell, the review process was handled objectively and no conflict of interest exists.

References

1. Gilbert MR, Dignam JJ, Armstrong TS, Wefel JS, Blumenthal DT, Vogelbaum MA, et al. A randomized trial of bevacizumab for newly diagnosed glioblastoma. N Engl J Med (2014) 370:699–708. doi:10.1056/NEJMoa1308573

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Dwan TM, Ownsworth T, Chambers S, Walker DG, Shum DH. Neuropsychological assessment of individuals with brain tumor: comparison of approaches used in the classification of impairment. Front Oncol (2015) 5:56. doi:10.3389/fonc.2015.00056

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Noll KR, Sullaway C, Ziu M, Weinberg JS, Wefel JS. Relationships between tumor grade and neurocognitive functioning in patients with glioma of the left temporal lobe prior to surgical resection. Neuro Oncol (2014) 17:580–7. doi:10.1093/neuonc/nou233

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Schagen SB, Klein M, Reijneveld JC, Brain E, Deprez S, Joly F, et al. Monitoring and optimising cognitive function in cancer patients: present knowledge and future directions. Eur J Cancer Care (2014) 12:29–40. doi:10.1016/j.ejcsup.2014.03.003

CrossRef Full Text | Google Scholar

5. Habets EJ, Kloet A, Walchenbach R, Vecht CJ, Klein M, Taphoorn MJ. Tumour and surgery effects on cognitive functioning in high-grade glioma patients. Acta Neurochir (2014) 156:1451–9. doi:10.1007/s00701-014-2115-8

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Smerbeck AM, Parrish J, Yeh EA, Weinstock-Guttman B, Hoogs M, Serafin D, et al. Regression-based norms improve the sensitivity of the national MS society consensus neuropsychological battery for pediatric multiple sclerosis (NBPMS). Clin Neuropsychol (2012) 26:985–1002. doi:10.1080/13854046.2012.704074

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Crawford JR, Garthwaite PH. Comparing patients’ predicted test scores from a regression equation with their obtained scores: a significance test and point estimate of abnormality with accompanying confidence limits. Neuropsychology (2006) 20:259–71. doi:10.1037/0894-4105.20.3.259

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Crawford JR, Garthwaite PH. Using regression equations built from summary data in the neuropsychological assessment of the individual case. Neuropsychology (2007) 21:611–20. doi:10.1037/0894-4105.21.5.611

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Crawford JR, Garthwaite PH, Ryan K. Comparing a single case to a control sample: testing for neuropsychological deficits and dissociations in the presence of covariates. Cortex (2011) 47:1166–78. doi:10.1016/j.cortex.2011.02.017

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: cognition, neuropsychology, assessment methods, brain tumors, neoplasms

Citation: Noll KR and Fardell JE (2015) Commentary: “Neuropsychological assessment of individuals with brain tumor: comparison of approaches used in the classification of impairment”. Front. Oncol. 5:188. doi: 10.3389/fonc.2015.00188

Received: 22 April 2015; Accepted: 05 August 2015;
Published: 18 August 2015

Edited by:

Kerrie Leanne McDonald, University of New South Wales, Australia

Reviewed by:

Han Shen, Children’s Cancer Institute Australia, Australia
Eng-Siew Koh, University of New South Wales, Australia

Copyright: © 2015 Noll and Fardell. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Kyle R. Noll, knoll@mdanderson.org

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.