Academia.eduAcademia.edu

Valid or not: A critique of Graver and Green

2020, Applied Neuropsychology: Adult

Disagreements in science and medicine are not uncommon, and formal exchanges of disagreements serve a variety of valuable roles. As identified by a Nature Methods editorial entitled "The Power of Disagreement" (2016), disagreements bring attention to best practices so that differences in interpretation do not result from inferior data sets or confirmation bias, "prompting researchers to take a second look at evidence that is not in agreement with their hypothesis, rather than dismiss it as artifacts." Graver and Green published reasons why they disagree with a recent clinical case report and a decades old randomized control trial characterizing the effect of an acute 2 mg dosing of lorazepam on the Word Memory Test. In this article, we formally responded to their commentary to further clarify the reasons for our data interpretations. These two opposing views provide an excellent learning opportunity, particularly for students, demonstrating the importance of careful articulation of the rationale behind certain conclusions from different perspectives. We encourage careful review of the original articles being discussed so the neuropsychologists can read both positions and decide which interpretation of the findings they consider most sound.

Applied Neuropsychology: Adult ISSN: (Print) (Online) Journal homepage: https://www.tandfonline.com/loi/hapn21 Valid or not: A critique of Graver and Green David W. Loring , Kimford J. Meador & Felicia C. Goldstein To cite this article: David W. Loring , Kimford J. Meador & Felicia C. Goldstein (2020): Valid or not: A critique of Graver and Green, Applied Neuropsychology: Adult, DOI: 10.1080/23279095.2020.1798961 To link to this article: https://doi.org/10.1080/23279095.2020.1798961 Published online: 31 Jul 2020. Submit your article to this journal Article views: 6 View related articles View Crossmark data Full Terms & Conditions of access and use can be found at https://www.tandfonline.com/action/journalInformation?journalCode=hapn21 APPLIED NEUROPSYCHOLOGY: ADULT https://doi.org/10.1080/23279095.2020.1798961 Valid or not: A critique of Graver and Green David W. Loringa,b, Kimford J. Meadorc, and Felicia C. Goldsteina a Department of Neurology, Emory University School of Medicine, Atlanta, GA, USA; bDepartment of Pediatrics, Emory University School of Medicine, Atlanta, GA, USA; cDepartment of Neurology & Neurological Sciences, Stanford University School of Medicine, Stanford, CA, USA KEYWORDS ABSTRACT Disagreements in science and medicine are not uncommon, and formal exchanges of disagreements serve a variety of valuable roles. As identified by a Nature Methods editorial entitled “The Power of Disagreement” (2016), disagreements bring attention to best practices so that differences in interpretation do not result from inferior data sets or confirmation bias, “prompting researchers to take a second look at evidence that is not in agreement with their hypothesis, rather than dismiss it as artifacts.” Graver and Green published reasons why they disagree with a recent clinical case report and a decades old randomized control trial characterizing the effect of an acute 2 mg dosing of lorazepam on the Word Memory Test. In this article, we formally responded to their commentary to further clarify the reasons for our data interpretations. These two opposing views provide an excellent learning opportunity, particularly for students, demonstrating the importance of careful articulation of the rationale behind certain conclusions from different perspectives. We encourage careful review of the original articles being discussed so the neuropsychologists can read both positions and decide which interpretation of the findings they consider most sound. Graver and Green (2020) express disagreement with our interpretation of a clinical case study and a randomized clinical trial, both of which included the Word Memory Test (WMT), a performance validity test (PVT) developed and commercialized by one of the commentary authors (i.e., Paul Green). While their commentary did not appear in either of the two journals in which our reports appeared (Archives of Clinical Neuropsychology, The Clinical Neuropsychologist), we appreciate the willingness of the Editors of Applied Neuropsychology: Adult to allow our response to Graver and Green to appear in the same journal. Our goal is not to change their position, but rather to respond formally to inaccuracies and to encourage readers, particularly students, to carefully consider both positions to better understand clinical case formulation from different perspectives. The two primary issues described by Graver and Green are that (1) our clinical patient with an established diagnosis of multiple sclerosis (MS) who failed two computerized performance validity tests displayed a pattern of neuropsychological performance that was inconsistent with that diagnosis, thus supporting the validity test interpretation of poor effort, and (2) the results of our randomized clinical trial examining the acute effects of 2 mg of lorazepam on the WMT reflect invalid performance rather than drug effects. Multiple sclerosis case report Our case report described a woman with an established diagnosis of MS who was referred for neuropsychological CONTACT David W. Loring [email protected] ß 2020 Taylor & Francis Group, LLC Lorazepam; multiple sclerosis; performance validity tests; word memory test evaluation of subjective memory concerns (Loring & Goldstein, 2019). Her findings included a neuropsychological profile consistent with an MS syndrome occurring in the context of failed stand-alone computerized PVTs. In an attempt to reconcile these two superficially inconsistent findings, we “speculate(d) that poor computerized (performance validity test) scores resulted from the disease-related features of MS, although we also discuss approaches to reconcile apparently contradictory performance validity testing (PVT) vs. neuropsychological results if the contributions of diseaserelated cognitive factors on PVTs scores are discounted. This case demonstrates the value of completing the assessment protocol despite obtaining PVT scores below publisher recommended cutoffs in clinical evaluations. If sub-threshold PVT scores are considered evidence of performance invalidity, it is still necessary to have an approach for interpreting seemingly credible neuropsychological test results rather than simply dismissing them as invalid” (p. 1192). One of the diagnostic dilemmas using PVT measures outside of medico-legal evaluations of mild traumatic brain injury (TBI) is the scarcity of appropriate validation studies to guide clinical interpretation. We were misquoted by Graver and Green as saying “there are no available data on the WMT in MS patients,” and actually state that the “WMT has not been systematically studied in MS,” (p. 1200) to point out the lack of appropriate validation in our specific clinical context (McWhirter et al., 2020). Unfortunately, most stand-alone and embedded performance validity measures have not been systematically characterized Emory Brain Health Center, 12 Executive Park, Atlanta, GA 30329, USA. 2 D. W. LORING ET AL. in homogenous neurological diagnostic groups across a spectrum of disease severity, and generalization to these populations needs to be made cautiously while trying to provide the highest quality patient care. After deciding that the general neuropsychological results must be invalid based upon their interpretation of subthreshold WMT scores, Graver and Green seek support for that interpretation without consideration of the consistency of the neuropsychological profile with the MS diagnosis. Given the well-established white matter pathophysiology of MS affecting frontal system efficiency (Grzegorski & Losy, 2017), Graver and Green’s citation of a prolonged Trail Making Part B is an illogical choice to infer evidence of poor effort. Graver and Green also characterize infrequency of poor scores compared to their expectations in support of the inference of poor effort, although the MS studies that they cite are actually studies of Clinically Isolated Syndrome (CIS) rather than MS (Uher et al., 2014; Viterbo et al., 2013). While suggestive of an eventual clinical diagnosis of MS, CIS commonly presents with optic neuritis, or an isolated brainstem or partial spinal-cord syndrome (Loring, 2015), and likely provides an under estimation of cognitive deficits in MS when based upon patterns obtained in CIS patients. While a review of neuropsychological aspects of MS is beyond the scope of this critique, several good reviews are available (Amato et al., 2008; Chiaravalloti & DeLuca, 2008; Oreja-Guevara et al., 2019; Sumowski et al., 2018). All of the embedded PVT measures in our patient’s neuropsychological assessment are based on peer-reviewed studies that characterize her performance levels as in the valid range. These embedded measures included Reliable Digit Span (Greiffenstein et al., 1994), Wisconsin Card Sorting Test (WCST) failure to maintain set errors (Greve et al., 2009), recognition scores for both Logical Memory and Visual Reproduction from the Wechsler Memory ScaleIV (Pearson, 2009), logistic regression estimation using the Rey Auditory Verbal Learning Test (Davis et al., 2012), and Complex Figure recognition (Lu et al., 2003). Although Graver and Green cite references suggesting the possibility of inadequate effort on several of these embedded measures, they are for measures that are often affected in patients with executive function and working memory impairment in MS. We again emphasize that embedded PVT measures have not been systematically examined in MS. Graver and Green present a lengthy discussion of their own unpublished MS data, but it is difficult to evaluate these data since their methodology, findings, and conclusions have not been subjected to a full formal peer review—even basic characteristics of their MS patients are unknown (e.g., disease duration, MRI lesion burden, MS subtype, or whether McDonald criteria (Thompson et al., 2018) were employed). Further, we note that despite information in Applied Neuropsychology: Adult’s Instructions to Authors to include a data availability statement, Graver and Green failed to include any declaration about how the data supporting their assertions could be obtained for independent review. What is lost in the Graver and Green discussion of our case report is the patient’s normal learning and memory scores, which enabled us to explicitly answer the referral question addressing her memory concerns, a fact that provided reassurance during her feedback session. Her normal memory scores on neuropsychological testing are inconsistent with a poor effort inference from WMT. Primum non nocere. Lorazepam effects on WMT Graver and Green spend considerable effort addressing what they perceive as incorrect conclusions in a healthy volunteer study of lorazepam effects on WMT. That study examined the effects of an acute 2 mg dose on WMT in a randomized, double-blind, placebo-controlled, crossover investigation in healthy volunteers published in 2011 in The Clinical Neuropsychologist. We concluded that “these data indicate that multiple WMT measures may be affected by acute (lorazepam) dosing, and provide additional evidence that potential latent variables and their effects on both (symptom validity test) performance and cognitive function should be part of the clinical decision-making process” (p. 799). Regarding the lorazepam study design, we have successfully employed a similar experimental design across multiple studies examining the cognitive side effects of anti-seizure medications (ASMs). Our neuropsychological studies in ASMs have involved a broad range of medications including brivaracetam, carbamazepine, gabapentin, lamotrigine, levetiracetam, lorazepam, phenobarbital, phenytoin, topiramate, valproate, with all studies producing results supporting our a priori experimental hypotheses of differential ASM cognitive side-effects (Meador et al., 1995, 1999, 2005, 2007, 2011, 2019). We also note that this experimental design and study findings have been accepted in well-established peerreview journals. We have conducted two independent studies demonstrating the negative cognitive effects of acute 2 mg administration of lorazepam (Loring et al., 2011; Meador et al., 2011); one of these studies also included neurophysiological measures demonstrating that acute 2 mg lorazepam decreases EEG alpha band peak power and also decreases beta band power reflecting decreased neuronal excitability (Meador et al., 2011). Neurocognitive findings in methodologically rigorous randomized double-blind, placebo-controlled trials cannot be simply dismissed because they do not conform to preconceived biases of the WMT test publisher. As proponents of data sharing, these data were gladly provided to a colleague for post-hoc exploratory analyses, and findings were presented in a poster presentation at the 2013 meeting of the American Society for Clinical Psychopharmacology (Rohling, 2013). While it is beyond the scope of this report to provide a scientific critique of a conference abstract, In brief, the project was an exploratory analysis in which 16 embedded “validity” indices were derived internally from the dataset, with no available validity of the derived “validity” indices themselves. Thus, the meaningfulness of these “validity” measures is unknown. The scatter plot of these findings (see Figure 1) displays Overall Test Battery Mean (OTBM) scores across experimental APPLIED NEUROPSYCHOLOGY: ADULT Figure 1. Scattergram of subjects characterized as having “invalid” peromances, represented by red X’s of Overall Test Battery Mean across trial conditions (Rohling, 2013). conditions, with subjects identified as having “invalid” profiles identified with a red X. It is difficult to reconcile the large number of “invalid” scores with OTBM scores in the normal range, broadly defined as OTBM scores greater than SS ¼ 85. Unfortunately, internally derived measures from this dataset suffer from contamination artifacts that cannot be disentangled. Most scientifically rigorous neuropsychologists will not rely on results from a post-hoc exploratory abstract to refute a peer-reviewed double-blind, placebo-controlled randomized trial, an experimental design considered to reflect the highest level of evidence by multiple systems characterizing evidence quality (e.g., Burns et al., 2011; French & Gronseth, 2008; Straus et al., 2018). Interested readers can download a detailed critique that identifies multiple specific flaws in the Rohling post-hoc analysis (Loring, 2013). Summary We acknowledge that there will be clinicians who disagree with our interpretation, and the MS case report was intended to highlight approaches when attempting to reconcile two superficially discrepant findings. Consequently, we devoted an entire section of our case report to “‘Ontological-Epistemological One-Worldness’ and Choosing Between Apparent Contradictory Results.” Faust’s (2003) provocative article introduces the concept of “ontologicalepistemological one-worldness,” that is, the belief that 3 “careful analysis and synthesis (of seemingly discrepant data) allows one to integrate them into meaningful or orderly results and patterns” (p. 430). Thus, it is easy to dismiss any unexpected variability simply as reflecting poor effort. As we note, however, neuropsychological evaluations often contain cognitive scores that are internally inconsistent (Brooks et al., 2009), and cognitively healthy individuals frequently have scores in the impaired range (Binder et al., 2009). We concluded our case report by stating that “this case suggests that computerized PVTs may be affected in some MS patients by disease-related decreased information processing speed and working memory impairment (emphasis added). Even for those who disagree, there remains a need for viewing the apparently valid neuropsychological test profile within a consistent interpretative framework that addresses superficially incompatible findings within some logical decision-making framework. Different recommendations regarding best practice after obtaining subthreshold scores on PVT measures will vary based upon the clinical context in which the evaluations are being performed (i.e., forensic vs. medical). However, this case demonstrates that credible patterns of neuropsychological performance addressing the primary referral question may be obtained in at least some neurologic patients with subthreshold PVTs scores. Clinical neuropsychological evaluation requires judicious use of clinical judgment, and a ‘one-size fits all’ approach to interpretation and assessment protocol when subthreshold PVT scores are obtained is insufficient to replace good clinical practice” (p. 1200). We appreciate this opportunity to provide lessons in clinical case formulation to our colleagues and peers, and to encourage all to read both positions critically and decide which interpretation of the findings they consider most sound. Articles to which Graver and Green refer, in addition to the 2013 American Society for Clinical Psychopharmacology PowerPoint and poster critique, may be downloaded for educational purposes at http://neurology. emory.edu/faculty/neuropsychology/loring_david.html. Disclosure statement No potential conflict of interest was reported by the author(s). References Amato, M. P., Zipoli, V., & Portaccio, E. (2008). Cognitive changes in multiple sclerosis. Expert Review of Neurotherapeutics, 8(10), 1585–1596. https://doi.org/10.1586/14737175.8.10.1585 Binder, L. M., Iverson, G. L., & Brooks, B. L. (2009). To err is human: “Abnormal” neuropsychological scores and variability are common in healthy adults. Archives of Clinical Neuropsychology : The Official Journal of the National Academy of Neuropsychologists, 24(1), 31–46. https://doi.org/10.1093/arclin/acn001 Brooks, B. L., Iverson, G. L., Feldman, H. H., & Holdnack, J. A. (2009). Minimizing misdiagnosis: Psychometric criteria for possible or probable memory impairment. Dementia and Geriatric Cognitive Disorders, 27(5), 439–450. https://doi.org/10.1159/000215390 Burns, P. B., Rohrich, R. J., & Chung, K. C. (2011). The levels of evidence and their role in evidence-based medicine. Plastic and 4 D. W. LORING ET AL. Reconstructive Surgery, 128(1), 305–310. https://doi.org/10.1097/PRS. 0b013e318219c171 Chiaravalloti, N. D., & DeLuca, J. (2008). Cognitive impairment in multiple sclerosis. The Lancet Neurology, 7(12), 1139–1151. https:// doi.org/10.1016/S1474-4422(08)70259-X Davis, J. J., Millis, S. R., & Axelrod, B. N. (2012). Derivation of an embedded Rey Auditory Verbal Learning Test performance validity indicator. The Clinical Neuropsychologist, 26(8), 1397–1408. https:// doi.org/10.1080/13854046.2012.728627 Faust, D. (2003). Holistic thinking is not the whole story: Alternative or adjunct approaches for increasing the accuracy of legal evaluations. Assessment, 10(4), 428–441. https://doi.org/10.1177/ 1073191103259534 French, J., & Gronseth, G. (2008). Lost in a jungle of evidence: We need a compass. Neurology, 71(20), 1634–1638. https://doi.org/10. 1212/01.wnl.0000336533.19610.1b Graver, C., & Green, P. (2020). Misleading conclusions about word memory test results in multiple sclerosis (MS) by Loring and Goldstein (2019). Adult. Greiffenstein, M. F., Baker, W. J., & Gola, T. (1994). Validation of malingered amnesia measures with a large clinical sample. Psychological Assessment, 6(3), 218–224. https://doi.org/10.1037/ 1040-3590.6.3.218 Greve, K. W., Heinly, M. T., Bianchini, K. J., & Love, J. M. (2009). Malingering detection with the Wisconsin Card Sorting Test in mild traumatic brain injury. The Clinical Neuropsychologist, 23(2), 343–362. https://doi.org/10.1080/13854040802054169 Grzegorski, T., & Losy, J. (2017). Cognitive impairment in multiple sclerosis – a review of current knowledge and recent research. Reviews in the Neurosciences, 28(8), 845–860. https://doi.org/10. 1515/revneuro-2017-0011 Loring, D. W. (2015). INS dictionary of neuropsychology and clinical neurosciences (2nd ed.). Oxford University Press. Loring, D. W., & Goldstein, F. C. (2019). If invalid PVT scores are obtained, can valid neuropsychological profiles be believed? Archives of Clinical Neuropsychology: The Official Journal of the National Academy of Neuropsychologists, 34(7), 1192–1202. https://doi.org/10. 1093/arclin/acz028 Loring, D. W., Marino, S. E., Drane, D. L., Parfitt, D., Finney, G. R., & Meador, K. J. (2011). Lorazepam effects on Word Memory Test performance: A randomized, double-blind, placebo-controlled, crossover trial. The Clinical Neuropsychologist, 25(5), 799–811. https://doi. org/10.1080/13854046.2011.583279 Loring, D. W. (2013). ASCP post-hoc analysis comments. http://neurology.emory.edu/faculty/neuropsychology/loring_david.html. Lu, P. H., Boone, K. B., Cozolino, L., & Mitchell, C. (2003). Effectiveness of the Rey-Osterrieth Complex Figure Test and the Meyers and Meyers recognition trial in the detection of suspect effort. The Clinical Neuropsychologist, 17(3), 426–440. https://doi. org/10.1076/clin.17.3.426.18083 McWhirter, L., Ritchie, C. W., Stone, J., & Carson, A. (2020). Performance validity test failure in clinical populations: A systematic review. Journal of Neurology, Neurosurgery, and Psychiatry. https:// doi.org/10.1136/jnnp-2020-323776 Meador, K. J., Gevins, A., Leese, P. T., Otoul, C., & Loring, D. W. (2011). Neurocognitive effects of brivaracetam, levetiracetam, and lorazepam. Epilepsia, 52(2), 264–272. https://doi.org/10.1111/j.15281167.2010.02746.x Meador, K. J., Gevins, A., Loring, D. W., McEvoy, L. K., Ray, P. G., Smith, M. E., Motamedi, G. K., Evans, B. M., & Baum, C. (2007). Neuropsychological and neurophysiologic effects of carbamazepine and levetiracetam. Neurology, 69(22), 2076–2084. https://doi.org/10. 1212/01.wnl.0000281104.55418.60 Meador, K. J., Loring, D. W., Moore, E. E., Thompson, W. O., Nichols, M. E., Oberzan, R. E., Durkin, M. W., Gallagher, B. B., & King, D. W. (1995). Comparative cognitive effects of phenobarbital, phenytoin, and valproate in healthy adults. Neurology, 45(8), 1494–1499. https://doi.org/10.1212/wnl.45.8.1494 Meador, K. J., Loring, D. W., Ray, P. G., Murro, A. M., King, D. W., Nichols, M. E., Deer, E. M., & Goff, W. T. (1999). Differential cognitive effects of carbamazepine and gabapentin. Epilepsia, 40(9), 1279–1285. https://doi.org/10.1111/j.1528-1157.1999.tb00858.x Meador, K. J., Loring, D. W., Vahle, V. J., Ray, P. G., Werz, M. A., Fessler, A. J., Ogrocki, P., Schoenberg, M. R., Miller, J. M., & Kustra, R. P. (2005). Cognitive and behavioral effects of lamotrigine and topiramate in healthy volunteers. Neurology, 64(12), 2108–2114. https://doi.org/10.1212/01.WNL.0000165994.46777.BE Meador, K. J., Seliger, J., Boyd, A., Razavi, B., Falco-Walter, J., Le, S., & Loring, D. W. (2019). Comparative neuropsychological effects of carbamazepine and eslicarbazepine acetate. Epilepsy & Behavior: E&B, 94, 151–157. https://doi.org/10.1016/j.yebeh.2019.02.034 Oreja-Guevara, C., Ayuso Blanco, T., Brieva Ruiz, L., Hernandez Perez, M., Meca-Lallana, V., & Rami o-Torrenta, L. (2019). Cognitive dysfunctions and assessments in multiple sclerosis. Frontiers in Neurology, 10, 581. https://doi.org/10.3389/fneur.2019.00581 Pearson. (2009). Advanced clinical solutions for WAIS-IV and WMS-IV: Clinical and interpretive manual. Pearson. Rohling, M. L. (2013). When are your trial data real? [Poster presented]. American Society for Clinical Psychopharmacology, Boca Raton, FL. Straus, S., Glasziou, P., Richardson, W. S., & Haynes, R. B. (2018). Evidence-based medicine. Elsivier. https://doi.org/10.1136/ebm.12.1.2-a Sumowski, J. F., Benedict, R., Enzinger, C., Filippi, M., Geurts, J. J., Hamalainen, P., Hulst, H., Inglese, M., Leavitt, V. M., Rocca, M. A., Rosti-Otajarvi, E. M., & Rao, S. (2018). Cognition in multiple sclerosis: State of the field and priorities for the future. Neurology, 90(6), 278–288. https://doi.org/10.1212/wnl.0000000000004977 The power of disagreement. (2016). Nature Methods, 13(3), 185–185. https://doi.org/10.1038/nmeth.3798 Thompson, A. J., Banwell, B. L., Barkhof, F., Carroll, W. M., Coetzee, T., Comi, G., Correale, J., Fazekas, F., Filippi, M., Freedman, M. S., Fujihara, K., Galetta, S. L., Hartung, H. P., Kappos, L., Lublin, F. D., Marrie, R. A., Miller, A. E., Miller, D. H., Montalban, X., … Cohen, J. A. (2018). Diagnosis of multiple sclerosis: 2017 revisions of the McDonald criteria. The Lancet Neurology, 17(2), 162–173. https://doi.org/10.1016/S1474-4422(17)30470-2 Uher, T., Blahova-Dusankova, J., Horakova, D., Bergsland, N., Tyblova, M., Benedict, R. H. B., Kalincik, T., Ramasamy, D. P., Seidl, Z., Hagermeier, J., Vaneckova, M., Krasensky, J., Havrdova, E., & Zivadinov, R. (2014). Longitudinal MRI and neuropsychological assessment of patients with clinically isolated syndrome. Journal of Neurology, 261(9), 1735–1744. https://doi.org/10.1007/s00415-0147413-9 Viterbo, R. G., Iaffaldano, P., & Trojano, M. (2013). Verbal fluency deficits in clinically isolated syndrome suggestive of multiple sclerosis. Journal of the Neurological Sciences, 330(1–2), 56–60. https:// doi.org/10.1016/j.jns.2013.04.004