Showing posts with label WAIS-III. Show all posts
Showing posts with label WAIS-III. Show all posts

Tuesday, March 8, 2016

Research Byte: Comparison of norms from three Spanish-language and one English-language WAIS-III tests (select subtests)

Norm Comparisons of the Spanish-Language and English-Language WAIS-III: Implications for Clinical Assessment and Test Adaptation.  Funes, Cynthia M.; Hernandez Rodriguez, Juventino; Lopez, Steven Regeser.  Psychological Assessment, Mar 7 , 2016, No Pagination Specified. http://dx.doi.org.ezp1.lib.umn.edu/10.1037/pas0000302

Abstract

  1. This study provides a systematic comparison of the norms of 3 Spanish-language Wechsler Adult Intelligence Scales (WAIS–III) batteries from Mexico, Spain, and Puerto Rico, and the U.S. English-language WAIS–III battery. Specifically, we examined the performance of the 4 normative samples on 2 identical subtests (Digit Span and Digit Symbol-Coding) and 1 nearly identical subtest (Block Design). We found that across most age groups the means associated with the Spanish-language versions of the 3 subtests were lower than the means of the U.S. English-language version. In addition, we found that for most age ranges the Mexican subsamples scored lower than the Spanish subsamples. Lower educational levels of Mexicans and Spaniards compared to U.S. residents are consistent with the general pattern of findings. These results suggest that because of the different norms, applying any of the 3 Spanish-language versions of the WAIS–III generally risks underestimating deficits, and that applying the English-language WAIS–III norms risks overestimating deficits of Spanish-speaking adults. There were a few exceptions to these general patterns. For example, the Mexican subsample ages 70 years and above performed significantly better on the Digit Symbol and Block Design than did the U.S. and Spanish subsamples. Implications for the clinical assessment of U.S. Spanish-speaking Latinos and test adaptation are discussed with an eye toward improving the clinical care for this community. (PsycINFO Database Record (c) 2016 APA, all rights reserved)

Monday, June 17, 2013

Error in Dr. James Flynn's (2009) WAIS-IV norming date: Quest blog post by Dr. Dale Watson



This is a guest blog post by Dr. Dale Watson.  The opinions expressed do not necessarily reflect the official position of the ICDP blog or the blogmaster.  However, it is of interest to note that the error Dr. James Flynn (2009) made in reporting the WAIS-IV norming date (here reported by Dr. Dale Watson) is true, and was also in a published review that I received a few days after I received Dr. Watson's guest post.  This second verification source (Kaufman, Dillon, & Kirsch, 2013) will be the subject of my next post.

Dr. Dale Watson's guest post 





In an article entitled, The WAIS-III and WAIS-IV: Daubert motions favor the certainly false over the approximately true, Dr. James Flynn analyzed data from a number of IQ tests, including the WAIS-R, WAIS-III, and WAIS-IV to estimate the rate of the “Flynn Effect” on the Wechsler scales in the U.S. over time.[i] He concluded, as have others, that in order to account for the obsolescence of aging IQ test norms, a “Flynn Effect” adjustment of 0.30 points per year from the date of a tests norming should be applied to the obtained IQ test scores (Flynn, 2009; Fletcher et al., 2010). For example, if the WAIS-III (normed in 1995) was administered to an individual in 2005, the obtained IQ should be downwardly adjusted by 0.30 x 10 or 3.0 points. Thus, an obtained IQ score of 72 would result in a Flynn-adjusted score of 69. Such adjustments have been recommended for use in Atkins evaluations (Flynn, 2009; Gresham & Reschly, 2011; cf Hagen et al., 2010).[ii]

Flynn compared the IQ scores obtained on the WAIS-III and the WAIS-IV in a sample of 240 examinees reported in the Technical and Interpretive Manual for the WAIS-IV (2008).[iii] The Technical Manual reported that the mean IQs differed by 2.9 points with the sample mean for the WAIS-IV being 100 and for the WAIS-III 102.9 (Wechsler, 2008, p. 75). However, because these IQ scores were calculated using different combinations of subtests, Flynn re-calculated the IQ scores utilizing the same combination of 11 subtest scores used on the WAIS-III to calculate the IQs. Flynn (2009) noted, “The list of subtests used to compute Full Scale IQ had not only changed, but had dropped from 11 to 10. But, once again, they gave the comparison group all 11 of the old WAIS-III subtests, and once again that was fortunate because it meant that the true obsolescence of the WAIS-III could be measured. I calculated the total standard score the group got on the same 11 WAIS-III and WAIS-IV subtests. Using the totals and the WAIS-III conversion table, I calculated Full Scale IQs for the two tests” (p. 102). 

In examining Flynn’s Table 2, it appears that these calculations included scores for the Picture Arrangement subtest for both the WAIS-III and WAIS-IV. However, the Picture Arrangement subtest is not included in the WAIS-IV so it is quite unclear how this calculation was performed. Moreover, there is a footnote to this table indicating that the “WAIS-IV estimate is eccentric in carrying over WISC-III subtests (and scoring vs. the WAIS-III tables)…” but the meaning of this statement is also uncertain. In addition, substitution of the Symbol Search subtest for Picture Arrangement appears to yield very similar results.

In any case, the point of this note is not to recalculate Flynn’s estimates but rather to point out what appears to be a discrepancy between WAIS-IV norming date provided by Flynn and that found in the Technical and Interpretive Manual for the WAIS-IV. Flynn indicated that the WAIS-IV was normed in 2006 (Table 1) whereas the Manual reported, “The WAIS-IV normative data was established using a sample collected from March 2007 to April 2008.” [iv] If we use 2007 as the mid-point norming date, the time between the norming of the WAIS-III and WAIS-IV is 12 years and not 11 as provided by Flynn. Using the Flynn 2006 date resulted in a calculated Flynn Effect between the WAIS-III and WAIS-IV of 0.306 points per year (+3.37 / 11 years). Using the norming date provided in the manual resulted in a calculated score of 0.281 points per year (+3.37 / 12 years). It is understood that this discrepancy of just 0.025 points is of little practical significance but it should be noted nonetheless. Moreover, the metaphorical splitting of hairs is not uncommon when discussing the Flynn Effect. Hagan et al. (2010) asserted, “Decades of FE research and testimony… depict the amount of this shift as a moving target. For example, Flynn (1998) once identified the annual shift as 0.25 rather than 0.30, but later testified in Ex Parte Eric Dewayne Cathey (2010) that 0.29 would be appropriate. Schalock et al. (2010) have called for an annual adjustment of 0.33” pp. 1-2.[v] Flynn has acknowledged that the results reported in his report are estimates for the Wechsler scales, writing, “It is quite possible that the rate of gain on Wechsler tests is 0.275 or 0.325 points per year” (Flynn, 2009, p. 104). The recalculation noted here is consistent with this judgment. Further, the weight of the available evidence, including that of a recent meta-analysis, continues to support the Flynn Effect adjustment of 0.3 points per year.[vi]



[i] Flynn, J. R. (2009). The WAIS-III and WAIS-IV: Daubert motions favor the certainly false over the approximately true. Applied Neuropsychology, 16(2), 98-104. doi: 10.1080/09084280902864360
[ii] Gresham, F. M., & Reschly, D. J. (2011). Standard of practice and Flynn Effect testimony in death penalty cases. Intellectual and Developmental Disabilities, 49(3), 131-140. doi: 10.1352/1934-9556-49.3.131
[iii] Wechsler, D. (2008). Wechsler Adult Intelligence Scale: Technical and interpretive manual (4th ed.). San Antonio, TX: Pearson.
[iv] Id., p. 22.
[v] Hagan, L. D., Drogin, E. Y., & Guilmette, T. J. (2010). IQ scores should not be adjusted for the Flynn Effect in capital punishment cases. Journal of Psychoeducational Assessment, 28(5), 474-476. doi: 10.1177/0734282910373343
[vi] Fletcher, J. M., Stuebing, K. K., & Hughes, L. C. (2010). IQ scores should be corrected for the Flynn Effect in high-stakes decisions. Journal of Psychoeducational Assessment, 28(5), 469-473. doi: 10.1177/0734282910373341

Sunday, January 27, 2013

Research Byte: Which is better measure of intelligence? WAIS-III or WAIS-IV

A new article comparing the changes from the WAIS-III to the WAIS-IV with implications for Atkins cases by Taub and Benson. Below is the abstract. Dr. Taub can be contacted via this link.

A previous IAP AP101 report dealing with WAIS-III/WAIS-IV structural changes is worth reading when reviewing this current article.

 

Friday, July 6, 2012

Roberts et al (2005) on Wechslers, WJ III and CHC theory




From the above excellent book. See select comments about the Wechslers, WJ III and CHC theory by Roberts et al. - a very good chapter in an excellent book.



Posted using BlogPress from Kevin McGrew's iPad
www.themindhub.com

Monday, April 2, 2012

CHC theory and the Wechsler IQ scales and test development and interpretation

In 1998 Dr. Dawn Flanagan and I published the Intelligence Test Desk Reference book which was the first thorough treatment of CHC theory (then called Extended Gf-Gc theory). This book is now out-of-print.




We then took the concepts from the ITDR and, together with Dr. Sam Ortiz, presented a cross-battery approach to interpreting the Wechsler batteries.




And again, this book is no longer in print. This also means we no longer receive any $ for sales (conflict of interest disclosure). Table of contents for first three chapters below (click on images to enlarge)






The research, theory, and conceptual material in the second book is nearly identical to the first, but it was presented in the context of how to upgrade interpretation and understanding of the Wechsler batteries according to the CHC framework. Since then the same CHC overview material has been tweaked and updated in a series of CHC cross-battery books by Flanagan et al. But, the foundation of CHC theory, and how it can be integrated within a conceptual framework of test development and interpretation, is largely the same in these newer CHC cross-battery books.

Thus, given that these "mother and father" books are no longer in print, I took the liberty of copying the first three chapters of the Wechsler oriented book and am now making them available for my readers (click here). I make this material available to provide psychologists who have not done much reading regarding CHC theory an opportunity to have access to the basic foundation of CHC theory to help them see how it can be applied to the interpretation of an intelligence battery (in this case the Wechslers). By choosing the Wechsler material this also helps understand how the Wechsler batteries are evolving (either implicitly or explicitly--see Keith and Reynolds, 2010) when viewed from the lens of CHC theory.

But, one must recognize that this material is a bit dated. An update of CHC theory was later published in 2005 (click here to access...plus some other chapters), and was again updated this year by Schneider and McGrew (click here).

However, the CHC chapter I provide in this blog post, particularly when placed in the context of the Wechsler batteries, provides a solid foundation for understanding CHC theory and its impact on contemporary intelligence test development and interpretation. My goal is to increase awareness of CHC theory and its relevance to psychological assessment and interpretation. My goal is to spur others to become more current re: this now dominant framework in the field of applied IQ testing.


- Posted using BlogPress from Kevin McGrew's iPad

Monday, June 13, 2011

Research brief: Gender differences in intelligence on the WAIS-III (Irwing, in press)




There has been no shortage of contemporary research on gender differences in cognitive abilities (click here for prior IQs Corner posts), and g (general intelligence) in particular. Irwing has a new article "in press" that contributes to this literature, both by reinforcing some prior findings...but also being at variance with other. The introduction provides a nice brief overview of some of the reasons (primarily methodological) for difference on the male-female g-difference research.

Double click on images to enlarge.








- iPost using BlogPress from my Kevin McGrew's iPad

Generated by: Tag Generator


Tuesday, March 15, 2011

Does the WAIS-III measure the same intellectual abilities in MR/ID individuals?

I have had a number of people send me copies of this article (see abstracts and journal info below), especially those who do work related to Dx of MR/ID in Atkins death penalty cases.

The abstract is self-explanatory--the authors conclude that the WAIS-III four-factor structure is not validated in an MR/ID population. I can hear a lawyer now--"so Dr. __________, according to MacLean et al. the WAIS-III doesn't measure the same abilities in individuals with MR/ID...so aren't your results questionable?"

A close read of the article suggests the results should be take with a serious grain of salt. In fact, the discussion is primarily a discussion of the various methodological and statistical reasons why the published 4-factor model may not have fit.

As is often the case when dealing with samples of convenience (the authors own words), especially samples of individuals at the lower end of the ability continuum, the variables often show significant problems with non-normality and skew. This is present in this sample. Given that we are dealing with SEM-based statistics, the problem is actually one of not meeting the assumption of multivariate normality. The variables also showed restricted SD's---restricted range of talent, a condition that dampens correlations in a matrix.

While doing extensive modeling research at the Institute for Community Integration at the University of Minnesota, an institute devoted to individuals with MR/ID/DD, I was constantly faced with data sets with these problems. As a result, I was constantly faced with model fit statistics that were much lower than the standard acceptable rules-of-thumbs for model fit statistics...which reflected the less than statistical and distributional robustness of such sample data. The best way to overcome the resultant low model fits (after trying transformations of the variables to different scales), was to compare the fit of competing models. The best fitting model, when compared to competing models, may still show a relatively poor absolute fit value (when compared to the standard rules of thumb), but by demonstrating that it was the best when compared to alternatives, the case could be made that it was still the best possible model given the constraints of the sample data.

This leads to the MAJOR flaw of this study. Although the authors discuss the sample problems above, they only tested one model...the WAIS-III four-factor model. They then looked at the absolute value of the fit statistics and concluded that the 4-factor model was not a good fit. I see this as a major flaw. Since the standard rules-of-thumb for absolute magnitude of fit stats may no longer hold in samples with statistical and distributional problems, they should have specified competing models (e.g., two-factor; CHC-model, single factor, etc.) and then compared the relative model fit statistics before rendering a conclusion.

Finally, as the authors correctly point out, the current results, even with the flaws above, may simply reflect the well-established finding that the differentiation of cognitive abilities is less for lower functioning individuals, and more for higher functioning. This is Spearman's Law of Diminishing Returns (SLODR) [Click here for an interesting recent discussion of SLODR]

Bottom line for the blogmaster--I judge the authors conclusions to be overstated for the reasons noted above, particularly the failure to compare the 4-factor model to alternative models. It is very possible that the 4-factor model may be the best fitting model given the statistical and distributional constraints of the underlying sample data.


Abstract

Intellectual assessment is central to the process of diagnosing an intellectual disability and the assessment process needs to be valid and reliable. One fundamental aspect of validity is that of measurement invariance, i.e. that the assessment measures the same thing in different populations. There are reasons to believe that measurement invariance of the Wechsler scales may not hold for people with an intellectual disability. Many of the issues which may influence factorial invariance are common to all versions of the scales. The present study, therefore, explored the factorial validity of the WAIS-III as used with people with an intellectual disability. Confirmatory factor analysis was used to assess goodness of fit of the proposed four factor model using 13 and 11 subtests. None of the indices used suggested a good fit for the model, indicating a lack of factorial validity and suggesting a lack of measurement invariance of the assessment with people with an intellectual disability. Several explanations for this and implications for other intellectual assessments were discussed.

- iPost using BlogPress from my Kevin McGrew's iPad

Generated by: Tag Generator


Monday, January 31, 2011

IQ test "practice effects"

A practice effect is a major psychometric issue in many Atkins cases, given that both the state and defense often test the defendant with the same IQ battery (most often a Wechsler), and often within a short test-retest interval. Click here to view all ICDP posts that mention practice effects.

Dr. Alan Kaufman has summarized the majority of the literature on practice effects on the Wechslers. He published an article in The Encyclopedia of Intelligence (1994; Edited by Robert Sternberg) that summarized the research prior to the third editions of the Wechsler scales. That article is available on-line (click here).

The most recent summary of the contemporary Wechsler practice effect research is in Lichtenberger and Kaufman (2009) Essentials of WAIS-IV Assessment (p. 306-309). The tables and text provide much about WAIS-IV and some about WAIS-III. The best source for WAIS-III is Kaufman and Lichtenberger, Assessing Adolescent and Adult Intelligence (either the 2002 second edition or the 2006 third edition), especially Tables 6.5 and 6.6 (2006 edition). Below are a few excerpts from the associated text from the 2006 edition

"Practice effects on Wechsler's scales tend to be profound, particularly on the Performance Scale" (p. 202)

"predictable retest gains in IQs" (p.202)

"On the WAIS-III, tests with largest gains are Picture Completion, Object Assembly, and Picture Arrangement"

"Tests with smallest gains are Matrix Reasoning (most novel Gf test), Vocabulary and Comprehension

Block Design improvement most likely due to speed variance--"on second exposure subjects may be able to respond more quickly, thereby gaining in their scores" (p. 204)

One year interval results in far less pronounced practice effects (p. 208).

"The impact of retesting on test performance, whether using the WAIS-III, WAIS-R, other Wechsler scales, or similar tests, needs to be internalized by researchers and clinicians alike. Researchers should be aware of the routine and expected gains of about 2 1/2 points in V-IQ for all ages between 16 and 89 years. They should also internalize the relatively large gain on P-IQ for ages 16-54 (about 8 to 8 1/2 points), andn the fact that this gain in P-IQ swindles in size to less than 6 points for ages 55-74 and less than 4 points for ages 75-889" (p. 209).

"Increases in Performance IQ will typically be about twice as large as increases in Verbal IQ for individuals ages 16 to 54" (p. 209)


Finally, the latest AAIDD manual provides professional guidance on the practice effect.


"The practice effect refers to gains in IQ scores on test of intelligence that result from a person being retested on the same instrument" (p. 38)

"..established clinical practice is to avoid administering the same intelligence test within the same year to the same individual because it will often lead to an overestimate of the examinee's true intelligence" (p. 38).



- iPost using BlogPress from my Kevin McGrew's iPad

Generated by: Tag Generator


Thursday, December 2, 2010

IQ test battery publication timeline: Atkins MR/ID Flynn Effect cheat sheet

As I've become involved in consulting on Atkins MR/ID death penalty cases, a frequent topic raised is that of norm obsolescence (aka, the Flynn Effect). When talking with others I often have trouble spitting out the exact date of publication of the various revisions of tests, as I keep track of more than just the Wechsler batteries (which are the primary IQ tests in Atkins reports). I often wonder if others question my expertise...but most don't realize that there are more IQ batteries out there than just the Wechsler adult battery....and, in particular, a large number of child normed batteries and other batteries spanning childhood and adulthood. Thus, I decided to put together a cheat sheet for myself..one that I could print and have in my files. I put it together in the form of a simple IQ battery publication timeline. Below is an image of the figure. Double click on it to enlarge.

An important point to understand is that when serious discussions start focusing on the Flynn effect in trial's, most often the test publication date is NOT used in the calculation of how obsolete a set of test norms are. Instead, the best estimate of the year the test was normed/standardized is used, which is not included in this figure (you will need to locate this information). For example, the WAIS-R was published in 1981...but the manual states that the norming occurred from May 1976 to May 1980. Thus, in most Flynn effect discussions in court cases, the date of 1978 (middle of the norming period) is typically used. This makes recall of this information difficult for experts who track all the major individually administered IQ batteries.

Hope this helpful...if nothing else...you must admit that it is pretty :)  Click on image to view.






- iPost using BlogPress from my Kevin McGrew's iPad

Tuesday, October 26, 2010

Court Decision: Ramirez v Ryan (AZ, 2010)---follow up FYI

In the previous "So close, yet so far away--Or how the burden of proof matters" guest post by Kevin Foley the Atkins case of Ramirez v AZ raised a number of significant issues such as IQ practice effects, the Flynn effect, the burden of proof, preponderance of the evidence standard, etc (see prior post for more details). 

This past month a decision was made in the Ramirez v Ryan (AZ, 2010) federal habeas corpus case in the  Arizona District Court.  The decision outlines the complexity of a number of interesting and recurring issues (e.g., the Flynn Effect; practice effect) in Atkins cases, and the difficulty possed to the courts by dueling experts with differing opinions.    Due to a lack of time I will not make any additional comments.  This is an FYI post.

Technorati Tags: , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ,

Saturday, October 23, 2010

French WAIS III study supports primary Gq interpretation of Arithmetic in adults

Interesting study with French WAIS III that provides additional support for quantitative knowledge (Gq) being the primary source of variance in understanding the Arithmetic subtest, as well as some processing speed (Gs) in adults. Click here for prior post on this topic.


Rozencwajg, P., Schaeffer, O., & Lefebvre, V. (2010). Arithmetic and aging: Impact of quantitative knowledge and processing speed. Learning and Individual Differences, 20(5), 452-458.

Abstract

The main objective of this study was to examine how quantitative knowledge (Gq in the CHC model) and processing speed (Gs in the CHC model) affect scores on the WAIS-III Arithmetic Subtest (Wechsler, 2000) with aging. Two age groups were compared: 30 young adults and 25 elderly adults. For both age groups, Gq was an important predictor of Arithmetic score variance (R² = 48% and R² = 45%, respectively). However, in line with Salthouse, the results showed that processing speed predicted Arithmetic scores only for the older adults, not for the younger ones (additional 9% of the variance for the elderly vs. 1% of the variance for the young adults). These results can clarify the ambiguous evolution of Arithmetic scores with aging: Arithmetic performance with aging seems to follow an intermediate path between Gc and Gf. This suggests that both Gq and Gs have an impact on Arithmetic in aging.

Additional quotes from the article

Today, “the CHC model (Cattell–Horn–Carroll theory of cognitive abilities) used extensively in applied psychometrics and intelligence testing during the past decade is a consensus model” (McGrew, 2005, p. 149). CHC is a hierarchical model (Fig. 1) with three strata: factor g (Stratum III), broad abilities (Stratum II), and narrow abilities (Stratum I). Broad CHC abilities (Stratum II) include Gf (fluid intelligence/reasoning), Gc (crystallized intelligence/knowledge), Gv (visual–spatial abilities), Gsm (short-term memory), Gs (cognitive processing speed), and Gq (quantitative knowledge). [Click on images to enlarge them]




In contemporary assessments of intelligence (Flanagan & Harrison, 2005), the Cattell–Horn–Carroll Theory (CHC model) plays an important role in interpreting the scores underlying the Wechsler Scale Subtests. There is some controversy, however, as to the constructs measured by each subtest. As stated above, authors disagree on how to classify Arithmetic in this model.

The first hypothesis tested here concerns the role of quantitative knowledge (Gq) in Arithmetic Subtest performance. Gq has been defined as the wealth (breadth and depth) of a person's “acquired store of declarative and procedural quantitative knowledge. Gq is largely acquired through the ‘investment’ of other abilities, primarily during formal educational experiences. It is important to recognize that RQ (narrow ability, Stratum I), which is the ability to reason inductively and deductively when solving quantitative problems, is not included under Gq, but rather is included in the Gf domain (broad ability, Stratum II). Gq represents an individual's store of acquired mathematical knowledge, not reasoning with this knowledge” (McGrew, 2005, p. 156).

Yet when we look at the performance curve with age (see Fig. 2), we can see firstly that the mean scores on Digit Span (Gsm) and Matrix Reasoning – which is a typical test of fluid intelligence (Gf) ([Schroeder and Salthouse, 2004] and [Verhaeghen, 2003]); – start to decline gradually at the age of 25, whereas the mean score on Arithmetic remains stable until age 70. Secondly, the mean score on Vocabulary – which is a typical test of crystallized intelligence (Gc) (Verhaeghen, 2003) – is close to the teenage level (age 16) after the age of 70, whereas performance drops well below that level on Arithmetic. Analyses of age effects on the WAIS-III subtests among American subjects indicate the same phenomena ([Ardila, 2007] and [Ryan et al., 2000]). Finally, Arithmetic performance with aging seems to follow an intermediate path between Gc and Gf (see Fig. 3). This result is similar to that found by Schroeder and Salthouse (2004), see their [Fig. 1] and [Fig. 2] p. 399 and 400): “All the factors were also influenced by knowledge (vocabulary), with the largest knowledge effects on the numeric/fluency factor” (p. 400).



.....the high correlations obtained between the scores on the Arithmetic Subtest and the new quantitative test, both for the young and older adults, support the hypothesis that the Arithmetic Subtest belongs to factor Gq in the CHC model ([Flanagan and Harrison, 2005] and [Flanagan and Kaufman, 2004])





- iPost using BlogPress from my Kevin McGrew's iPad

Friday, July 16, 2010

Cattell-Horn-Carroll (CHC) based intelligence testing in Atkins cases

I've written frequently about the importance of psychologists involved in Atkins cases to get more up-to-date re: contemporary psychometric intelligence theory, especial the consensus Cattell-Horn-Carroll (CHC) theory of intelligence (click here for all prior posts, this one included, so tagged at the ICDP blog).  The gap between contemporary intelligence theory and testing practice in legal settings is troubling and needs to be narrowed.

Today I'm excited to announce the availability of a "taking stock" intelligence test/theory article that supports my frequent and persistent message...and it is NOT by me...but by other highly respected intelligence test research scholars.

I'm excited to announce that the special issue of Psychology in the Schools, Current Research in Cattell-Horn-Carroll-Based Assessment (guest editors where Jocelyn Newton and myself), is now published.  Yippeee.  To be honest, Dr. Newton deserves the major credit....she did all the heavy lifting and I road her coat tails.  Also thanks to Dr. David McIntosh for suggesting and overseeing the special issue

A review of the TOC can be found by clicking here

The article that I highly recommend as must reading is listed below (along with a link to a copy).

Keith, T. & Reynolds, M. (2010).  Cattell-Horn-Carroll abilities and cognitive tests:  What we've learned from 20 years of research.  Psychology in the Schools, 47(4), 2010, 635-650 (also available via click here)
This article reviews factor-analytic research on individually administered intelligence tests from a Cattell-Horn-Carroll (CHC) perspective. Although most new and revised tests of intelligence are based, at least in part, on CHC theory, earlier versions generally were not. Our review suggests that whether or not they were based on CHC theory, the factors derived from both new and previous versions of most tests are well explained by the theory. Especially useful for understanding the theory and tests are cross-battery analyses using multiple measures from multiple instruments. There are issues that need further explanation, of course, about CHC theory and tests derived from that theory. We address a few of these issues including those related to comprehension-knowledge (Gc) and memory factors, as well as issues related to factor retention in factor analysis.
Although the WAIS-IV is not specifically treated in the article, the authors do address the changing (and more CHC-like foundation) of the Wechsler batteries via research on the WISC-III/IV (which the WAIS-IV has been following in terms of structural organization since the WAIS-III/IV).  The inescapable conclusion is that psychologists involved in Atkins cases need to "get current" on CHC theory and CHC-based test interpretation.  It is nice to no longer be the sole voice yelling into the halls of justice.

The article makes reference to a master table of CHC abilities and definitions included in the introduction to this special issue I co-authored with Dr. Newton.  It available here.


Technorati Tags: , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ,

Friday, April 23, 2010

New MR/ID malingering research study: Existing measures not that good for MR/ID

Shandera, A. L., Berry, D. T. R., Clark, J. A., Schipper, L. J., Graue, L. O., & Harp, J. P. (2010). Detection of Malingered Mental Retardation. Psychological Assessment, 22(1), 50-56. (click here to view)

 Emphasis in abstract and conclusion added by IDCP blogmaster


Abstract

In a cross-validation of results from L. O. Graue et al. (2007), standard psychological assessment instruments, as well as tests of neurocognitive and psychiatric feigning, were administered under standard instructions to 24 participants diagnosed with mild mental retardation (MR) and 10 demographically matched community volunteers (CVH). A 2nd group of 25 community volunteers was instructed to malinger MR (CVM) during testing. CVM participants obtained Wechsler Adult Intelligence Scale (3rd ed.; D. Wechsler, 1997) Full Scale Intelligence Quotient scores that were significantly lower than the demographically similar CVH group but comparable to the MR group, suggesting that CVM subjects feigned cognitive impairment. On the basis of standard cutting scores from test manuals or published articles, of the 11 feigning measures administered, only the Test of Memory Malingering (TOMM; T. N. Tombaugh, 1996) retention trial had a specificity rate >.90 in the MR group. However, the 2nd learning trial of the TOMM, as well as a short form of the Digit Memory Test (T. J. Guilmette, K. J. Hart, A. J. Guiliano, & B. E. Leininger, 1994), approached this level of specificity, with both at .88. These results raise concerns about the specificity rates at recommended cutting scores of commonly used feigning tests in defendants with MR.
Conclusion
Overall, although there were some helpful findings in the present study, broadly speaking, neurocognitive feigning measures derived primarily on the basis of traumatic brain injury samples do not seem to work well in patients with MR. Given the fairly consistent reports to this effect in the published literature, it may well be that a new approach is needed to develop feigning detection instruments that are both sensitive and specific for the evaluation of possible malingered MR. Research assessing a broad array of possible approaches to this issue is clearly needed by the forensic clinical community.
Prior ICDP malingering-related posts, inclusive of the current, can be found by clicking here

Technorati Tags: , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ,

Tuesday, April 13, 2010

Guest commentary on the Retrospective Assessment of MR: Effectively Addressing Atkins Questions: Dr. Timothy Derning

Blogmaster introduction and comments:  Below is a guest blog post by Dr. Timothy Derning in response to the recent court ruling regarding Johnston v Fl (click here for prior post that includes links to all prior posts and documents).  This is a longer than usual blog post, but I believe it is worth the space.  Also, I would LOVE to see other professionals (who practice in the area of Atkins cases) offer similar post-hoc analysis of Atkins court decisions.  They can be very educational and instructive.  Such commentaries can serve a valuable function of encouraging discussion, the exchange of ideas, and professional debate.  That being said, if folks would like to engage in further dialogue regarding Dr. Derning's post, I  suggest joining (in not already a member) the public Atkins MR/ID Death Penalty discussion listserv and exchange thoughts, etc. in that medium.  Info regarding the listserv and a sign-up box are near the top of the blogroll on the right-hand side of this blog.  Thanks Dr. Derning for the post.

---------------------------

A brief review of the state of Florida decision regarding David Eugene Johnston dated April 5, 2010. Mr. Johnston is on Florida's death row. The issue before the court was whether Mr. Johnston has mental retardation. The court's decision was that Mr. Johnston does not have mental retardation. The defense presented four experts. The State presented two experts.

Mr. Johnston is 61 years old in 2010. He has been given a number and variety of intelligence (IQ) tests throughout his life, beginning in 1967, at age 7, when he was given the Stanford Binet, form LM, and received a 57 IQ. He was administered the WISC twice, once in 1972, FSIQ = 65 (at age 12); and two years later in 1974, FSIQ = 80 (age 14). It's important to note that the examiner in 1967 made a comment in the report that the IQ score of 57 most likely represented a depressed estimate of intellectual functioning due to an unhealthy home environment, moderate to severe perceptual problems and/or brain damage, and severe emotional disturbance.  This examiner stated Johnston’s intellectual ability and potential was "possibly within the lower dull normal range, normal level." Likewise, the 1972 examiner acknowledged that the WISC FSIQ of 65 was in the retarded range, but stated the results were suspect due to possible emotional problems.  The 1972 examiner estimated that Johnston’s ability "would be more in keeping with the slow learner or low average range rather than the mentally retarded." In 1974 the evaluator who administered a WISC and reported a FSIQ of 80, commented that young Mr. Johnston (then age 14) was cooperative and engaged. These comments played a significant role in this court's opinion.

Before proceeding its is worthwhile to pause and recall that the definition of mental retardation (aka intellectual disability) has three parts or prongs: subaverage intellectual functioning (a valid IQ score of approximately 70 + 5 points; 1 SEM); demonstrated deficits in daily adaptive living ability, and onset that begins during development (before age 18). Neither etiology, nor genetics, nor congenital deficits are mentioned or considered in the definition.

In 1988 (when Mr. Johnston was 28 years old) he was tested again with the adult Wechsler. On the WAIS-R there was a large split between Johnston’s Verbal IQ of 75 and Performance IQ of 101. Such a wide difference is statistically rare and unexpected, so much so that the full-scale IQ is regarded as uninterpretable (meaningless). Nonetheless, one of the state's experts in 2009/10 calculated a full-scale IQ 83 for WAIS-R.  Mr. Johnston was next given (at age 40) a newer Wechsler (WAIS-III) in 2000. His WAIS-III FSIQ score was 76. Another WAIS-III was administered in 2005, when Johnston was 45 years old.  On this WAIS-III his FSIQ score was 82 (or 84 - both scores are reported in this opinion). Finally, in July 2009 Mr. Johnston was administered the latest edition of the adult Wechsler IQ test (WAIS-IV, 2009), on which he received a FSIQ score of 61. In short, the pattern of Mr. Johnston's IQ scores is highly variable, ranging from a low of 57 to a high of 84 and then back down to 61. The differences among the IQ scores presented a thorny problem: which scores to accept, which to reject? How to rest comfortably with an opinion about Johnston’s level of general intellectual ability?

The defense experts presented a variety of arguments supporting the opinion that Mr. Johnston is a person with mental retardation; most centered on the IQ scores. One theme among these arguments was the interfering effects of the "Flynn Effect," a statistical phenomenon that IQ scores artificially increase over time in tests that have not been renormed for a number of years.  The defense experts believed that the Flynn Effect could account for the variable IQ scores. The other theme argued for the influence of "practice effects," which refers to the fact that the more an individual takes the same test, the more familiar they become with that test, and the more likely IQ scores will increase artificially due to practice (familiarity). While these reasons were offered, the court’s opinion does not report the experts’ explanations as to how or why these factors should influence an MR/ID Dx (or not) in Mr. Johnston's case---only that it is known that the Flynn Effect and practice effects can be variables that must be considered when evaluating a history of IQ scores. Hopefully, a more complete and relevant explanation was offered during testimony.

Much of the argument from the experts (on both sides) centered on the Wechsler IQ test itself, its validity, its psychometrics, and a comparison of scores between the WAIS-III in WAIS-IV. The defense experts argued that the WAIS-IV (2009) is a superior test to the others, uses a four-factor model to derive IQ scores, and is a superior measure of intelligence compared to the WAIS-III and it’s two-factor model of interpretation. The defense experts argued that the WAIS-IV, and consequently the 61 IQ obtained from Mr. Johnston in 2009, represented the "gold standard" of intelligence testing, and provides the best indicator of his true intellectual functioning, thus meeting the legal and clinical standard for mental retardation (actually only the first prong of the definition).

With respect to the variability in IQ scores over Mr. Johnston's lifetime, the defense experts argued that such variability is to be "expected" as there is much variability among very low IQ scores.  The defense experts then discounted (or gave little weight to) the higher 1974 and 1988 IQ scores "because those tested not reflect the most current testing data." One defense expert said he could not find the actual 1974 report, did not know where it was administered, or who administered the test, and therefore didn't trust the validity of the IQ scores.

The defense expert who administered the WAIS-IV in 2009 also administered the TOMM test as a check against malingering in order to demonstrate the validity of the WAIS-IV FSIQ. Adaptive ability was addressed by several defense experts.  One expert interviewed the mother and brother of the defendant, another administered the Adaptive Behavior Assessment System, Second Edition (ABAS-II) and reported that the defendant scored very low, 4 or less, in all 10 scales of the ABAS-II.

On the other hand the state's experts testified that they assessed Mr. Johnston in 2005 (they did not examine him in 2009, but reviewed the reports of the defense experts). One state expert administered the WAIS-III in 2005 (FSIQ = 82 or 84) and both experts concluded that Mr. Johnston was not a person with mental retardation. Both experts gave greater weight to the 1974 WISC IQ FSIQ score of 80, as they noted the examiner's positive remarks describing the defendant as alert, cooperative, friendly, verbally expressive, and exhibiting self-confidence during the testing.

The defense experts had a more awkward argument to maintain, having to weigh Mr. Johnston’s lower IQ scores more heavily, while giving less weight to higher IQ scores for various reasons. They also had to walk the gauntlet that the WAIS-III, an established and comprehensive measure of intelligence, was not a "piece of junk," while trying to give greater weight and emphasis to the 61 IQ from the WAIS-IV. In spite of various defense experts’ arguments, one of the most important and influential pieces of information came from a state's expert who testified that the correlation between the WAIS-III and the WAIS-IV is .94 "or almost perfect, which signified that the WAIS-III was measuring the same constructs as the WAIS-IV and there was a great deal of overlap between the two instruments, making them almost "identical."" One defense expert who argued for the superiority of the WAIS-IV over the WAIS-III was unable to cite the correlation between the two tests as provided in the WAIS-IV test manual, saying that the correlation was probably "mid-.8", which is about .10 lower in magnitude than is actually the case. Knowing the precise correlation (.94) allowed the state’s expert to testify convincingly that the technical concerns raised by the defense about the "two factor model" versus the "four factor model" were relatively insignificant.

The state’s experts also made a salient point that the TOMM test, administered to establish the validity of the WAIS-IV IQ test performance, was given at a much different time, and the court noted,"... the TOMM was not administered properly in that administering the TOMM and the WAIS-IV some two or three months apart, (so that) the ability to make an extrapolation from one test to the other was lost." True enough.

Also of significance, when addressing the drop in IQ from 2005 (FSIQ 82 to 84) to 2009 (FSIQ 61), the state’s experts examined individual responses to IQ test questions and found unexpected inconsistencies, such as when the defendant answered 4+5 =9 in 2005, but when asked to solve a similar problem in 2009 said, “I can’t add.” Also, Johnston could identify Martin Luther King, Jr. in 2005 but never heard of him in 2009.

While the defense experts attempted to assess adaptive ability, no defense expert interviewed anyone at the prison. The court took note of omission and expressed concern that the defense, "... did not interview any personnel at the Department Of Corrections who would have been familiar with Defendant on a day-to-day basis to further assess this issue." The court found the absence of current first-hand information was a significant weakness in the assessment data in spite of claims that adaptive test data (from the ABAS) indicated adaptive deficits. In addition, the court found the information from the defendant's mother and brother provided, "... far too little information and were too distant in time to have any probative value."

Overall, it would seem that this Atkins opinion regarding the presence of mental retardation turned on several factors in the decision. It was important to the court that IQ scores from the WAIS-III and the WAIS-IV are "virtually identical" [blogmaster coment--click here for CHC analysis of each instruments FS IQ composition] so that all technical arguments about the superiority of one score over another, and arguments advancing the psychometric superiority of the WAIS-IV, became irrelevant. This opinion noted that the “Flynn effect” arguments were made by defense experts, but the court document unfortunately provides no additional information about how the significance of the Flynn effect or practice effects were relevant to Mr. Johnston’s mental retardation claim.

This case is instructive on a number of points. For example, it can be problematic for a contemporary evaluator when childhood IQ scores are accompanied by comments from the earlier evaluator that dismiss a low IQ as not being "representative" of the true functioning of the youngster. It's often the case that less skilled examiners don't trust their own test data and tend to superimpose their own "clinical impressions" that a youngster is not mentally retarded for one reason or another. Usually bias, lack of training or experience, misinformation, or not having the benefit of 40+ years of additional research, play a large part in this clinical interpretation of IQ scores. However, comments about a youngster coming from a dysfunctional home, an unhealthy home environment, and/or having emotional problems, must be considered and given appropriate weight in the retrospective evaluation process. Additionally, when previous evaluators note that a youngster was fully cooperative and engaged during testing, that, too, must be weighed accordingly, especially when there is a noticeable increase in the IQ score. This last point was clear in this case.

In Mr. Johnston’s history of IQ scores the 1967 Stanford Binet IQ of 57 can be seen as an "outlier"-- an extremely low IQ score that is inconsistent with all other reported IQ scores. Nor is there data to support significant adaptive deficits throughout Mr. Johnston's life (i.e., very low functioning consistent with someone having an IQ 57, a very low score). Therefore, subsequent evaluators should consider the 1967 examiner’s comment that the 57 IQ score most likely represented a depressed estimate of intellectual ability as (likely) an accurate caveat. Likewise, the 1972 examiner's similar observation that emotional problems depressed the IQ score must be considered accordingly in the retrospective analysis of IQ scores. The judge in Mr. Johnston’s case found the state’s experts’ reasoning compelling. (Importantly, this judge also found the state’s experts’ explanations more detailed and credible with respect to secondary factors that could depress IQ scores in the past and in the present day, namely, anxiety about his impending execution.) The defense experts failed to overcome the “common sense” questions of the low IQ scores from Mr. Johnston’s childhood: an emotionally distraught youngster living in an unhealthy family environment is not be expected to perform at optimal levels when solving intelligence problems.

Evaluating the adaptive abilities of a defendant who has been living on death row for a number of years presents significant challenges to a contemporary evaluator, not the least of which is collecting valid and reliable information from collateral sources who know how the individual functions. As noted in this opinion, family members may be too unfamiliar and removed from current functioning to provide useful information. Additionally, they may be biased in favor of the defendant. Likewise prison personnel may not be able to provide the kind of information needed in such a limited and structured environment; they, too, may present a different bias toward normalcy. Additionally, prison personnel may not be made available to the evaluator. The court’s opinion in Mr. Johnston's case suggests, however, that it is important to make an honest effort to collect information from contemporary collateral sources, weighing and evaluating the validity of the information after it is collected, or at least after an honest attempt is made.

From a distance (and without benefit of copies of the oral testimonies) it would appear that the defense experts became “blinded” by the bright lights of the IQ test arguments. There is considerable intelligence testing research and expertise to draw upon from the extant literature. Another potential “blinding” of the defense experts is the fact that when one can establish the validity of a higher IQ score, well above the IQ range established for subaverage intellectual functioning (IQ approximately 70), the IQ score alone may have sufficient power to “conclusively refute the mental retardation diagnosis both legally and clinically” as conceded by defense experts. For this reason, a valid higher IQ score can be a “deal breaker” for the first prong of the MR definition, and thus the whole MR claim. Further assessment is not required. It is for this reason, however, that IQ scores are sometimes given more weight and emphasis than they deserve in Atkins arguments. Not infrequently, as in Mr. Johnston’s case, it is not easy to examine a retrospective history of IQ test scores and definitively establish or refute mental retardation, especially in a retrospective evaluation that spans decades, using various tests, and has been conducted by multiple examiners. In such cases retrospective evaluators may need to look elsewhere for data or information to form an opinion, namely, adaptive functioning, the “middle child” of the MR criteria.

The Johnston opinion is a good illustration of the difficulty of evaluating the subaverage intellectual functioning prong of the MR definition in the presence of multiple inconsistent IQ scores. These are typical cases that show up at the doorsteps of psychological experts. Atkins defendants with consistent IQ score histories in the 60’s or 80’s are easier to assess one way or the other. However, someone with Mr. Johnston’s IQ history confounds efforts to reach a firm conclusion regarding subaverage intellectual functioning. Technical expertise regarding psychometric issues may or may not help to untie the knot. In Mr. Johnston’s case most of the tests used were from Wechsler batteries (WISC, WAIS-R, -III, -IV), which is somewhat unusual; often a variety of brief, nonverbal only, group administered, or discontinued IQ tests are present in the defendants records. In this case it would seem the (over-) focus on the Wechsler IQ score validity took precedence among the defense experts’ opinions. Yet, when all is said and done, adaptive functioning (the second prong of the MR definition) may provide the clarity and more accurate insight necessary to evaluate a defendant’s overall functioning with respect to a finding of mental retardation.

This case also highlights the difficulties often encountered by the retrospective Atkins evaluator in both IQ and adaptive deficit domains. Perhaps Mr. Johnston’s argument for a finding of mental retardation would have been more compelling if the presence of adaptive deficits had been more thoroughly documented and presented (assuming the truth of such deficits exists). As this Atkins opinion demonstrates, arguments supporting a finding of mental retardation must balance expertise and technical knowledge about intelligence testing against practical and common sense ‘everyday’ considerations. Common sense sometimes leads and sometimes misleads, but it is always a useful foundation and context for an expert’s curiosity, evaluation focus, and final opinions. It certainly was the thread that ran through this carefully reasoned opinion.

As a result of their significant intellectual deficits, people with mental retardation have difficult lives of a particular kind. The difficulties they may experience “getting by” and “fitting in” can make them more vulnerable to criminal influences. This was the concern originally expressed by the U.S. Supreme Court in the Atkins 2002 decision when it ruled against the execution of individuals with mental retardation. The defense experts in this case had an uphill climb. They chose a thorny path. Nonetheless, the larger lesson from this case is not about psychometric technicalities, but about presenting the (in)adequacy of Mr. Johnston’s life, such as it is, or is not. That information was available to both sides. Sometimes the IQ measurement question cannot be answered to the desired level of certainty. The state’s experts in this case were direct and parsimonious. They did not lose sight of the practical issues of the case and the judge was persuaded.

In the end it is always the impaired life and deficient (dependent and limited) daily functioning that is the hallmark of mental retardation. The burden of proof for mental retardation was on Mr. Johnston; the default position was the absence of mental retardation. When the IQ score waters are muddied, as in this Atkins claim, experts must broaden their focus to include other data that may allow one to better see the forest, not just the trees. The state’s experts provided a relatively direct and persuasive context for their opinions. In this case the defense’s psychometric arguments did not carry the day and other data and explanations were not compelling. As noted, Mr. Johnston’s mental retardation claim was not an easy one to establish, and in the end the judge was not persuaded.

[Thank you to Drs. Kevin McGrew and Greg Olley for the generosity of their time making comments and editing suggestions]
Technorati Tags: , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ,