There’s a wonderful statistical discussion of Michael Winerip’s NYT article critiquing the use of value-added modeling in evaluating teachers, which I referenced in a previous post. I wanted to highlight some of the key statistical errors in that discussion, since I think these are important and understandable concepts for the general public to consider.

**Margin of error**: Ms. Isaacson’s 7th percentile score actually ranged from 0 to 52, yet the state is disregarding that uncertainty in making its employment recommendations. This is why I dislike the article’s headline, or more generally the saying, “Numbers don’t lie.” No, they don’t lie, but they do approximate, and can thus mislead, if those approximations aren’t adequately conveyed and recognized.**Reversion to the mean**: (You may be more familiar with this concept as “regression to the mean,” but since it applies more broadly than linear regression, “reversion” is a more suitable term.) A single measurement can be influenced by many randomly varying factors, so one extreme value could reflect an unusual cluster of chance events. Measuring it again is likely to yield a value closer to the mean, simply because those chance events are unlikely to coincide again to produce another extreme value. Ms. Isaacson’s students could have been lucky in their high scores the previous year, causing their scores in the subsequent year to look low compared to predictions.- Using
**only 4 discrete categories**(or ranks) for grades:- The first problem with this is the
**imprecision**that results. The model exaggerates the impact of between-grade transitions (*e.g.*, improving from a 3 to a 4) but ignores within-grade changes (*e.g.*, improving from a low 3 to a high 3). - The second problem is that this exacerbates the
**nonlinearity**of the assessment (discussed next). When changes that produce grade transitions are more likely than changes that don’t produce grade transitions, having so few possible grade transitions further inflates their impact.

Another instantiation of this problem is that the imprecision also exaggerates the**ceiling effects**mentioned below, in that benefits to students already earning the maximum score become invisible (as noted in a comment by journalist Steve Sailer:Maybe this high IQ 7th grade teacher is doing a lot of good for students who were already 4s, the maximum score. A lot of her students later qualify for admission to Stuyvesant, the most exclusive public high school in New York.

But, if she is, the formula can’t measure it because 4 is the highest score you can get.

- The first problem with this is the
**Nonlinearity**: Not all grade transitions are equally likely, but the model treats them as such. Here are two major reasons why some transitions are more likely than others.- Measurement
**ceiling effects**: Improving at the top range is more difficult and unlikely than improving in the middle range, as discussed in this comment:

Going from 3.6 to 3.7 is much more difficult than going from 2.0 to 2.1, simply due to the upper-bound scoring of 4.

However, the commenter then gives an example of a natural ceiling rather than a measurement ceiling. Natural ceilings (

*e.g.*, decreasing changes in weight loss, long jump, reaction time,*etc.*as the values become more extreme) do translate into nonlinearity, but due to physiological limitations rather than measurement ceilings. That said, the above quote still holds true because of the measurement ceiling, which masks the upper-bound variability among students who could have scored higher but inflates the relative lower-bound variability due to missing a question (whether from carelessness, a bad day, or bad luck in the question selection for the test). These students have more opportunities to be hurt by bad luck than helped by good luck because the test imposes a ceiling (doesn’t ask all the harder questions which they perhaps could have answered). **Unequal responses to feedback**: The students and teachers all know that some grade transitions are more important than others. Just as students invest extra effort to turn an F into a D, so do teachers invest extra resources in moving students from below-basic to basic scores.

More generally, a fundamental tenet of assessment is to inform the students in advance of the grading expectations. That means that there will*always*be nonlinearity, since now the students (and teachers) are “boundary-conscious” and behaving in ways to deliberately try to cross (or not cross) certain boundaries.

- Measurement
**Definition of “value”**: The value-added model described compares students’ current scores against predictions based on their prior-year scores. That implies that earning a 3 in 4th grade has no more value than earning a 3 in 3rd grade. As noted in this comment:

There appears to be a failure to acknowledge that students must make academic progress just to maintain a high score from one year to the next, assuming all of the tests are grade level appropriate.

Perhaps students can earn the same (high or moderate) score year after year on badly designed tests simply through good test-taking strategies, but presumably the tests being used in these models are believed to measure actual learning. A teacher who helps “proficient” students earn “proficient” scores the next year is still teaching them something worthwhile, even if there’s room for more improvement.

These criticisms can be addressed by several recommendations:

*Margin of error.*Don’t base high-stakes decisions on highly uncertain metrics.*Reversion to the mean.*Use multiple measures. These could be estimates across multiple years (as in multiyear smoothing, as another commenter suggested), or values from multiple different assessments.*Few grading categories.*At the very least, use more scoring categories. Better yet, use the raw scores.*Ceiling effect.*Use tests with a higher ceiling. This could be an interesting application for using a form of dynamic assessment for measuring learning potential, although that might be tricky from a psychometric or educational measurement perspective.*Nonlinearity of feedback.*Draw from a broader pool of assessments that measure learning in a variety of ways, to discourage “gaming the system” on just one test (being overly sensitive to one set of arbitrary scoring boundaries).*Definition of “value.”*Change the baseline expectation (either in the model itself or in the interpretation of its results) to reflect the reality that earning the same score on a harder test actually does demonstrate learning.

Those are just the statistical issues. Don’t forget all the other problems we’ve mentioned, especially: the flaws in applying aggregate inferences to the individual; the imperfect link between student performance and teacher effectiveness; the lack of usable information provided to teachers; and the importance of attracting, training, and retaining good teachers.

Great article. Some thoughts below:

– The problem with the ceiling effect is huge problem that could be relatively easily fixed by extending the top end of the assessments.

– Are they using non-linear models? If not, why not?

– The bureaucratic top-down approach is problematic, especially given the lack of transparency in the assessments. Schools should allow the local communities to assign different weights to the different aspects of education.

– It should not be too difficult to clearly explain VAM to anyone who understands basic algebra. The educational policymakers need to put in the time and effort to involve all the education stakeholders, not just quant wonks.