Now I Understand Why Bill Gates Didn’t Want The Value-Added Data Made Public

It all makes sense now.

At first I was a bit surprised that Bill Gates and Michelle Rhee were opposed to publicizing the value-added data from New York City and other cities.

Could they be experiencing twinges of a bad conscience?

No way.

That’s not it. Nor do these educational Deformers think that value-added mysticism is nonsense. They think it’s wonderful and that teachers’ ability to retain their jobs and earn bonuses or warnings should largely depend on it.

The problem, for them, is that they don’t want the public to see for themselves that it’s a complete and utter crock. Nor to see the little man behind the curtain.

I present evidence of the fallacy of depending on “value-added” measurements in yet another graph — this time using what NYCPS says is the actual value-added scores of all of the many thousands of elementary school teachers for whom they have such value-added scores in the school years that ended in 2006 and in 2007.

I was afraid that by using the percentile ranks as I did in my previous post, I might have exaggerated or distorted how bad “value added” really was.

No worries, mate – it’s even more embarrassing for the educational deformers this way.

In any introductory statistics course, you learn that a graph like the one below is a textbook case of “no correlation”. I had Excel draw a line of best fit anyway, and calculate an r-squared correlation coefficient. Its value? 0.057 — once again, just about as close to zero correlation as you are ever going to find in the real world.

In plain English, what that means is that there is essentially no such thing as a teacher who is consistently wonderful (or awful) on this extremely complicated measurement scheme. How teacher X does one year in “value-added” in no way allows anybody to predict how teacher X will do the next year. They could do much worse, they could do much better, they could do about the same.

Even I find this to be an amazing revelation. What about you?

And to think that I’m not making any of this up. (unlike Michelle Rhee, who loves to invent statistics and “facts”.)

PS:

I neglected to give the links to where you can find the raw data. (Warning: some of these spreadsheets are enormous); Here they are:

http://www.ny1.com/content/top_stories/156599/now-available–2007-2010-nyc-teacher-performance-data#doereports
or, if you prefer a shorter URL, try this one:

http://tinyurl.com/836a8cj

The URI to TrackBack this entry is: https://gfbrandenburg.wordpress.com/2012/03/03/now-i-understand-why-bill-gates-didnt-want-the-value-added-data-made-public/trackback/

RSS feed for comments on this post.

52 CommentsLeave a comment

  1. Reblogged this on educatorvoices and commented:
    If this data analysis by a New York teacher is based on real data as he claims then it suggests that rating teachers by the VAM scores is no more valid that raning/ rating teachers by their numerology score. perhaps we could campaign for this.

    Like

  2. I’m stealing your post –giving you credit, but stealing it. I can’t guarantee people will link over here.

    This is rather devastating to the claims of value-added, isn’t it?

    Wow.

    Like

  3. Reblogged this on Millard Fillmore's Bathtub and commented:
    You wanted evidence that Michelle Rhee’s plans in Washington, D.C., were not coming to fruition, that the entire scheme was just one more exercise in “the daily flogging of teachers will continue until morale improves?”

    G. G. Brandenburg ran the numbers. It isn’t pretty.

    Like

  4. This is intriguing. Can you share your raw data sources?

    Like

  5. […] Reblogged from GFBrandenburg's Blog: […]

    Like

  6. Really good research, and I really like your hypothesis about Gates and Rhee. Maybe teachers should shift our tactics and release more of this data to show just how flawed it really is- kind of like the Chevy Volt- sounds great, but no once it is out everyone sees its limited value.

    Like

  7. […] ego invested to turn back now, which to me is insane. But check out Brandenburg’s post here and his analysis of the value added data. Share this:Like this:LikeBe the first to like this post. Tags: brandenburgh, data, edreform, […]

    Like

  8. […] GFBrandenburg’s blog explains why Bill Gates didn’t want the VAM scores made public. They’re total bullshit and blow the cover off of his reform […]

    Like

  9. What Rhee and Gates will never figure out is the ultimate variables….students, their desires, their backgrounds, and their issues.

    Like

    • Here, here! (or is that hear, hear!) (Or should it be See! See!)

      Like

    • the perfect comment. game over.

      Like

  10. What a mess. It looks like somebody threw up.

    Like

  11. I interpret this result to mean that 6% of “Value Added” (whatever that is) would be associated with the combination of “teacher&school”. That means 94% would be attributed to everything else like “students”. Logic does not dictate blaming the teachers.

    Personally, I think it’s poverty, and the data I’ve plotted on NC school performance vs. free lunch fraction bears this out.
    I’ve also plotted SAT scores versus family income across North Carolina counties and posted it at:

    noyes ES proficiency graph


    Across the range of county average income, the SAT scores increase more than 150 points from low to high income. I also show data on school performance vs free lunch fraction, and high-free-lunch schools show just 60% of the performance of those with low-free-lunch fractions.
    Any discussion of “standardized testing” has to include a discussion of “socioeconomic status”.

    Liked by 1 person

  12. I don’t think Rhee and Gates were against LAT publication of teacher ratings; do you have sources for that? Now they are b/c of all the bad blow-back but I don’t recall them speaking out either way at the time.

    Like

    • You may be right. I have definitely read comments by them on the NYC data, and not on the LA data.
      I’ll fix that.

      Like

  13. As I said on Strauss’ blog on 2/26:

    “I do believe this type of exposure is why people like Michelle Rhee were against publicizing the NY scores. She knows just how shaky they are, thus just how bogus the IMPACT system is that she and current Chancellor Henderson instituted here in DC.

    They don’t want to have to defend actual data, because they know they can’t. Supposing some teachers were fired here for poor performance based on questionable VAM scores. Imagine all the bad publicity, just as reformers are trying to have these measures used across the country.”
    http://www.washingtonpost.com/blogs/answer-sheet/post/a-simple-question-teachers-should-now-ask-about-their-profession/2012/02/26/gIQAwMplcR_blog.html#comments

    Like

  14. These values do not tell the true story. There will be a different group of students from year to year. Culture, economic conditions, previous education or training will invalidate the data.

    Like

    • The kids in a given school generally do not change very much from one year to the next.

      Like

      • It is true that the overall population of a school won’t likely change a great deal from year to year. However, the population and dynamics of an individual class within that school can, and does, change dramatically. The mix of personalities, abilities and exceptionalities can vary widely.

        Like

      • Kids do change dramatically from year to the next. This is what is disturbing when we compare value-added data from year to year. Moreover, by so doing, we may be playing into Gates’ program. Pointing out the absurdity of the data, so obvious to anyone who has taught, does acknowledge, at least tacitly, that the data exists. What we should be saying (you are) is such data is meaningless. Good teachers do get best results, but the circumstances accompanying the kids makes any evaluation based on test scores dead on arrival.
        I don’t know if Gates and company know this (anyone who thinks increasing class size helps educate probably does not) but the fact remains that he has way too much influence on American schools. You should also consider that the deck could be stacked against teachers who fail the sycophant biopsy. Their rolls could be stacked in a manner that insures teacher failure. I know of principals that do this. Proving it is another matter.

        Like

  15. […] found a blog from a retired math teacher and followed it. Share this:TwitterFacebookLike this:LikeBe the first […]

    Like

  16. 2007 scores were inflated. Thus (through no fault of your own) I think this graph is essentially meaningless.

    Year to year value-added correlation is about .35 (range .30-.50). Year to career value added correlation is .6.

    Not a meaningless tool. A misused one.

    Like

    • But these inflated scores are being used to falsely rate teachers, right?

      Thus, the graph is incredibly meaningful to teachers.

      In DC, we have some teachers being rated on scores inflated by administrator cheating and cover-up of the cheating. This type of deception can have a huge effect on teachers, especially when the scores plummet the next year when the cheating stops and teachers get blamed for for the declining achievement.

      Like

      • Meaningless for the purposes of abstract statistical analysis. Deeply harmful to publish for the reasons you state and many others.

        Like

  17. Not sure if I’m convinced that VA is not a useful indicator (albeit as with every number trying to illustrate something complex it is flawed). You’re trying to find a trend over two years, but may be VA is like the weather: erratic over the short term but consistent over the long-term.

    Like

    • If we had some sort of medical test that was as unreliable as this VA stuff, we would say it was utterly useless. What good is a test for condition X if it has an error rate of 57% or more? None, and any doctor who put faith in it ought to have his/her head examined.

      Like

      • That’s not an accurate analogy – we’re talking about performance measures.

        In medicine (btw, I’m from London, UK, so that’s my ref point), there are performance measures such as ‘waiting times’, or ‘surgery deaths per thousand’. These are also deeply flawed, but tell a story of some kind.

        I think that VA has purpose, but probably not, as this debate is right to highlight, as an accountability measure; probably as a self-reflective tool and a school development measure (among many others). Eg a teacher might want to know the difference s/he has made compared to the previous year and reflect accordingly.

        And if VA is going to be used for public reporting, first it shouldn’t be published teacher-by-teacher – it’ll just kills motivation and is flawed – but it could be collated school-by-school or district-by-district. VA often shows that schools in poorer areas are doing a better job than well funded private schools who get their kudos from the educationally rich home environment.

        It is also worth diving into VA more. In the UK, until recently, we had Contextual Value-Added, which factored into the VA figure a metric for poverty and ESL. As you can imagine, the tough urban school often were found to perform the best in terms of the impact teachers were making on children.

        There is another massive problem with all this data though and that’s how it effects teaching and learning. Publishing data based on test scores tells a very partial picture, but will influence the direction of teaching. In UK we’ve seen it in practices such as teaching to the test and targeting particular groups of children to ensure certain benchmarks are met.

        Like

      • I would love to hear more about what is happening in the UK with education. Sounds like some of the same misguided ideas on how to deform education are crossing the north atlantic in both directions. Do you know of any thoughtful, insightful blogs or websites on what is happening with schooling or social policy in the UK or elsewhere in Europe? I just started reading Pasi Stahlberg’s “Finnish Lessons” yesterday.

        Like

  18. Glad to see this, the problem is that children are NOT a product, they cannot be treated with one treatment and are not meant to turn out the same way.

    And if the poster above did think about this more, this is CHILDREN, not medical issues, how would you want your children to be educated?? Step into a school and see what it’s like, where is the creativity, the projects, the children’s voice etc. Why are these business people forcing our children through the schools to think the same way, respond the same way???

    And when does society think about the poverty issue?? It is not just the fault of the school system, until we all take responsibility and face it, we will always have children who are facing this huge issue.

    Like

  19. Sean — the Value-Added results are being used in DC to rate teachers, in some cases to ineffective or minimally effective, based on faulty data.

    I hasn’t been published, but it has ruined some careers.

    Like

  20. […] these performance evaluations systems themselves do not perform as expected:  Here it is, “Now I understand why Bill Gates didn’t want the value-added data made public“: It all makes sense […]

    Like

  21. What would a chart look like, were a measure like this actually to work in distinguishing good from bad teachers?

    Like

    • It would look like a cigar- shaped blob that forms a 46-degree angle with the two axes, going thru the origin, from upper right to lower left.

      Like

  22. […] year essentially explains nothing about what’s going to happen next year. Or look here, or here, or here. Or here  or here or here (by Gary Rubenstein) Would you trust a medical test of some […]

    Like

  23. […] on how this “value-added” model of teacher evaluations from these test scores is a “complete and utter crock” when analyzing the […]

    Like

  24. How are you matching the teachers for that graph?

    The xls files that I just downloaded don’t have unique teacher ids. In fact there is no teacher id in the Teacher 1 and Teacher 2 columns as there seems to be in the simple report forms (the pdf format reports).

    thanks

    Like

    • From what I understand, each line represents a unique teacher.
      If you have a PDF version of much the same data, I’d love to see it!!

      Like

  25. Couldn’t reply to the chain above, but in response to your request for UK based blogs on social policy and education:

    This is a really good journalist’s blog: http://www.educationbynumbers.org.uk/blog/

    This is a great researcher – best follow him on twitter:
    http://web.me.com/dylanwiliam/Dylan_Wiliams_website/Welcome.html

    This is good think tank that looks at a broad range of issues (I’ve filtered their research page to only show ed publications):
    http://www.ippr.org/research-projects/12?megafilter=education

    And this is these are the two main UK ed newspaper:
    http://www.guardian.co.uk/education
    http://www.tes.co.uk/

    Also, if you’re on Twitter you can check me our here (https://twitter.com/#!/ShirazC) and see who from the UK I follow.

    Hope that’s helpful!

    Like

  26. […] Cuomo, Arne Duncan, and Bill Gates all agree that it may not be best to make all teacher evaluations […]

    Like

  27. […] would Cuomo, Arne Duncan, and Bill Gates all agree that it may not be best to make all teacher evaluations […]

    Like

  28. I was wondering if you ever considered changing
    the structure of your website? Its very well written; I love
    what youve got to say. But maybe you could a little more in the way of content so people could
    connect with it better. Youve got an awful lot of text for only having one or two
    pictures. Maybe you could space it out better?

    Like

  29. […] "It all makes sense now. At first I was a bit surprised that Bill Gates and Michelle Rhee were opposed to publicizing the value-added data from New York City and other cities. The problem, for them, is that they don’t want the public to see for themselves that it’s a complete and utter crock. Nor to see the little man behind the curtain." | by Scott Mcleod  […]

    Like

  30. […] Looking at the same exact teachers, in the same exact schools, teaching the same subjects there was no correlation between teachers’ 2005-06 scores and their 2007-08 scores. There was no correlation between a […]

    Like

  31. […] Looking at the same exact teachers, in the same exact schools, teaching the same subjects there was no correlation between teachers’ 2005-06 scores and their 2007-08 scores. There was no correlation between a […]

    Like

  32. Reblogged this on ΕΝΙΑΙΟ ΜΕΤΩΠΟ ΠΑΙΔΕΙΑΣ.

    Like

  33. I just retired after 41 years as a chemistry/physics/math teacher. The first thing I thought of when I saw this graph is the probability diagram in quantum mechanics for finding an s orbital. It’s a sphere, and as we all know, a sphere is symmetrical, so it goes no place. And that’s where “value added” data should end up going.

    Like

  34. Statistics? Are they using statistics? Are they using research based theories to make sweeping changes? If so, I haven’t seen the studies.

    Like

  35. […] had no consistency from one subject to the next. See my blog  (not all on NY City) here, here, here,  here, here, here, here, here,  here, here, and here. See Gary R’s six part series on […]

    Like

  36. […] Rubinstein and I have dissected the value-added scores for New York City public school teachers that were computed and released by […]

    Like


Leave a comment