Do Charter Schools Cherrypick Students?

Yesterday I looked into quantitatively measuring the rumor I've been hearing for years, namely that charter schools cherrypick students - get rid of troublesome ones, keep well-behaved ones, and so on.

Here are two pieces of anecdotal evidence. There was a "Got To Go" list of students at one charter school in the Success Academy network. These were troublesome kids that the school was pushing out.

Also, I recently learned that Success Academy doesn't accept new kids after the fourth grade. Their reasoning is that older kids wouldn't be able to catch up with the rest of the kids, but on the other hand it also means that kids kicked out of one school will never land there. This is another form of selection.

Now that I've said my two examples I realize they both come from Success Academy. There really aren't that many of them, as you can see on this map, but they are a politically potent force in the charter school movement.

Also, to be clear, I am not against charter schools as a concept. I love the idea of experimentation, and to the extent that charter schools perform experiments that can inform how public schools run, that's interesting and worthwhile.

Anyhoo, let's get to the analysis. I got my data from this DOE website, down at the bottom where I clicked "citywide results" and grabbed the following excel file:

With that data, I built an iPython Notebook which is on github here so you can take a look, reproduce my results with the above data (I removed the first line after turning it in to a csv file), or do more.

From talking to friends of mine who run NYC schools, I learned of two proxies for difficult students. One is 'Percent Students with Disabilities' and the other is 'Percent English Language Learners' (I also learned that charter schools' DBN code starts with 84). Equipped with that information, I was able to build the following histograms:

Percent Students with Disabilities, non-Charter:

 

Percent Students with Disabilities, Charter:

Percent English Language Learners, non-Charter:

 

Percent English Language Learners, Charter. Please note that the x-axis differs from above:

 

I also computed statistics which you can look at on the iPython notebook. Finally, I put it all together with a single scatterplot:

 

The blue dots to the left and all the way down on the x-axis are mostly test schools and "screened" schools, which are actually constructed to cherrypick their students.

The main conclusion of this analysis is to say that, generally speaking, charter schools don't have as many kids with disabilities or poor language skills, and so when we compare their performance to non-charter schools, we need to somehow take this into account.

A final caveat: we can see just by looking at the above scatter plot that there are plenty of charter schools that are well inside the middle of the blue cloud. So this is not a indictment on any specific charter school, but rather a statistical statement about wanting to compare apples to apples.

Update: I've now added t-tests to test the hypothesis that this data comes from the same distribution. The answer is no.

Those very small numbers are the p-values which are much smaller than 0.05. Other t-tests give similar results (but go ahead and try them yourself):

screen-shot-2015-11-19-at-11-06-13-am.png

Debiasing techniques in science

My buddy Ernie Davis just sent me an article, published in Nature, called How scientists fool themselves – and how they can stopIt's really pretty great - a list of ways scientists fool themselves, essentially through cognitive biases, and another list of ways they can try to get around those biases.

There's even an accompanying graphic which summarizes the piece:

 

I've actually never heard of "blind data analysis" before, but I think it's an interesting idea. However, it's not clear how it would exactly work in a typical data science situation, where you perform exploratory data analysis to see what the data looks like, and you form a "story" based on that.

One thing they mentioned in the article but not in the graphic is the importance of having your research open sourced, which I think is the way to let the "devil's advocacy" and "team of rivals" approaches actually happen in practice.

It's all the rage nowadays to have meta analyses. I'd love for someone to somehow measure the ability of the above debiasing techniques to see which work well, and under what circumstances.