Skip to main content

NY City Public Schools, and what they might tell us about the SAT

Recently, I received a message from Akil Bello who pointed out a data visualization he had seen.  It was originally posted to Reddit, but later was edited to eliminate the red-green barrier that people with color-blindness face.  The story was here, using a more suitable blue-red scheme.

There's nothing really wrong with visualizing test scores, of course.  I do it all the time.  But many of the comments on Reddit suggest that somehow the tests have real meaning, as a single variable devoid of any context.  I don't think that's a good way to analyze data.

So I went to the NY City Department of Education to see what I can find.  There is a lot of good stuff there, so I pulled some of it down and began taking a look at it.  Here's what I found.

On the first chart, I wanted to see if the SAT could be described as an outcome of other variables, so I put the average SAT score on the y-axis, and began with a simple measure: Eighth grade math and English scores on the x-axis. Hover over the regression line, and you'll see an r-squared of about .90.

Scientists would use the term "winner, winner, chicken dinner" when getting results like this.  It means, for all intents and purposes, that if you know a high school's mean 8th grade achievement scores, you can predict their SAT scores four years later with amazing accuracy.  And--here's the interesting thing--the equation holds for virtually every single school.  There are few outliers.

Ponder that.

But critics of the SAT also say that the scores are reflective of other things, too; an accumulation of social capital, for instance.  So use the control at the bottom to change the value on the x-axis.  Try economic need index, or percentage of students in temporary housing, or percentage of the student body that are White or Asian. The line may go up (positive correlation) or down (negative) but you'll always see the schools with the highest scores tend to have the characteristics you'd expect.

Jump to the second tab.  This is more a response to the Reddit post: The top map shows the ZIP codes and a bubble, indicating the number of schools in that ZIP.  The bottom map shows every school arrayed on two poverty scales: Economic Index and Percent in Temporary Housing.  The color shows the mean SAT score in the school (Critical Reading plus Math, on a 1600-point scale.)  Purple dots represent higher scores.

Use the ZIP highlighter, and you'll see the top map show only that bubble, and the bottom will show the schools in it.

Got the lesson?  Good.  Now, think about why the colleges with high median test scores a) have them, and b) tend to produce students with high GRE and MCAT and LSAT scores,  and c) point to excellent outcomes for their students.

And let me know what you think.






Comments

Popular posts from this blog

Educational Attainment and the Presidential Elections

I've been fascinated for a while by the connection between political leanings and education: The correlation is so strong that I once suggested that perhaps Republicans were so anti-education because, in general, places with a higher percentage of bachelor's degree recipients were more likely to vote for Democrats. The 2024 presidential election puzzled a lot of us in higher education, and perhaps these charts will show you why: We work and probably hang around mostly people with college degrees (or higher).  Our perception is limited. With the 2024 election data just out , I thought I'd take a look at the last three elections and see if the pattern I noticed in 2016 and 2020 held.  Spoiler: It did, mostly. Before you dive into this, a couple of tips: Alaska's data is always reported in a funky way, so just ignore it here.  It's a small state (in population, that is) and it's very red.  It doesn't change the overall trends even if I could figure out how to c...

First-year student (freshman) migration, 2022

A new approach to freshman migration, which is always a popular post on Higher Ed Data Stories. If you're a regular reader, you can go right to the visualization and start interacting with it.  And I can't stress enough: You need to use the controls and click away to get the most from these visualizations. If you're new, this post focuses on one of the most interesting data elements in IPEDS: The geographic origins of first-year (freshman) students over time.  My data set includes institutions in the 50 states and DC.  It includes four-year public and four-year, private not-for-profits that participate in Title IV programs; and it includes traditional institutions using the Carnegie classification (Doctoral, Masters, Baccalaureate, and Special Focus Schools in business, engineering, and art/design. Data from other institutions is noisy and often unreliable, or (in the case of colleges in Puerto Rico, American Samoa, and other territories, often shows close to 100% of enro...

Education Levels in the US, by State and Attainment

Attainment has always been an interesting topic for me, every since I first got stunned into disbelief when I looked at the data over time.  Even looking at shorter periods can lead to some revelations that many don't make sense at first. Here is the latest data from NCES, published in the Digest of Education Statistics . Please note that this is for informational purposes only, and I've not even attempted to visualize the standard errors in this data, which vary from state-to-state.  There are four views year, all looking at educational attainment by state in 2012 and 2022.   The first shows data on a map: Choose the year, and choose the level of attainment.  Note that the top three categories can be confusing: BA means a Bachelor's degree only; Grad degree means at least a Master's (or higher, of course); and BA or more presumably combines those two.  Again, standard errors might mean the numbers don't always add up perfectly. The second shows the data o...