Saturday, July 28, 2018

Five Questions and Answers on the 2018 GA Milestones

It’s Georgia Milestones scores time again!  I always look forward to the state releasing scores on the annual exams.  It gives a good reason to check in on whether different initiatives are having their intended effect.  As I explained a few years ago, the score releases have so many numbers and are hard enough to interpret that they tend to provide lots of fodder for misguided newspaper headlines, tweets, and emoji high-fives.  In general, the conversation overstates how much fundamentals in a district or school change from one year to the next.    

Below I’ve tried to take a relatively rigorous look at five questions that interest me about Atlanta Public Schools.

It’s worth caveating that all the comments below are based on test scores.  If you believe test scores are effective measures of what we want schools to teach students, you may find these conclusions useful.  If not, nothing below will likely be of interest to you because this is about the Georgia Milestones.


      1. Has achievement in Atlanta Public Schools (APS) changed?

Not really. 

For each of the past seven years, APS has performed similarly relative to the state.  Back in 2011, the average student in the district scored 0.62 grade levels behind the state.  In 2018, the average student scored 0.55 grade levels behind the state.  The highest score occurred in 2015 and the lowest in 2012.

In general, district scores jump around a bit from one year to the next.  The graph below shows how changes in APS scores compare to changes for all the other districts in the state.  From this graph it becomes clear that what's remarkable about APS is how little scores have moved.

One reason APS scores haven’t moved much is that it’s a large district.  Small districts tend to see more variation in their scores from one year to the next.  However the stability in APS scores is noticeable even relative to similarly sized districts.  Among large metro districts, Fulton, Gwinnett, Cobb, Clayton and Forsyth all saw score changes that were more substantial than those in APS (both up and down).

A second reason that APS scores haven’t changed much is that the needs of students served by the district remain relatively stable.  The district is slowly gentrifying.  Over the past 4 years, APS has produced a consistent measure of needs for test takers in the district.[i]  That measure is called the challenge index and you can read more about it here.  It fell slightly from 60 to 58 over the past four years.

Overall, performance in APS is about the same.  It has perhaps ticked up very slightly, and its students have become slightly less needy, but overall the changes are incredibly modest.



      2. Is achievement in APS more equitable than it used to be?

No.

Outcomes in APS are not more equitable today than they were a few years ago.  Across the country, huge gaps exist between poor students and their wealthier peers.  For the nation as a whole, students at the wealthiest schools score about 6 grade levels above students at the poorest schools.

Consider the graph below.  A perfectly equitable district would expect to see student outcomes be independent of how wealthy or poor a school is.  In APS, the slope is similar to the country as a whole.  The slope has not changed over the past few years.

Equity in the district is about the same as it has been.



      3Are the turnaround partnerships increasing achievement?

Possibly.

APS launched a turnaround plan a few years ago.  One component of the plan transfers management of 6 district schools to charter operators (Purpose Built, Kindezi, and KIPP).  The partnerships are being phased in over several years.  Results are available now to allow us to look at the 2016 and 2017 partnerships to see how they’ve done.

One way to analyze the impact of these partnerships is to look at how the schools were doing before the partnership and compare that to how they are doing after the partnership.  The graph below does that for Thomasville Heights, the only partnership rolled out in the fall of 2016.    

The results show that scores at Thomasville were 0.4 and 0.3 grade levels higher in 2017 and 2018, respectively, than they were in 2016, the last year before the partnership.  This is encouraging, but scores at schools change for all kinds of reasons, so its helpful to think about how these changes compare to other schools in the district (shown in grey).  We can’t say for sure that changes from one year to the next aren’t just noise unless we see that the changes are outside of the normal range that schools experience.

The change from 2016 to 2017 is greater than the change that occurred for 96% of other APS schools.  The change from 2016 to 2018 is greater than the change that occurred for 70% of other APS schools.  So, I would say that the best evidence is that the turnaround at Thomasville has had a positive impact, but the improvements are not so dramatic that they are unheard of.  For example, Burgess (a traditional public school) saw scores rise faster. 


In the fall of 2017, three additional partnerships began at Gideons, Slater, and Price.  A similar analysis is shown for these schools below.  Two saw scores improve, while one saw scores decline.  The greatest increase was at Gideons.  Scores increased by 0.6 grade levels from 2017.  This improvement was greater than 96% of APS schools.  Hope-Hill was the only traditional public school that saw a similarly large increase in scores.


One might wonder if the improvement in scores at partnership schools is driven by changing student populations.  Maybe wealthier families who would not have considered the school before are willing to give it a shot under new management.  I did the same analysis above, controlling for changes in the challenge index, and the results were consistent.  Click these links to see those analyses for 2016 and 2017 partners.


      4. Are targeted interventions increasing achievement?

Not much.

A second component of the APS turnaround strategy was to provide additional resources to several schools that had performed poorly in the past.  The plan included small-group tutoring and mentoring of principals. 

Overall, there is not much evidence that these investments are paying off.  In 2017 about half the schools saw scores rise while the other half saw scores fall.  In 2018, results look a little better, but that picture doesn’t hold up once you control for changes in the challenge index.


It seems like year-to-year changes at these schools may be driven by noise and school-specific things (for example leadership or staffing changes) rather than the targeted intervention strategy.  The biggest increase among the targeted schools occurred at BAMO.  The biggest decline occurred at Scott.


      5. Are charter operators beating the odds?

KIPP and Drew are.  Others are not.

About twenty percent of APS K-8 students now attend a locally-approved charter school.  In 2018, students at KIPP scored a half a grade level above what would be expected of them based on the needs of the students who attend.  Drew students also performed better than would be expected by 0.4 grade levels.

All other charter operators performed worse than would be expected based on the students who attend.  The worst performance occurred at Atlanta Classical.  Students at that school scored 1.2 grade levels behind what would be expected.  In math skills, which are often thought to change more with school inputs, the performance was 1.8 grade levels behind.  Even though Atlanta Classical has one of the wealthiest student populations in the state, about half the schools in the state score better at math.  Students at the traditional public school, Jackson Elementary, have a similar challenge index.  However, in math, students at Jackson score 2.5 grade levels ahead of Atlanta Classical.  

The graph below shows a history of performance by charter operator.  For some, performance has varied from year to year.  For others, it has been consistently above or below expectations. 




For more analysis and to investigate specific schools, visit APS Insights.    



[i] Before this, a different measure (Free and Reduced Lunch) was generally used.  Over time, this measure became less and less accurate due to changes in federal policy, so districts and researchers moved toward measuring student needs in more accurate ways.

8 comments:

  1. Thank you for sharing this information,very helpful and insightful. Much appreciated.

    ReplyDelete
  2. Is it logistically impossible to measure by student or by class? I have never understood why current 3rd graders are compared to last year's 3rd graders instead of being compared to last year's 2nd graders. Wouldn't it make more sense to compare scores for the same group of students over time? I realize retained students would skew the results. With all the individual tracking computers can now do, why isn't this being done for students to measure school effectiveness?

    ReplyDelete
    Replies
    1. Good points - yeah ideally student level data would be used and that opens up some better methods. This is just using school-grade-subject data. The benefit is it’s all publicly available, but there is a downside in terms of rigor/methods available.

      Delete
  3. This is great analysis! Thanks for sharing! Just a clarifying question: When you use the phrase "Beating the Odds" in #5, you aren't referring to the state's metric (also called "Beating the Odds") that compares a predicted CCRPI score for a school to the actual CCRPI score the school earns, correct? I'd love to know more about how you are using the school-grade-subject data to quantify grade levels (I'm guessing mean scale score maybe?)

    ReplyDelete
    Replies
    1. Correct. I'm using the mean scale score (normalized into grade levels so that it is easier to interpret) and then comparing that to the expectation from the challenge index (which is calculated for the students in each test-grade-subject combination).

      Delete
  4. Very helpful and insightful. Wondering how you come to the definitive conclusion that KIPP is beating the odds. Just based on the graph above? Given the year to year variation within school, 3 of 4 years above 0 looks like a fairly probable outcome even if they were no better than average.

    ReplyDelete
    Replies
    1. When I said which charters beat the odds, I meant for 2018. You are right that some of the history has not always matched 2018.

      Delete