I don't think we need to blow up how we measure schools...yet

I don't think we need to blow up how we measure schools...yet

Back in the spring I asked whether assessment is holding back the Science of Reading. I picked on DIBELS, writing:

In Kindergarten, DIBELS tests (1) letter naming, (2) phonemic segmentation (if I say 'am' you would say /a/ /m/), (3) nonsense word fluency (/h/ /a/ /p/ is pronounced 'hap'), (4) and word reading fluency (which includes lots of sight words).
I'm not a psychometrician. My understanding, though, is that letter naming may be less important initially than pronouncing phonemes from letters, segmentation may be less important than blending, nonsense word fluency may be less important than blending real words, and sight words are not a key indicator of early word reading fluency.

I pointed out that a large assessment company might have an incentive not to follow the latest research if that meant rapidly changing the nature of the assessments that company produces, since if the company did so, the company's customers might see surprising decreases in student performance and choose a competing product.

I got some pushback on that from folks I trust. Again, I'll emphasize, I'm not a psychometrician. But I'll certainly be curious if real psychometrician-economist double majors choose to pull on this thread.

Regardless, if, in that article, I was pushing on a small brick at the bottom of the education assessment wall, in his latest post, "We Measure Our School Systems the Wrong Way" Tim Daly advocates knocking down the whole wall.

Tim writes:

It's time for a major correction. We should define the success of our education system based on outcomes for 25 year-olds, not 18 year-olds...By capping the responsibility of K-12 systems at the end of high school, we give them a free pass for the dismal life outcomes many students - particularly those from less privileged backgrounds - are experiencing in their 20s.

Instead, Tim argues, we should measure (1) educational attainment, employment with a living wage, civic engagement (ie. voter registration), safety and lawfulness (ie. criminal record).

I disagree. Well not entirely. I do think longitudinal outcome measurement is an important frame for public policy, but I don't think this is the most important problem we face. Not by a long shot.

Implicit in Tim's argument is the assumption that if we had all the data on the outcomes we care about and the political consensus to manage to those outcomes we could work backwards and design our schools to meet those outcomes. I'm skeptical, particularly since we have struggled massively to design our schools to meet the far more concrete outcomes that Tim criticizes--test scores and high school graduation.

Let's say we agreed that the longitudinal outcome we cared about was voter registration (to pick one that Tim suggested). We wouldn't optimize for it by making kindergarteners recite the Pledge to Register as a Voter every morning during circle time. Instead, I think most folks would agree that we'd teach them how to read, since that is a precursor skill to voting that is also age-appropriate.

Tim wants to move the goal line with his suggested outcomes. But we have not proven that we know how to get to the ten yard line, since national 4th grade reading scores are exactly where they were 30 years ago. And we called it a crisis back then.


Courtesy of

And if we can't predictably, efficiently and scalably get to the ten yard line, we'll never make it to the goal line, no matter where we choose to draw that ultimate goal.

The good news is that everyone--parents, voters, policy experts, governors believe that we need to get all students reading by third grade. There is no argument about where to draw that line.

We also know, based on decades of research, that beyond first and second grade, learning how to read becomes much harder and more expensive than it is in kindergarten and first grade.

So we should direct every available resource toward teaching children to read in kindergarten and first grade and be open to redesigning staffing, space, and schedules (hello, paraprofessionals!) to making that happen. And then, at long last, our time will become well spent debating whether to measure school districts by life and academic outcomes at 18 or at 25.

Abbey Vannoy

Communications | Impact | Raising Curious, Kind Kids

3 个月

Another non-psychometrician weighing in on this incredibly complex topic. Even worse, without any rigorous evidence.?? I have two 2nd graders in public school and spent nearly a decade as a grants management consultant working with LEAs, school leaders, and education nonprofits. In both personal and professional settings, I find it striking how much human noise remains embedded in assessments -- how they're selected/adopted, administered, interpreted, and communicated. On one end, we have the cheerleaders of specific assessments who interpret and communicate assessment results like gospel (giving Lucy Caulkins fandom vibes a la Sold a Story) without fully appreciating any assessment's limitations. At the other, complete indifference - a box to be checked or a mean's to an end, i.e., getting much-needed resources into persistently under-resourced classrooms. Even when one assessment dominates, e.g. DIBELS, we seem so far from connecting it from curriculum to real skills. And we stray from assessments' primary purpose to allocate finite educational resources for maximum impact. All with a whole lotta noise. Changing the definition of success wouldn't overcome the noise; the added data burdens may worsen it.

回复

要查看或添加评论,请登录

Matt Pasternack的更多文章

  • How can districts respond to the cyber charter earthquake?

    How can districts respond to the cyber charter earthquake?

    If you read K-12 headlines, and I asked you what's the biggest earthquake hitting schools this year, you might say:…

    4 条评论
  • Tutoring programs are not widgets

    Tutoring programs are not widgets

    Matthew A. Kraft Beth Schueler Grace Falken just released an important meta-analysis of 282 randomized controlled…

    5 条评论
  • Tech and Edtech

    Tech and Edtech

    When I was getting started in education reform, there was Linda Darling-Hammond saying "STOP". When I was getting…

    5 条评论
  • A tale of two studies

    A tale of two studies

    For those of us interested in large-scale social science research, this summer was dizzying. The promise of this…

    2 条评论
  • In defense (praise?) of silver bullets

    In defense (praise?) of silver bullets

    I still remember my first day of my first job out of the classroom in 2007. After 3 years of teaching middle school ELA…

    6 条评论
  • "What did we get for it?"

    "What did we get for it?"

    If you haven't yet listened to Michael Horn, Diane Tavenner and Stacey Childress's most recent Class Disrupted podcast…

    11 条评论
  • Our randomized controlled trial journey

    Our randomized controlled trial journey

    As companies and organizations grow and ossify, they tend to take fewer risks and make decisions more slowly. There's a…

    26 条评论
  • Writing Well

    Writing Well

    In business we generally preach concision, concision, concision. Ok, maybe just one concision :) Mostly it's because…

    1 条评论
  • The jig is up

    The jig is up

    It's counterintuitive that K12 education has numerous studies demonstrating programs and products with strong positive…

    12 条评论
  • Is assessment holding back the Science of Reading?

    Is assessment holding back the Science of Reading?

    I've wanted to write this post for a while, but haven't found the right impetus. I'll settle for a recent update from…

    5 条评论

社区洞察

其他会员也浏览了