The jig is up

The jig is up

It's counterintuitive that K12 education has numerous studies demonstrating programs and products with strong positive effects (see Johns Hopkins University School of Education Evidence for ESSA and U.S. Department of Education What Works Clearinghouse) and yet NAEP (national assessment) scores have been virtually flat for five decades. In medicine, drugs with strong efficacy change health outcomes at scale. Why don't we see this in education?

If you're starting from first principles, you might wonder:

  • Perhaps education studies just aren't that replicable. A product or intervention with a strong effect in one place might not produce the same effect at scale with uneven implementations.
  • Perhaps there's an information problem in the education market and despite best efforts of initiatives like Evidence for ESSA, the products and services with the best results don't spread and instead schools spend limited budget on slickly marketed products with inferior results.
  • Perhaps the most effective education products and services are so expensive that they can't scale even if they work. In education there's nothing like the US health insurance system that can scale extremely expensive solutions.

All three scenarios are pretty dispiriting. But, Laurence Holt just pointed out and Dan Meyer amplified (pun intended) a new scenario that I hadn't considered, The 5 Percent Problem.

Laurence Holt asks the same question I did:

Those gains, and many others like them reported each year, are impressive. Since use of these tools is widespread, one could be forgiven for asking why American students are not making impressive gains in math achievement. John Gabrieli, an MIT neuroscientist, declares himself “impressed how education technology has had no effect on . . . outcomes.”

He digs into a Khan Academy study with an impressive 0.26 standard deviation effect that only applied to the 4.7% of students who used the program 30 minutes per week as recommended, and goes on to cite several other similar studies.

If education studies sometimes feel like going down the rabbit hole, Laurence Holt 's discovery feels like we've reached the bottom, and now we might know how to go back up. When education providers were gaming uncontrolled studies, we placed a premium on controlled trials. But it turns out you can game controlled trials too by excluding nearly all the students from the results.

Districts want interventions that significantly impact all of their targeted students, and if students aren't motivated enough to engage with the intervention in the first place, districts shouldn't be persuaded that the intervention is effective.

The school year is ending, which means it's results season! As researchers and companies publish studies let's keep a close eye on how much of the treatment population is included in the results.




Andy M.

ML/AI, Data & Product

9 个月

Nice post! If you haven't listened to John List I think you would enjoy his discussion of the Voltage Effect https://podcasts.apple.com/us/podcast/john-list-on-scale-uber-and-the-voltage-effect/id135066958?i=1000571078949

赞
回复
David Harper

Data Scientist | Quantitative & Mixed-Methods Researcher | Research Manager | Linguist | PhD

9 个月

This is a nice summary of recent discussions, and it highlights why the WWC prefers studies that employ an intent-to-treat (ITT) analysis. ITT analysis is important because it helps to avoid overestimating the effectiveness of an intervention, and it accounts for non-compliance and non-adherence that often occur in real-world situations. This is also known as “Once randomized, always analyzed."

Dr. Rachel Schechter

Founder of Learning Experience Design (LXD) Research

9 个月

There are all types of product research. Some tell you the recipe for success (hopefully replicable) and others let you know if something new is better than what is happening with typical materials. Each have their value - and we need to keep the sample sizes clear, so the reader knows who is being cut and why.

Dayna Laur, Ed.D.

Founding Partner at Project ARC, PBC | Transforming Educator Professional Learning

9 个月

Most EdTech is junk and falls into the substitution level of SAMR. I blame developers for not understanding the underlying pedagogical foundations that have the power to transform education. However, I also blame organizations such as ISTE for allowing these companies to push their crap products at their conference. 95% of what is on that conference floor is not worth a fraction of the companies' prices. A few years ago, I participated in a famous study that I dropped out of because the researchers told the teachers to do something I completely disagreed with. That action completely negated the purpose of the study and introduced a variable that was NOT reported in the final results. Yet, that study receives a ton of press as an exemplar, making my blood boil. While not explicitly related to EdTech, I also conducted a systematic literature review of 747 studies and found that 79% of the researchers had zero clue about what they were conducting research about. The results were all published and, in many cases, highly touted. Yet, the underlying pedagogical foundations were utterly inaccurate. Intuitively, I believe there is a correlation between a lack of pedagogical understanding and EdTech product development.

Andrew Kaiser

Founder/CEO Educated AI ~ School Principal (Retired) ~ LearningGarden.ai

9 个月

Technology is a tool and not a pill to swallow. Edtech is only as good as the educator makes it. Tutorbots and all the magic of AI won't change this as humans are both predictable and unpredictable at the same time. Good luck figuring us out AI. ??

要查看或添加评论,请登录

Matt Pasternack的更多文章

  • I don't think we need to blow up how we measure schools...yet

    I don't think we need to blow up how we measure schools...yet

    Back in the spring I asked whether assessment is holding back the Science of Reading. I picked on DIBELS, writing: In…

    2 条评论
  • How can districts respond to the cyber charter earthquake?

    How can districts respond to the cyber charter earthquake?

    If you read K-12 headlines, and I asked you what's the biggest earthquake hitting schools this year, you might say:…

    4 条评论
  • Tutoring programs are not widgets

    Tutoring programs are not widgets

    Matthew A. Kraft Beth Schueler Grace Falken just released an important meta-analysis of 282 randomized controlled…

    5 条评论
  • Tech and Edtech

    Tech and Edtech

    When I was getting started in education reform, there was Linda Darling-Hammond saying "STOP". When I was getting…

    5 条评论
  • A tale of two studies

    A tale of two studies

    For those of us interested in large-scale social science research, this summer was dizzying. The promise of this…

    2 条评论
  • In defense (praise?) of silver bullets

    In defense (praise?) of silver bullets

    I still remember my first day of my first job out of the classroom in 2007. After 3 years of teaching middle school ELA…

    6 条评论
  • "What did we get for it?"

    "What did we get for it?"

    If you haven't yet listened to Michael Horn, Diane Tavenner and Stacey Childress's most recent Class Disrupted podcast…

    11 条评论
  • Our randomized controlled trial journey

    Our randomized controlled trial journey

    As companies and organizations grow and ossify, they tend to take fewer risks and make decisions more slowly. There's a…

    26 条评论
  • Writing Well

    Writing Well

    In business we generally preach concision, concision, concision. Ok, maybe just one concision :) Mostly it's because…

    1 条评论
  • Is assessment holding back the Science of Reading?

    Is assessment holding back the Science of Reading?

    I've wanted to write this post for a while, but haven't found the right impetus. I'll settle for a recent update from…

    5 条评论

社区洞察

其他会员也浏览了