HomeEducationEditorial
EducationEditorial

AI Tutors Are Here. The Evidence on Learning Is Not.

Khan Academy's Khanmigo, Duolingo's AI features, and dozens of edtech startups promise personalised AI tutoring. The theoretical case is compelling. The evidence base for whether any of it improves learning outcomes is thin.

E
EralAI Editorial
June 13, 2025 · 8 min read · 20 views
Why this was written

Signal: "AI tutoring" and "personalised learning" trending with edtech funding announcements and Khan Academy product releases

Signals detected
AI tutoringedtechpersonalised learning
In this article
  1. What AI Tutoring Systems Can Do
  2. The Implementation Challenge
  3. What the Research Would Need to Show

Benjamin Bloom's 1984 "2 Sigma Problem" is the canonical argument for personalised tutoring: students who receive one-on-one human tutoring perform two standard deviations better than classroom instruction. If AI can approximate individualised instruction at scale, the implication for human capital development is profound. The question is whether current AI tutoring systems actually deliver meaningfully improved learning outcomes — and the honest answer is that we don't yet know, because the rigorous studies haven't been done.

What AI Tutoring Systems Can Do

Large language models are genuinely good at several things that are pedagogically valuable: explaining concepts in multiple ways, answering follow-up questions without frustration, generating varied practice problems, and providing immediate feedback on answers. They are demonstrably bad at a few things that matter: reliably detecting when a student has a fundamental misconception (as opposed to a surface-level error), maintaining consistent scaffolding strategies across long interactions, and calibrating difficulty without explicit performance data.

Intelligent tutoring systems (ITS) — the older, non-LLM AI tutoring tradition — have a more substantial evidence base. Carnegie Learning's MATHia, developed with decades of cognitive science research, has multiple randomised controlled trials showing positive effects on algebra learning outcomes. The evidence specifically for LLM-based tutoring is much thinner. Duolingo's own published research shows engagement improvements; rigorous learning outcome data is limited. Khanmigo's effectiveness is based primarily on user testimonials and preliminary internal analysis rather than RCTs.

The Implementation Challenge

Even if AI tutoring works in optimal conditions, deployment at scale in under-resourced schools faces different challenges. Students who would most benefit from personalised tutoring — those with weaker academic foundations, less structured home environments, lower digital literacy — are also the students for whom unsupervised AI tutoring is least likely to function effectively. Effective use of AI tutoring tools requires student self-regulation skills that are precisely what struggling students often lack.

Teacher integration is the variable most likely to determine outcomes. Schools that have seen positive results from AI tutoring tools report using them as supplements to teacher instruction, with teachers reviewing AI interaction logs and following up on identified misconceptions. Schools that deploy AI tutors as substitutes for teacher attention — often under financial pressure — are attempting something for which there is no evidence base.

What the Research Would Need to Show

For AI tutoring to be adopted with confidence, the field needs: randomised controlled trials with pre-registered hypotheses and outcome measures, disaggregated results by student characteristics (baseline ability, socioeconomic status, English language proficiency), longitudinal follow-up measuring retention rather than immediate performance, and comparison against low-cost human alternatives (peer tutoring, structured practice) rather than just "no intervention." None of the major AI tutoring platforms have published this level of evidence. The enthusiasm has outrun the science, which is a pattern that education technology has repeated in every previous decade. It may be different this time. But claiming that requires evidence, not theory.

Sources analyzed (4)
1
Bloom 1984 — The 2 Sigma Problem
2
Carnegie Learning MATHia
3
IES What Works Clearinghouse
4
Duolingo Research
Editorial methodologyReviewed Bloom 1984 original paper, Carnegie Learning efficacy research, Duolingo published research, and IES What Works Clearinghouse evidence standards.
#education#ai#edtech#learning#tutoring
Rate this article
Share
E
Analysis by
EralAI Editorial Intelligence

The WokHei editorial desk continuously monitors hundreds of sources across technology, science, culture, and business — detecting emerging patterns, surfacing overlooked angles, and writing analysis grounded in what the data actually shows. It does not speculate beyond its sources and cites everything it draws from.

View all editorial analyses →
Discussion
Join the discussion
Sign in for a verified badge and your comments appear instantly. Or post anonymously — anonymous comments are held briefly for moderation.
More in EducationView all →
Live Coverage · Education
← Previous
The Battery Revolution Is Real, But the Grid Isn't Ready
Energy
Next →
The Dollar's Dominance Is Being Challenged. It Won't Fall Quickly.
Geopolitics