Research Studies

How Tutor Co-Pilot Systems Scale Teaching Capacity Worldwide

Stanford’s National Student Support Accelerator ran the largest randomized trial to date. Researchers embedded tutor co-pilot systems within 900 tutors serving 1,800 students. Overall mastery rose four percentage points over control groups. Moreover, students paired with lower-rated tutors gained nine points. World Bank teams replicated positive effects in Nigerian secondary English classes. The AI assistant there delivered 0.31 standard deviation growth within six weeks. Consequently, analysts equated the short program to almost two years of schooling.

Tags

Why One-on-One Tutoring Outperforms Two-on-One

A new study by Stanford researchers Hsiaolin Hsieh, David Gormley, Carly D. Robinson, and Susanna Loeb suggests why one-on-one tutoring has been found to produce double the gains in student learning than two-on-one tutoring.

Analyzing 16,629 transcripts from 2022-23 school year tutoring sessions from an earlier study that established the greater gains under one-on-one tutoring, the researchers examined how tutors allocated their time and attention across both one-on-one and two-on-one formats. The tutoring sessions focused on early literacy and served kindergarten through second grade students, with 510 students receiving one-on-one tutoring and 570 students receiving two-on-one tutoring. All students met with their tutor online for 20 minutes during the school day, four times per week.

Why Hasn’t Tutoring Been More Effective?

The most recent of these, from researchers at Stanford University’s SCALE Initiative, examined math and reading tutoring programs in a large, urban district during the 2022-23 and 2023-24 school years. Neither led to overall gains in academic achievement.

But when researchers dug deeper into the data, they identified implementation problems that could be driving these null effects.

The findings add to a growing body of evidence around tutoring in a post-COVID landscape that suggests the effectiveness of a program hinges on the nitty-gritty details of how it is run—how often students meet with their tutors, for instance, or whether lessons are tailored to their specific needs.

Studying these implementation details could help school systems build more effective tutoring initiatives in the long run, said Elizabeth Huffaker, an assistant professor of educational leadership and policy at the University of Florida, and the lead author on the SCALE paper.

Tutor CoPilot: A Human-AI Approach for Scaling Real-Time Expertise

Generative AI, particularly Large Language Models (LLMs), can expand access to expert guidance in domains like education, where such support is often limited. We introduce Tutor CoPilot, a Human-AI system that models expert thinking to assist tutors in real time. In a randomized controlled trial involving more than 700 tutors and 1,000 students from underserved communities, students with tutors using Tutor CoPilot were 4 percentage points more likely to master math topics (p0.01). Gains were highest for students of lower-rated tutors (+9 p.p.), and the tool is low-cost (about $20/tutor/year). Analysis of over 350,000 messages shows Tutor CoPilot promotes effective pedagogy, increasing the use of probing questions and reducing generic praise. In this work we show the potential for human-AI systems to scale expertise in a real-world domain, bridge gaps in skills, and create a future where high-quality education is accessible to all students.

Lessons from a Failed Texas Tutoring Program

Experts view the findings as a cautionary tale of how tutoring can go wrong.

The district had to wait on background checks for tutors, many students were still chronically absent and the tutoring sessions often conflicted with other lessons or special events. As a result, students didn’t receive the 30 hours or more required under a state law mandating tutoring for those who failed the annual state test. Instead of five days a week as planned, 81% of the students attended tutoring three or fewer days, and most students worked with a different tutor every time they attended a session.

The findings reinforce the importance of protecting the time students are supposed to receive tutoring, said Elizabeth Huffaker, an assistant professor of education at the University of Florida and the lead author of the study.

High-dosage models — featuring individualized sessions held at least three times a week with the same, well-trained tutor — can still “drive really significant learning gains,” she said, “but in the field, things are always a little bit more complicated.”

Tags

Virtual tutoring is here to stay. New research points to ways to make it better.

Two studies from Stanford University’s National Student Support Accelerator released Wednesday used natural language processing technologies to review transcripts from tens of thousands of hours of virtual tutoring sessions. Their goal: to better understand exactly what happens between tutors and students in these sessions.

One study examines the impact of disruptions as revealed through tutor comments, such as “You can’t see me? I’m not sure why you can’t see me” or “Sorry. Did you say something? It was hard to hear.”

Researchers found that 19% of available time was lost to disruptions, whether from technological issues, distracted students, or background noise. Time lost to disruptions was even greater when tutors were working with more than one student, especially if one of the students entered the session late.

Tags

How is ChatGPT impacting schools, really? Stanford researchers aim to find out

A new collaboration between Stanford’s SCALE and OpenAI, the creator of ChatGPT, strives to better understand how students and teachers use the popular AI platform and how it impacts learning

Education is one of the fastest-growing use cases of AI products. Students log on for writing assistance, brainstorming, image creation, and more. Teachers tap into tools like attendance trackers, get curriculum support to design learning materials, and much more.

Yet despite the rapid growth – and potential – a substantial gap remains in knowledge about the efficacy of these tools to support learning. 

A new research project from the Generative AI for Education Hub at SCALE, an initiative of the Stanford Accelerator for Learning, aims to help fill that gap by studying how ChatGPT is used in K-12 education. In particular, the research will examine how secondary level teachers and students use ChatGPT. 

What’s in a Contract? How Outcomes-Based Contracting Reshapes School District–Vendor Relationships

In this study, we analyze the contracts between districts and vendors of instructional services and products to understand how relationships between these parties are structured. We compare three types of contracts: those developed with the support of SEF’s Outcomes-Based Contracting (OBC) Cohort program, those between the same districts and other vendors without SEF support, and those involving the same vendors but with other districts that did not receive SEF assistance. During the cohort experience, participating districts received guidance from SEF’s Center for Outcomes-Based Contracting. The total cost of hosting each district in the cohort was $30,000, of which districts contributed $15,000, with the remaining expenses covered by SEF through philanthropic funding. We use the emerging OBC framework as a baseline to understand the extent to which traditional district–vendor contracts already incorporate elements of the OBC approach and other information pertinent to vendor quality and alignment.

The Impact of High-Impact Tutoring on Student Attendance: Evidence from a State Initiative

This study provides compelling evidence that tutoring can do more than boost test scores; it can actually get students back in the classroom. On average, students were 1.2 percentage points less likely to be absent on days when they were scheduled to receive tutoring, suggesting that they are motivated to participate in tutoring. This impact was even greater for middle schoolers and students who’d missed more than 30% of school days the prior year. The study also found that the design matters: tutoring only improved attendance when it combined at least two evidence-based features like small groups, frequent sessions, and in-school delivery.