CONVOLEARN: Fine-Tuning Dialogic AI Tutors

Dataset of 2,134 teacher-student dialogues labeled across six learning-sciences dimensions to train and evaluate dialogic tutoring behavior in LLMs.

Most tutoring LLMs optimize for being “helpful”—which usually means giving answers directly. That’s at odds with effective tutoring, which draws out student thinking rather than replacing it. This project builds training data grounded in learning-sciences theory of dialogic tutoring to both evaluate and improve AI tutor behavior.

What’s in it

CONVOLEARN is a dataset of 2,134 tutor-student dialogues (~20 turns each) with dimension-level labels across six core aspects of dialogic tutoring, grounded in knowledge-building theory:

  • Cognitive engagement — prompting reasoning, not answers
  • Formative assessment — probing understanding in real time
  • Accountability — holding students to disciplinary norms
  • Cultural responsiveness — recognizing diverse ways of knowing
  • Metacognition — making thinking visible
  • Power dynamics — distributing authority

These six dimensions break down into 21 subdimensions, all situated in middle school Earth Science. Teacher turns are authored by credentialed teachers; student turns are simulated.

Approach

Dimension-labeled dialogues support classifier training, contrastive learning, and model fine-tuning.

Unlike binary “good/bad” labels, the dataset preserves quality spectra across dimensions. This supports multiple uses: training classifiers to evaluate dialogue quality, contrastive learning to surface pedagogical patterns, and direct model fine-tuning.

Results

Classifiers trained on CONVOLEARN correlate with expert-coded instructional quality in authentic classroom transcripts (r = .118–.258, all p < .05)—evidence that semi-synthetic training data captures real pedagogical signal. Fine-tuned Mistral-7B shows measurably more dialogic behavior, rated by credentialed teachers as competitive with strong proprietary baselines.

Why it matters

Most tutoring benchmarks measure correctness and fluency—not whether models actually support learning. CONVOLEARN provides open infrastructure for pedagogically aligned evaluation, with training signal to steer models away from “helpful assistant” defaults toward dialogue that builds understanding.