Edodo
All posts

Published January 15, 2026

The Formative Assessment Revolution: How AI Closes Learning Gaps in Days, Not Weeks

You don't have a formative assessment problem.

You have a feedback latency problem.

In international schools, your students learn fast.

Your timetable moves faster.

And your marking pile moves… whenever it can.

So the kids who most need timely feedback often get it last, not first.

By the time they see your comments, the class has already shifted to the next concept, the next text type, the next lab.

That isn't a pedagogy issue.

It's a bandwidth issue.

And that's why AI is suddenly showing up in formative assessment conversations—not because it "grades faster," but because it can close the loop while learning is still happening.


What's actually at stake (for you, not for policy documents)

If you're a teacher, coach, or curriculum leader, you already know the line:

"Assessment for learning."

But in real life, the version you live with is closer to:

"Assessment after learning… if I survive the week."

The consequences are predictable.

You reteach too late.

Students practise the wrong thing for too long.

The confident students get even more confident.

The quiet strugglers go invisible.

And your best professional judgement gets replaced by a crude proxy: the final score.

Black & Wiliam's classic review is blunt about what works: formative assessment improves achievement, with typical effect sizes reported in the 0.4 to 0.7 range, and it can particularly help low achievers, narrowing spread while raising overall performance.

The problem is not belief.

It's execution at scale.


The "black box" problem isn't theoretical

Black & Wiliam describe classrooms as a "black box": systems try to change inputs (standards, tests, policies) and hope outputs improve, without enough attention to what happens inside the classroom minute-to-minute.

Their key move is simple:

Formative assessment isn't "more quizzes."

It's using evidence to adapt teaching and learning activities—in time for that adaptation to matter.

That definition matters in AI conversations.

Because most AI use in schools right now isn't formative assessment.

It's automated scoring.

Those aren't the same thing.


The real opportunity: AI can shorten the formative cycle

The U.S. Department of Education's Office of Educational Technology frames the opportunity cleanly: AI can enhance feedback loops, but only if we keep "humans in the loop" and treat AI as augmentation, not replacement.

This is the key shift:

You don't use AI to outsource teaching decisions.

You use AI to surface better evidence faster so that your decisions improve.

That's the move that fits international schools: high expectations, complex learners, multilingual classrooms, and high relational demands.


The 4-Step AI Formative Loop (that actually works in real schools)

Below is a practical loop you can run in any subject area.

It stays aligned with Wiliam's five strategies and avoids the trap of "AI = grading bot."

1) Plan the target (not the task)

Start by naming the learning target and what success looks like.

Not "write a paragraph."

But "use evidence to support a claim" or "justify a method choice."

This mirrors the "plan learning targets → interpret evidence → adjust goals" cycle described in formative assessment process models.

2) Elicit evidence from everyone (fast)

Use short, diagnostic prompts.

Not long assignments.

You want something every student can do in 3–6 minutes.

Wiliam's work emphasizes engineering questions/tasks that elicit evidence and making participation visible (so "radiator kids" can't hide).

A simple classroom technique example is "traffic lights" (green/yellow/red) to surface understanding and prompt metacognition.

3) Let AI do the first-pass pattern finding

This is the highest-leverage use of AI:

Not final grades.

Not judgement.

Pattern detection.

Have AI cluster responses into:

  • correct concept, weak explanation
  • misconception A
  • misconception B
  • language barrier vs conceptual barrier (you still verify)

Then you, the teacher, review the patterns quickly.

The point is that you're no longer reading 26 near-identical errors one by one before you notice the trend.

You see the trend first.

4) Close the loop with "feed forward"

Feedback must create thinking.

Wiliam's summary of the research on feedback is clear: feedback can even make performance worse when it becomes ego-involving (grades, comparison), and comments-only tends to outperform grades-with-comments because students focus on the grade first.

So use AI to draft feedback in a "next step" format.

But you keep:

  • the relational tone
  • the contextual knowledge
  • the follow-up task design

Then build a 10-minute "redo" window in the next lesson.

No redo time = no formative impact.


Objection handling (the honest concerns leaders raise)

"AI feedback will be generic."

It will—if you ask for generic feedback.

But Wiliam's own framing of effective feedback is "feedback that moves learning forward," not feedback that sounds polished.

Generic praise is useless.

Specific next steps tied to success criteria are useful.

AI is often better at consistency.

You are better at meaning.

"What about bias and safety?"

That concern is valid.

The U.S. Department of Education explicitly warns about risks like bias, privacy, and over-automation, and recommends transparency, human oversight, and assessment expertise to reduce bias.

So the leadership move is not "ban or buy."

It's: pilot with guardrails.


Students don't learn what you teach—they learn what they retrieve

One reason formative assessment works is that it forces retrieval.

And retrieval changes memory.

Roediger & Karpicke summarize decades of evidence on the "testing effect": retrieval practice enhances later retention more than additional study, even without feedback in many cases.

That matters because AI makes high-frequency, low-stakes retrieval feasible without wrecking teacher workload.

But only if you close the loop.


What changes in 30 days if you do this well

Imagine four weeks from now.

You walk into class already knowing:

  • which misconception is rising
  • which students are bluffing
  • which students need a language scaffold vs a concept reteach

Your feedback stops being a weekend activity.

It becomes a lesson design input.

Students stop asking, "Is this good?"

They start asking, "What's my next move?"

That's not fantasy.

That's what happens when feedback latency drops and the loop closes.


Getting started

Pick one lesson this week.

Do one micro-check.

  1. Write one learning target in student-friendly language.
  2. Ask one diagnostic question (short response, not MCQ only).
  3. Use AI to cluster responses into 3–5 patterns.
  4. Start the next lesson with: "Here are the top two patterns. Let's fix them."

Timebox it.

One cycle.

One class.

Then scale.