Monthly Archives: April 2026

AI-Assisted Assessment in Practice

This week I sat down with two Grade 7 students to conference on a math performance task, and we did something I want to make a habit of: we all read the feedback together and each of us annotated the documents as we went — their observations alongside mine.

The task and the evaluation were both AI-generated — a rooftop garden design problem spanning fractions, algebra, data analysis, and budgeting, aligned to all four categories of the Ontario Growing Success Achievement Chart. The feedback was detailed and criterion-referenced, with specific commentary for each question and conference prompts prepared in advance. Our focus was on that feedback loop: what did the assessment reveal, what did it miss, and what did the students themselves have to say about it.

As the three of us read through the documents together and something stopped us. The students had originally submitted a colour-coded grid diagram, but the AI had assessed the PDF without picking up the colour, effectively missing the diagram entirely and assuming it hadn’t been completed. Reading through the feedback together, the students flagged it.

We looked at the original, counted the squares section by section, and it became clear that every area matched their proposed redesign from Part B exactly — vegetables at 24 squares, herbs at 16, pollinator at 14, seating at 6, totalling 60 squares and 240 m² without a remainder. To produce that, you have to know your target area, divide by 4 m² per square, then design a compact rectangular region with that exact integer count, and do it for all four sections at once. That is spatial and proportional reasoning applied deliberately to a real constraint.

So we did what made sense: we redid the diagram with proper labels and resubmitted it for re-evaluation. The AI had flagged their use of the redesigned areas throughout the rest of the task as a cascading error — with the labelled diagram now in front of us, the re-evaluation told a different story.

Their answer was essentially: we proposed a new garden design in Part B, so we applied that new design everywhere — to the tray calculation, the budget, the diagram, all of it. If the plan changed, the whole plan should change.

The AI had read that as an error — careless copying of the wrong values. But when you heard them explain it, the logic held up. They weren’t being careless. They were being consistent. They made a design decision and followed it through. That is a different thing entirely, and it took the three of us together to surface it. That reasoning is coherent. It also cost them marks because the task instructions pointed back to the original areas, and we talked through exactly why, and what they’d watch for next time. Then they wrote their own feedback into the document alongside mine — what made sense, what they’d do differently, what they knew they’d got right.

The AI produced feedback precise enough to be worth conferencing over, but it was the three of us together — reading, questioning, counting squares — that turned the assessment into actual learning. One of the students explained their design decision and she was right about the logic, even though the task had different expectations. No model caught that. We did.

AI in the classroom isn’t about replacing the teacher. It’s about raising the floor of what’s possible so the ceiling — the conversation, the moment of genuine inquiry, the student explaining their reasoning to an adult who is actually listening — can happen more often, with better material underneath it.


What the AI Did Well — and What Only We Could Do

Grade 7 Mathematics · Rooftop Garden Performance Task
 

What AI contributed

Designed a rich, multi-strand performance task aligned to the Ontario curriculum in a fraction of the time
Generated criterion-referenced feedback for every question with strengthening and extending comments
Prepared conference prompts and next steps in advance
Re-evaluated when the diagram was resubmitted with labels — revised the score and the interpretation
Raised the floor: made a high-quality assessment process possible in the time available
 

What only we could do

Noticed the colour-coded diagram had been missed because the AI read a PDF and couldn’t see colour
Counted the squares together and understood what the precision of the layout actually meant
Asked the question: was this a deliberate design decision, not an error?
Understood that what the AI called a cascading error was actually a deliberate design decision — once they changed the plan, they applied it everywhere, and that only became clear when we talked
Co-authored the record: students wrote their own feedback into the document alongside the teacher’s

The ceiling is the conversation. AI raised the floor — better material, more detailed feedback, faster turnaround. But the learning happened when three people sat in a room, read the document together, and one student said something that made the teacher pause. No model catches that. That’s the point.


AI-Assisted Assessment & Feedback Conference

How a performance task, AI evaluation, and a three-way conference produced better learning than any one alone
1
AI-Generated Performance Task
A rooftop garden design problem — fractions, algebra, data analysis, budgeting — aligned to all four Growing Success Achievement Chart categories and the Ontario Math curriculum (2020).

AI builds the floor

2
Students Complete the Task
Two Grade 7 students work through all four parts, submitting written responses, calculations, and a colour-coded grid diagram of their garden layout.

Student work

3
AI-Generated Criterion-Referenced Evaluation
Detailed feedback aligned to each success criterion — what was done well, where marks were lost, descriptive strengthening and extending feedback, and conference prompts prepared in advance.

AI raises the detail

4
Three-Way Feedback Conference
Teacher and both students read the evaluation together, out loud. Each annotates the document. The students flag a colour-coded diagram the AI had missed. Squares are counted. The reasoning behind a design decision is surfaced that no model had identified.

The conversation the document made possible

5
Student Re-Submission
Students redraw the diagram with proper labels and resubmit for re-evaluation. The AI revises its assessment. The score changes. More importantly, the students understand exactly why.

Student agency

6
Assessment as Learning
Students write their own feedback into the document alongside the teacher’s. The record of the learning is co-authored. Growing Success fulfilled — not as compliance, but as genuine practice.

Assessment as learning · Growing Success

 


Rich Baxter is a Grade 7 teacher at John McCrae Public School (TDSB), Founder & CEO of Innovation in Education, and creator of the Lern2ern platform. He is shortlisted for the 2025 QS Reimagine Education Award in Nurturing Employability.

#AssessmentAsLearning #GrowingSuccess #AIinEducation #OntarioEducation #HighImpactPractices #StudentVoice #FeedbackConference #TeacherLeadership #EdTech #HumanPlusAI