AI in Assessment and Feedback

Lessons learnt from the Jisc AI Assessment Pilot

Feedback is one of the most powerful tools an educator has. John Hattie described it as “the most powerful single influence on student achievement.” Yet across every sector of education, delivering feedback well remains one of the hardest things to do consistently.

Watch the webinar

Watch the full webinar recording for the complete discussion and sector perspectives behind the key themes explored in this article.

The reasons are familiar. Marking is time-consuming, cognitively demanding and relentless. In many cases, educators are spending up to 10 hours per week on assessment-related tasks, with peaks significantly higher during key submission periods. When workloads are this heavy, something has to give. Too often, it is the quality, depth or timeliness of feedback that suffers.

At the same time, AI is already being used in assessment, often informally and without clear institutional guidance. The real risk is no longer whether AI is used. It is whether it is used in a consistent, governed and pedagogically sound way.

A sector at a turning point

Across schools, colleges and universities, AI adoption is accelerating, but guidance has not kept pace. In many cases, AI is being used under the radar, without shared expectations or clear boundaries. This creates inconsistency across teams and raises important questions around safeguarding, fairness and professional standards.

The challenge for institutions is not whether to engage with AI. It is whether adoption will be unstructured and reactive, or deliberate and strategically led.

A national pilot grounded in real practice

The Jisc AI Assessment Pilot brought together a group of colleges, universities and post-16 providers from across the UK to explore how AI could be used responsibly in assessment and feedback.

Crucially, this was not a controlled experiment. Educators embedded AI into live teaching and assessment workflows, testing its impact on feedback quality, consistency across cohorts, turnaround time, and staff workload.

“Workload reduction on its own isn’t necessarily something that is a priority to students…”

“AI is very good… at providing larger volumes of feedback. The timeliness… is something that’s really important.”

Tom Moule, Senior AI Specialist, Jisc

What emerged was something the sector has been lacking: a shared, evidence-informed understanding of AI in assessment, grounded in real practice.

The feedback gap: what needs to change

Before looking at solutions, the pilot reinforced a long-standing issue. There is a persistent gap between what feedback should achieve and what it often delivers.

In practice, feedback struggles in three key areas:

  • Clarity – Students are not always clear on how to improve.
  • Usefulness – Feedback can feel generic or retrospective.
  • Timeliness – Feedback often arrives too late to influence learning.

This is not simply an operational challenge. It is a pedagogical one. When feedback is delayed by weeks, students may no longer clearly remember their thinking, making it significantly harder to apply that feedback to future work. Timely feedback is not a convenience. It is essential to learning.

What the pilot revealed so far…

Tom Moule, Senior AI Specialist, Jisc

One of the most important insights from the pilot was that workload reduction alone is not the main driver.

This reframes the conversation. The value of AI is not simply that it saves time. It is that it enables more consistent marking, greater depth of feedback, and faster turnaround for students.

“The feedback is generally consistent… there isn’t a zero variance, but there is a very low variance.”

Tom Moule, Senior AI Specialist, Jisc

For leaders concerned with quality assurance and standardisation, this is significant.

Why formative assessment led the way

The pilot also showed where AI is currently most effective. Formative assessment naturally emerged as the primary use case. In this context, AI enabled faster feedback cycles, more detailed commentary, and increased opportunities for student improvement.

This is where the impact on learning becomes most visible.

Improving workflow without adding complexity

Peter Kilcoyne, Managing Director, TeacherMatic

A major barrier to adoption is not capability, but workflow friction. If AI tools add complexity, they will not be used.

“The Advanced Feedback Generator has vastly improved workflows.”

Peter Kilcoyne, Managing Director, TeacherMatic

The ability to process multiple submissions and generate structured feedback at scale transformed what had previously been a time-intensive process. This reinforces a key principle: AI must integrate into existing practice, not sit alongside it.

Why general AI alone is not enough

A key question raised during the pilot was simple: why not just use tools like ChatGPT or Copilot?

In practice, general AI assumes a level of expertise that is not always present. It relies on educators to construct effective prompts, apply pedagogical frameworks, and interpret and refine outputs. At scale, this leads to inconsistent outputs, uneven practice across teams, and challenges in governance and oversight.

What the pilot demonstrated is that success depends not just on the AI itself, but on how it is structured, guided and embedded into teaching practice.

Human judgement remains central

Across all institutions, there was clear alignment on one principle: AI should support professional judgement, not replace it.

“These tools are not to be used to automate and delegate the marking process. They are there to support and to speed up that process.”

Tom Moule, Senior AI Specialist, Jisc

Every successful implementation observed in the pilot shared one non-negotiable principle: the educator remains the final decision-maker.

 

Real impact in the classroom

Alessio Corso, Associate Professor

Working with large cohorts of first-year engineering students, Alessio Corso used AI-supported feedback within a structured, rubric-driven approach. He developed detailed criteria aligned to specific components of the assignment and used AI to support consistent, criteria-based feedback across the cohort.

“What really impressed me was, first of all, the comprehensiveness of the feedback…”

Alessio Corso, London South Bank University

AI-supported feedback provided structured comments aligned to assessment criteria, clear identification of strengths and areas for improvement, and greater consistency across a large cohort. This level of structured feedback would be difficult to replicate consistently at scale using traditional marking approaches alone.

Crucially, Alessio maintained full oversight, reviewing outputs to ensure alignment with academic standards and his own professional judgement.

Lisa Sayce, PCET Lecturer

From a Further Education perspective, Lisa Sayce’s experience highlighted both the practical efficiency gains and the pedagogical impact.

“The consistency and the quality of my feedback has been really, really enhanced from using TeacherMatic.”

Lisa Sayce, Further Education Lecturer

Working with trainee teachers submitting regular lesson reflections, Lisa used AI to support high-volume feedback tasks that would traditionally require significant time. The result was not just faster marking, but better feedback.

Feedback became more structured, more detailed, and more aligned to professional expectations. AI supported her in identifying patterns, linking theory to practice, and providing more targeted developmental feedback.

“The consistency and the quality of my feedback has been really, really enhanced from using TeacherMatic.”

Lisa Sayce, Further Education Lecturer

“Students have said it feels more personalised and constructive than feedback they’ve received in other contexts.”

Lisa Sayce, Further Education Lecturer

This is particularly significant. Feedback is often criticised for being impersonal, yet here it was perceived as more personalised, clearer and more actionable.

Stephen Clarke, Head of Digitally Enhanced Learning

“It’s a very, very easy to grasp the tool… but they are very powerful tools.”

Stephen Clarke, London South Bank University

If AI requires advanced technical expertise, adoption will remain limited. If it supports and scaffolds practice, it becomes accessible across teams.

Adoption: the real challenge

The pilot also highlighted a key reality. Most educators are not early adopters. They sit within the persuadable middle: open to innovation, but cautious, and requiring reassurance and evidence.

The challenge is not convincing innovators. It is supporting the majority of staff who need clarity, confidence and practical guidance.

Why the SAFE Framework was developed

Esam Baboukhan, Director, TeacherMatic

“AI-supported feedback is emerging at pace… but the honest reality is that guidance has not kept up.”

Esam Baboukhan, TeacherMatic

“AI should support professional judgment and not replace it.”

Esam Baboukhan, TeacherMatic

One of the clearest challenges identified through the pilot was not capability, but uncertainty. AI-supported feedback is evolving rapidly, yet guidance has not kept pace.

Educators are asking: What is acceptable? Where are the boundaries? How do we ensure fairness and consistency? This matters because assessment sits at the heart of academic integrity, professional judgement, and institutional trust.

This is why the SAFE Framework was developed. Not as a theoretical model, but as a practical, institution-ready starting point.

Introducing the SAFE Framework

SAFE is built around four core principles:

  • Safeguarding data and privacy
  • Augmenting professional judgement
  • Fairness and inclusion
  • Ethical and transparent practice

Its purpose is simple: to help institutions move from uncertainty to confident, structured adoption of AI in assessment and feedback.

Continue the conversation

Help shape responsible AI in assessment

If you are exploring AI in assessment, this is an opportunity not just to adopt, but to help shape how it is implemented across the sector.