Skip to main content

AI for assessment

AI for assessment

Generative AI can support aspects of assessment design by helping learning designers draft assessment ideas, generate examples, propose question formats, and refine instructions or criteria language. It can be a useful drafting partner, but it should not be relied upon uncritically for validity, fairness, or alignment.

Where AI can help

AI can assist with:

    drafting assessment task ideas generating question stems proposing rubric language refining assessment instructions generating examples of evidence or performance suggesting formative assessment activities checking whether an assessment appears aligned to stated outcomes

    These uses are most valuable when AI is helping to surface options and language, rather than determining assessment quality by itself.

    Good practice

    When using AI for assessment:

      Begin with the learning outcomes
      Ensure the assessment is tied to what learners are actually expected to know or do.

      Use AI to generate options
      Ask for alternative task formats, clearer wording, or possible criteria rather than assuming the first output is suitable.

      Review validity and authenticity
      Check whether the task genuinely measures the intended outcome and reflects meaningful evidence of learning.

      Refine the language
      AI can help make instructions, questions, and rubric statements clearer, but clarity should not come at the expense of precision.

      Check fairness and level
      Make sure the assessment is appropriate for the learner group, the course level, and any programme or industry expectations.

      Risks and limitations

      AI can produce assessment content that appears polished but is weak in important ways. Common problems include:

        shallow alignment to learning outcomes poor or inconsistent cognitive level unrealistic or generic task contexts criteria that sound good but are difficult to apply reliably invented standards or references assessment ideas that are too broad, too narrow, or too easy to game

        These risks make human review essential, especially for summative assessment.

        Example uses

        Example 1: Improving assessment instructions

        A designer has a strong task idea but the instructions are too wordy or unclear. AI can suggest a cleaner structure with:

          purpose task instructions expected output submission requirements success criteria summary

          The designer then checks that the final wording remains accurate and institutionally appropriate.

          Example 2: Drafting rubric language

          AI can help convert rough notes into draft criteria language, such as turning “shows understanding of audience and purpose” into more explicit descriptors. The designer still needs to test whether the rubric is usable and aligned.

          Practical guidance

          AI can be helpful in assessment design when it is used to:

            generate alternatives improve wording draft structures stress-test alignment produce first-pass rubric or question language

            It should not be treated as the final authority on assessment quality, validity, or compliance.

            Use AI in assessment design as a drafting and review aid, while keeping final responsibility for alignment, fairness, and quality with the learning designer.