AI for assessment
AI for assessment
Generative AI can support aspects of assessment design by helping learning designers draft assessment ideas, generate examples, propose question formats, and refine instructions or criteria language. It can be a useful drafting partner, but it should not be relied upon uncritically for validity, fairness, or alignment.
Assessment design has direct consequences for workload, evidence quality, learner experience, and academic integrity. For that reason, AI should be used as a support tool for exploring or refining assessment ideas, not as an authority on whether an assessment is educationally sound.
Where AI can help
AI can assist with:
- drafting assessment task ideas
- generating question stems
- proposing rubric language
- refining assessment instructions
- generating examples of evidence or performance
- suggesting formative assessment activities
- checking whether an assessment appears aligned to stated outcomes
These uses are most valuable when AI is helping to surface options and language, rather than determining assessment quality by itself.
Good practice
When using AI for assessment:
-
Begin with the learning outcomes
Ensure the assessment is tied to what learners are actually expected to know or do. -
Use AI to generate options
Ask for alternative task formats, clearer wording, or possible criteria rather than assuming the first output is suitable. -
Review validity and authenticity
Check whether the task genuinely measures the intended outcome and reflects meaningful evidence of learning. -
Refine the language
AI can help make instructions, questions, and rubric statements clearer, but clarity should not come at the expense of precision. -
Check fairness and level
Make sure the assessment is appropriate for the learner group, the course level, and any programme or industry expectations.
Risks and limitations
AI can produce assessment content that appears polished but is weak in important ways. Common problems include:
- shallow alignment to learning outcomes
- poor or inconsistent cognitive level
- unrealistic or generic task contexts
- criteria that sound good but are difficult to apply reliably
- invented standards or references
- assessment ideas that are too broad, too narrow, or too easy to game
These risks make human review essential, especially for summative assessment.
Example uses
Example 1: Improving assessment instructions
A designer has a strong task idea but the instructions are too wordy or unclear. AI can suggest a cleaner structure with:
- purpose
- task instructions
- expected output
- submission requirements
- success criteria summary
The designer then checks that the final wording remains accurate and institutionally appropriate.
Example 2: Drafting rubric language
AI can help convert rough notes into draft criteria language, such as turning “shows understanding of audience and purpose” into more explicit descriptors. The designer still needs to test whether the rubric is usable and aligned.
Relationship to other design work
Assessment design should remain connected to writing learning outcomes and objectives, course level alignment, summative assessment planning, and tasks and activities.

Example: an AI-generated assessment design image showing rubric drafting and thinking about authentic evidence of learning.
Practical guidance
AI can be helpful in assessment design when it is used to:
- generate alternatives
- improve wording
- draft structures
- stress-test alignment
- produce first-pass rubric or question language
It should not be treated as the final authority on assessment quality, validity, or compliance.
Use AI in assessment design as a drafting and review aid, while keeping final responsibility for alignment, fairness, and quality with the learning designer.
No comments to display
No comments to display