GenAI: Design
This section on Design is for where you have a longer time frame and it is possible to design the assessment from scratch or fully redesign the existing assessment.
As GenAI tools continue to advance, it is becoming increasingly difficult to detect instances of GenAI misuse in assessments and detection will become increasingly difficult, even impossible. Dawson (2021) argues that if we set restrictions on technologies that are not feasible to detect, the integrity and validity of assessment is likely to be reduced, not enhanced. Therefore, attempting to create assessments that GenAI tools cannot complete, or where we prohibit use, is not a sustainable approach to assessing. Instead, assessments should provide opportunities for students to utilise GenAI tools to support their knowledge and skill development (Rudolph, et al., 2023).
When redesigning assessments, make students aware of how GenAI can and cannot inform the ways in which they respond to the task. This approach will instruct them in the utilisation of current tools and technologies demanded in the workplace. This will also remove the difficulty of having to identify prohibited GenAI use in assessment (Dobrin, 2023).
If there is a long time frame, for example, designing a new unit design, or where changes to an existing unit can be made in UCMS, this allows opportunity to rethink assessment in response to GenAI, including assessment type and Unit Learning Outcomes (ULOs). The modified ULOs where appropriate could also be designed to incorporate components of local or personal contexts and foster critical thinking and creativity. When you are incorporating ULO revisions as part of assessment redesign, you will need to ensure that the revision of the ULOs maintains academic rigour and relevance as well as alignment with the Course Learning Outcomes relevant to the level of study for the unit. Here is an example:
Original ULO across an undergraduate course (level 1, 2 or 3) | Revised ULO for a Level 1 unit (1000) | Revised ULO for a Level 2 or 3 unit (2000 or 3000) |
Discuss the effects of climate change on the environment. | Discuss the effects of climate change on the environment broadly and using recent local examples. Explanation: We've expanded the original ULO to include not only an understanding of climate change effects generally but also a discussion of recent local examples. The core knowledge of comprehending the effects of climate change on the environment remains unaltered, ensuring that foundational knowledge remains intact. But, by encouraging discussions of climate change within students' immediate surroundings, we not only enhance engagement and relevance but also introduce complexity that makes it challenging for students to accomplish this ULO using GenAI. The change to the ULO allows for the marking criteria to include the quality of discussion on the recent local examples. The changes are still consistent with an undergraduate foundational coursework unit (Level 1). | Critically discuss the effects of climate change on the environment broadly and using recent local contexts. Explanation: This revised ULO emphasises the higher-order requirement for the discussion. It challenges students to consider climate change effects in a deeper, critical manner generally and in recent local contexts. The focus is shifted towards higher-order cognitive skills, such as critical thinking, less likely to be automated by GenAI. In addition, students need to critically engage with recent local events and discuss these, something that GenAI would be less likely capable of doing well. The changes to the ULO also provide scope for higher weighting in the marking criteria for the ‘human’ components: critical thinking and current local contexts. The ULO reflects the level of study required for discipline-building or integrated coursework units (Level 2 or 3). |
Once your ULOs are ready, you can refer to the tables presented in the following section and select the cognitive level that aligns with the ULOs being measured and identify potential assessment task types. You can then consider the mitigation strategies you need to implement to address GenAI-related risks specific to your chosen cognitive level.
Designing assessment for GenAI resilience
Please refer to the following tables that are categorised based on the cognition level to be achieved (which should align to the ULOs being measured by the assessment).
Each table contains information on the suitable assessment types that could be used to measure the achievement of the cognition level, the conditions that may lead to AI breaches, and the potential mitigation strategies. This resource is adapted from University of Wollongong (2023) and UNSW (2023).
Understand - Demonstrate knowledge and comprehension
Suitable assessment types | Conditions that may lead to Academic Integrity breaches | Potential mitigation strategies |
---|---|---|
Quiz Short-answer questions Summary writing Identifying main ideas Checklists Glossary development | Reusing briefs/questions without modification from one term to another. Questions assessing lower order thinking skills (Bearman et al., 2023). |
For more information: 7 tips to make your quizzes AI-resilient this Spring - LX at UTS |
Apply - Transfer knowledge to practical situations
Suitable assessment types | Conditions that may lead to Academic Integrity breaches | Potential mitigation strategies |
---|---|---|
Case study Short written responses Problem-solving scenarios Simulations Role-playing Calculations Project proposal Research proposal Workbook Handouts Annotated bibliography Group work | Task that requires responses to common question or problems which can be addressed through simple searching or derived via GenAI (Bearman et al., 2023). Use of paraphrasing tools by students that limit plagiarism detectability. |
|
Analyse - Think critically and make judgements
Suitable assessment types | Conditions that may lead to Academic Integrity breaches | Potential mitigation strategies |
---|---|---|
Essay Blog/articles Discussion board Journal Literature review Project report Report Wiki Project-based assessments Debates Video presentations Industry live and static pitches/briefs | Task that centres around producing a single output or response. Use of paraphrasing tools by students that limit plagiarism detectability. Not using assessment formats that require critical thinking, personal insights and analysis makes it easier for students to rely solely on GenAI-generated content. |
|
Evaluate - Synthesise information, generate original ideas and design new solutions
Suitable assessment types | Conditions that may lead to Academic Integrity breaches | Potential mitigation strategies |
---|---|---|
Portfolio activities Presentations/poster presentations Reflections Digital storytelling Reflective journals Self-assessments/reviews Pitching ideas for change Reviewing ideas for change Debates Ethical dilemmas Critical review Oral assessments | Single output-focused task where students use GenAI (such as DALL-E, Midjourney) without permission to produce the artefact. Use of paraphrasing tools by students that limit plagiarism detectability. Not using assessment formats that require critical thinking, personal insights and analysis makes it easier for students to rely solely on GenAI-generated content. |
|
Create - Perform procedures and showcase techniques
Suitable assessment types | Conditions that may lead to Academic Integrity breaches | Potential mitigation strategies |
---|---|---|
Hand-on lab experiments Physical demonstration Role play or simulation Lab report Observation of real or simulated professional practice | Students might use pre-generated instructions or procedures obtained from GenAI tools or other sources, passing them off as their own work. Students might use GenAI tools or other external applications to guide their technique execution, bypassing the need for genuine skill demonstration. |
|
Quick tips for assessment design and GenAI
To minimise risk to academic rigour and academic integrity:
Consider using assessment tasks such as: | In the assessment task instructions: | In the rubric: |
Synchronous oral presentations where there is the opportunity for Q&A | Ask students to connect their responses to module content and materials, resources and readings. | Where possible, increase weighting for criteria requiring higher-order cognitive skills. |
Multi-modal submission | Use current events and local issues. | Where possible, minimise marks for content that can be readily achieved to an acceptable standard through GenAI. Have a go yourself to see what outputs can be generated. |
Face-to-face/ oral interactive assessment, e.g. interview, viva voce, practicals | Require students to use personal examples and experiences where there is evidence of the experience, for example, a critical reflection on placement journal. | Where possible, increase the weighting for criteria requiring input from personalised experiences and local recent events that can be substantiated. |
Innovative industry-based assessments where GenAI use is part of industry practice | Purposefully use GenAI as a tool for completing aspects of the assessment, for example:
| Where possible, allocate e-marks for process-based responses rather than just the output, i.e. the grade descriptors for the relevant criteria should reflect the student's explanation and rationalisation of the processes involved, not just the end product. |
An important note: With any assessment design, there needs to be clear alignment with the ULOs being measured. This includes alignment of the marking criteria to the ULOs. Grading needs to reflect the achievement of the ULOs being assessed.
Once you have considered the Design aspects of your assessment, continue with the rest of the assessment design process by clicking on Analyse, Act, Inform, Educate, Check and Evaluate.
Note: Given the rapidly evolving nature of GenAI technologies and largely opinion-based and low-level evidence on emerging practices for use in higher education, this resource represents the status quo at the time of writing (August 2023). As changes to policies and technology develop, and evidence for best practice emerges, practice recommendations outlined here are likely to continue to change and develop.