EVENT DETAILS
AGENDA
SPEAKERS
REGISTER HERE
IN PERSON SEMINAR
Digital Innovation and Disruption in Educational Assessment: A realignment of humans and machines?
Join a group of assessment professionals and academic researchers in the field of educational and psychological assessments to discuss examples of innovative practice in educational assessment
Save the date
Tuesday 25th November 2025
4:30 – 6:00pm
Location
Pavilion Room, Hughes Hall
CB1 2EW
Event details
The space of educational assessment is being rapidly transformed with new techniques and technologies in digital data and artificial intelligence. This in-person seminar will present examples of innovative practice in digital educational assessments and consider their potential to re-align the relationship between humans and machines. Our speakers will illustrate and discuss new techniques that are currently being applied in assessment, involving new forms of digital automation, digital sensors, and AI-based models and agents.
This in-person seminar will involve three brief presentations followed by questions and discussion. It will not be recorded.
Agenda
16:30 – 16:35 Dr Martina Kuvalja: Welcome and introduction
16:35 - 16:55 Dr Peter Andrews and Dr Linhai Zhang: Developing a Human-Centred Virtual Examiner Assistant
Automated marking performs well for constrained response items such as multiple choice and short answer questions, but reliably assessing extended responses and questions involving multimedia remains challenging. Recent advances in computer vision and large language models (LLMs) now enable richer semantic and multimodal understanding, offering new possibilities for more transparent and trustworthy automation.
The EPSRC supported Prosperity project, undertaken in collaboration between AQA and King’s College London, aims to develop a virtual AI examiner assistant founded on a modular architecture. The project aims to provide a reliable and transparent baseline for quality assurance, while also modelling examiner cognition to identify where AI can assist rather than replace human judgement. Drawing on principles of Human Centered AI and distributed cognition, we explore how expert markers attend to, interpret, and integrate evidence across the marking process.
We will present early conceptual models for the virtual examiner assistant and preliminary findings from a behavioural eye gaze study with experienced markers. These findings inform the design of our modular AI system and guide the development of support tools that reduce ambiguity and cognitive load in high stakes decision making. Overall, Prosperity seeks to augment examiners by improving consistency, limiting bias, and enhancing transparency while preserving human agency and oversight.
16:55 - 17:15 Dr Britt He: Identifying Meaningful Patterns with Unstructured Log Data in Large-Scale Assessments
Computer-based assessments offer unprecedented opportunities to capture rich, granular log data through human–machine interactions, providing deeper insights into test-takers’ strategies and behaviors. This multidimensional sequential log data, encompassing actions, time intervals, eye-tracking, and more, poses challenges for traditional unidimensional sequence models. This talk presents an overview of sequence mining techniques applied to unstructured process data, with the goal of advancing item design and measurement frameworks.
Two case studies are highlighted to illustrate the application of sequence mining in educational assessments. The first explores missing response patterns in a computational thinking assessment using Dynamic Time Warping (DTW) method, identifying potential causes of nonresponse and predicting missing values based on process data. The second investigates behavioral features of resilient students in scientific inquiry tasks, applying the Time-Warped Longest Common Subsequence (T-WLCS), a method originally developed for musical retrieval—to uncover interpretable patterns in student interactions.
The presentation will also discuss the broader potential of integrating AI techniques with log data in large-scale assessments, exploring opportunities for more adaptive, inclusive, and behaviorally informed measurement systems.
17:15 - 17:35 Professor Bryan Maddox: Weak Signals and Wild Cards: An anticipatory framework for the validation of emergent immersive, interactive, adaptive and sensor-based assessments
Assessment theory and practice have developed in relation to particular contexts of practice, and with technological modes. In the context of new and emergent digital technologies, including sensor-based, process oriented, interactive and adaptive designs, and with computational psychometrics and artificial intelligence we ask: to what extent to the axioms of assessment theory and organising concepts and principles that used to underpin assessment design and validity practice hold in relation to emergent, and future assessment designs and assessment practice?
To answer we draw on methods and concepts in anticipatory futures thinking, and the influential field of French anthropology of technology (Andre Leroi-Gournan 1943, Gilbert Simondon, 2017 Bernard Steigler, 1994). The phrase – ‘weak signals and wild cards’ – indicates both the relevance of anticipatory futures think though. It imagines and considers radically different assessment possibilities and scenarios in rapidly changing and highly uncertain technological, ecological and political contexts and the increased significance of digital ‘sensor’ technologies.
17:35 – 18:00 Questions and discussion
Speakers
Organisers
Organiser: Digital Education Futures Initiative
Contact person: Judith Hannam
Email: jhh51@hughes.cam.ac.uk
Register here





