D 2026

Measuring the Impact of Student Gaming Behaviors on Learner Modeling

LIU, Qinyi; Lin LI; Valdemar ŠVÁBENSKÝ; Conrad BORCHERS; Mohammad KHALIL et al.

Základní údaje

Originální název

Measuring the Impact of Student Gaming Behaviors on Learner Modeling

Autoři

LIU, Qinyi; Lin LI; Valdemar ŠVÁBENSKÝ; Conrad BORCHERS a Mohammad KHALIL

Vydání

New York, NY, USA, Proceedings of the 16th Learning Analytics and Knowledge Conference (LAK '26), 11 s. 2026

Nakladatel

ACM

Další údaje

Jazyk

angličtina

Typ výsledku

Stať ve sborníku

Stát vydavatele

Spojené státy

Utajení

není předmětem státního či obchodního tajemství

Forma vydání

elektronická verze "online"

Odkazy

URL, URL

Označené pro přenos do RIV

Ne

Organizace

Fakulta informatiky – Masarykova univerzita – Repozitář

DOI

https://doi.org/10.1145/3785022.3785036

Klíčová slova anglicky

Adaptive Learning System; Knowledge Tracing; Data Poisoning Attacks; Adversarial Machine Learning; Gaming the System

Návaznosti

GN25-15839I, projekt VaV.
Změněno: 16. 1. 2026 00:51, RNDr. Daniel Jakubík

Anotace

V originále

The expansion of large-scale online education platforms has yielded vast amounts of student interaction data for knowledge tracing (KT). KT models estimate students’ concept mastery from interaction data, but the models' performance is sensitive to input data quality. Gaming behaviors, such as excessive hint use, may misrepresent students’ knowledge and undermine model reliability. However, systematic investigations of how different types of gaming behaviors affect KT remain scarce, and existing studies rely on costly manual analysis that does not capture behavioral diversity. In this study, we conceptualize gaming behaviors as a form of data poisoning, defined as the deliberate submission of incorrect or misleading interaction data to corrupt a model’s learning process. We design Data Poisoning Attacks (DPA) to simulate diverse gaming patterns and systematically evaluate their impact on KT model performance. Moreover, drawing on advances in DPA detection, we explore unsupervised approaches to enhance the generalizability of gaming behavior detection. We find that KT models performance tend to decrease especially for random guess behaviors. Our findings provide insights into the vulnerabilities of KT models and highlight the potential of adversarial methods for improving the robustness of learning analytics systems.
Zobrazeno: 2. 5. 2026 23:05