Date
Publisher
arXiv
As artificial intelligence tools become ubiquitous in education, maintaining
academic integrity while accommodating pedagogically beneficial AI assistance
presents unprecedented challenges. Current AI detection systems fail to control
false positive rates (FPR) and suffer from bias against minority student
groups, prompting institutional suspensions of these technologies. Watermarking
techniques offer statistical rigor through precise $p$-values but remain
untested in educational contexts where students may use varying levels of
permitted AI edits. We present the first adaptation of watermarking-based
detection methods for classroom settings, introducing conformal methods that
effectively control FPR across diverse classroom settings. Using essays from
native and non-native English speakers, we simulate seven levels of AI editing
interventions--from grammar correction to content expansion--across multiple
language models and watermarking schemes, and evaluate our proposal under these
different setups. Our findings provide educators with quantitative frameworks
to enforce academic integrity standards while embracing AI integration in the
classroom.
What is the application?
Who is the user?
Who age?
Why use AI?
Study design
