Date
Publisher
arXiv
The increasing availability and use of artificial intelligence (AI) tools in
educational settings has raised concerns about students' overreliance on these
technologies. Overreliance occurs when individuals accept incorrect
AI-generated recommendations, often without critical evaluation, leading to
flawed problem solutions and undermining learning outcomes. This study
investigates potential factors contributing to patterns of AI reliance among
undergraduate students, examining not only overreliance but also appropriate
reliance (correctly accepting helpful and rejecting harmful recommendations)
and underreliance (incorrectly rejecting helpful recommendations). Our approach
combined pre- and post-surveys with a controlled experimental task where
participants solved programming problems with an AI assistant that provided
both accurate and deliberately incorrect suggestions, allowing direct
observation of students' reliance patterns when faced with varying AI
reliability. We find that appropriate reliance is significantly related to
students' programming self-efficacy, programming literacy, and need for
cognition, while showing negative correlations with post-task trust and
satisfaction. Overreliance showed significant correlations with post-task trust
and satisfaction with the AI assistant. Underreliance was negatively correlated
with programming literacy, programming self-efficacy, and need for cognition.
Overall, the findings provide insights for developing targeted interventions
that promote appropriate reliance on AI tools, with implications for the
integration of AI in curriculum and educational technologies.
What is the application?
Who is the user?
Who age?
Why use AI?
Study design
