Date
Publisher
arXiv
The integration of large language models (LLMs) into education offers
significant potential to enhance accessibility and engagement, yet their high
computational demands limit usability in low-resource settings, exacerbating
educational inequities. To address this, we propose an offline
Retrieval-Augmented Generation (RAG) pipeline that pairs a small language model
(SLM) with a robust retrieval mechanism, enabling factual, contextually
relevant responses without internet connectivity. We evaluate the efficacy of
this pipeline using domain-specific educational content, focusing on biology
coursework. Our analysis highlights key challenges: smaller models, such as
SmolLM, struggle to effectively leverage extended contexts provided by the RAG
pipeline, particularly when noisy or irrelevant chunks are included. To improve
performance, we propose exploring advanced chunking techniques, alternative
small or quantized versions of larger models, and moving beyond traditional
metrics like MMLU to a holistic evaluation framework assessing free-form
response. This work demonstrates the feasibility of deploying AI tutors in
constrained environments, laying the groundwork for equitable, offline, and
device-based educational tools.
What is the application?
Who is the user?
Who age?
Why use AI?
Study design
