Search and Filter

Submit a research study

Contribute to the repository:

Add a paper

Technical – Computational

Ai-Driven Predictive Models For Optimizing Mathematics Education Technology: Enhancing Decision-Making Through Educational Data Mining And Meta-Analysis

This paper explores the challenge of achieving consistent effectiveness in integrating Mathematics Education Technology (MET) in K-12 classrooms, focusing on factors such as technology type, timing, and instructional strategies. It highlights the difficulties novice teachers face in optimizing MET compared to experienced educators, emphasizing the need to better understand the ideal duration and application of MET in various teaching settings. This study proposes using Artificial Intelligence (AI) to predict and optimize MET effectiveness, aiming to enhance student achievement.

Edumod-Llm: A Modular Approach For Designing Flexible And Transparent Educational Assistants

With the growing use of Large Language Model (LLM)-based Question-Answering (QA) systems in education, it is critical to evaluate their performance across individual pipeline components. In this work, we introduce {\model}, a modular function-calling LLM pipeline, and present a comprehensive evaluation along three key axes: function calling strategies, retrieval methods, and generative language models. Our framework enables fine-grained analysis by isolating and assessing each component.

Kidspeak: A General Multi-Purpose LLM For Kids' Speech Recognition And Screening

With the rapid advancement of conversational and diffusion-based AI, there is a growing adoption of AI in educational services, ranging from grading and assessment tools to personalized learning systems that provide targeted support for students. However, this adaptability has yet to fully extend to the domain of children's speech, where existing models often fail due to their reliance on datasets designed for clear, articulate adult speech.

Advisingwise: Supporting Academic Advising In Higher Education Settings Through A Human-In-The-Loop Multi-Agent Framework

Academic advising is critical to student success in higher education, yet high student-to-advisor ratios limit advisors' capacity to provide timely support, particularly during peak periods. Recent advances in Large Language Models (LLMs) present opportunities to enhance the advising process. We present AdvisingWise, a multi-agent system that automates time-consuming tasks, such as information retrieval and response drafting, while preserving human oversight.

Physicseval: Inference-Time Techniques To Improve The Reasoning Proficiency Of Large Language Models On Physics Problems

The discipline of physics stands as a cornerstone of human intellect, driving the evolution of technology and deepening our understanding of the fundamental principles of the cosmos. Contemporary literature includes some works centered on the task of solving physics problems - a crucial domain of natural language reasoning. In this paper, we evaluate the performance of frontier LLMs in solving physics problems, both mathematical and descriptive. We also employ a plethora of inference-time techniques and agentic frameworks to improve the performance of the models.

Llm Targeted Underperformance Disproportionately Impacts Vulnerable Users

While state-of-the-art large language models (LLMs) have shown impressive performance on many tasks, there has been extensive research on undesirable model behavior such as hallucinations and bias. In this work, we investigate how the quality of LLM responses changes in terms of information accuracy, truthfulness, and refusals depending on three user traits: English proficiency, education level, and country of origin. We present extensive experimentation on three state-of-the-art LLMs and two different datasets targeting truthfulness and factuality.

Pedagogy-Driven Evaluation Of Generative Ai-Powered Intelligent Tutoring Systems

The interdisciplinary research domain of Artificial Intelligence in Education (AIED) has a long history of developing Intelligent Tutoring Systems (ITSs) by integrating insights from technological advancements, educational theories, and cognitive psychology. The remarkable success of generative AI (GenAI) models has accelerated the development of large language model (LLM)-powered ITSs, which have potential to imitate human-like, pedagogically rich, and cognitively demanding tutoring.

Pustak Ai: Curriculum-Aligned And Interactive Textbooks Using Large Language Models

Large Language Models (LLMs) have demonstrated remarkable capabilities in understanding and generating human-like content. This has revolutionized various sectors such as healthcare, software development, and education. In education, LLMs offer potential for personalized and interactive learning experiences, especially in regions with limited teaching resources.

Ai-Enabled Grading With Near-Domain Data For Scaling Feedback With Human-Level Accuracy

Constructed-response questions are crucial to encourage generative processing and test a learner's understanding of core concepts. However, the limited availability of instructor time, large class sizes, and other resource constraints pose significant challenges in providing timely and detailed evaluation, which is crucial for a holistic educational experience. In addition, providing timely and frequent assessments is challenging since manual grading is labor intensive, and automated grading is complex to generalize to every possible response scenario.

Consistently Simulating Human Personas With Multi-Turn Reinforcement Learning

Large Language Models (LLMs) are increasingly used to simulate human users in interactive settings such as therapy, education, and social role-play. While these simulations enable scalable training and evaluation of AI agents, off-the-shelf LLMs often drift from their assigned personas, contradict earlier statements, or abandon role-appropriate behavior. We introduce a unified framework for evaluating and improving persona consistency in LLM-generated dialogue.