ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.05276
44
0

Enhancing LLM-Based Short Answer Grading with Retrieval-Augmented Generation

7 April 2025
Yucheng Chu
Peng He
Hang Li
Haoyu Han
Kaiqi Yang
Yu Xue
Tingting Li
Joseph Krajcik
Jiliang Tang
    AI4Ed
ArXivPDFHTML
Abstract

Short answer assessment is a vital component of science education, allowing evaluation of students' complex three-dimensional understanding. Large language models (LLMs) that possess human-like ability in linguistic tasks are increasingly popular in assisting human graders to reduce their workload. However, LLMs' limitations in domain knowledge restrict their understanding in task-specific requirements and hinder their ability to achieve satisfactory performance. Retrieval-augmented generation (RAG) emerges as a promising solution by enabling LLMs to access relevant domain-specific knowledge during assessment. In this work, we propose an adaptive RAG framework for automated grading that dynamically retrieves and incorporates domain-specific knowledge based on the question and student answer context. Our approach combines semantic search and curated educational sources to retrieve valuable reference materials. Experimental results in a science education dataset demonstrate that our system achieves an improvement in grading accuracy compared to baseline LLM approaches. The findings suggest that RAG-enhanced grading systems can serve as reliable support with efficient performance gains.

View on arXiv
@article{chu2025_2504.05276,
  title={ Enhancing LLM-Based Short Answer Grading with Retrieval-Augmented Generation },
  author={ Yucheng Chu and Peng He and Hang Li and Haoyu Han and Kaiqi Yang and Yu Xue and Tingting Li and Joseph Krajcik and Jiliang Tang },
  journal={arXiv preprint arXiv:2504.05276},
  year={ 2025 }
}
Comments on this paper