Training Large Language Models for Retrieval-Augmented Question Answering through Backtracking Correction

0citations
Project
0
Citations
#1957
in ICLR 2025
of 3827 papers
4
Authors
3
Data Points

Abstract

Despite recent progress in Retrieval-Augmented Generation (RAG) achieved by large language models (LLMs), retrievers often recall uncorrelated documents, regarded as "noise" during subsequent text generation. To address this, some methods train LLMs to distinguish between relevant and irrelevant documents using labeled data, enabling them to select the most likely relevant ones as context. However, they remain sensitive to noise, as LLMs can easily make mistakes when the selected document is noisy. Some approaches increase the number of referenced documents and train LLMs to perform stepwise reasoning when presented with multiple documents. Unfortunately, these methods rely on extensive and diverse annotations to ensure generalization, which is both challenging and costly. In this paper, we proposeBacktracking Correctionto address these limitations. Specifically, we reformulate stepwise RAG into a multi-step decision-making process. Starting from the final step, we optimize the model through error sampling and self-correction, and then backtrack to the previous state iteratively. In this way, the model's learning scheme follows an easy-to-hard progression: as the target state moves forward, the context space decreases while the decision space increases. Experimental results demonstrate thatBacktracking Correctionenhances LLMs' ability to make complex multi-step assessments, improving the robustness of RAG in dealing with noisy documents.

Citation History

Jan 26, 2026
0
Jan 27, 2026
0
Jan 27, 2026
0