Weijia Shi University of Washington, Sewon Min University of Washington, Michihiro Yasunaga Stanford University 3 KAIST, Minjoon Seo, Rich James Meta AI, Mike Lewis Meta AI, Luke Zettlemoyer University of Washington Meta AI, Wen-Tau Yih Meta AI (2023)
The paper introduces REPLUG, a retrieval-augmented language modeling framework that improves the performance of large language models (LLMs) by treating them as black boxes and integrating a tuneable retrieval model. Unlike prior approaches that modify the LMs, REPLUG simply prepends retrieved documents to the input context, allowing it to enhance any existing black-box LM. The authors demonstrate that this approach significantly improves the performance of models like GPT-3 and Codex, reporting enhancements of up to 6.3% and 5.1% on respective tasks. The proposed methodology also includes REPLUG LSR, a training mechanism that utilizes the language model to better adapt the retrieval strategy, further boosting performance on benchmarks such as MMLU. REPLUG has been shown to increase language modeling effectiveness and improve accuracy across different applications.
This paper employs the following methods:
The following datasets were used in this research:
The authors identified the following limitations: