TY - JOUR
T1 - In-Context Retrieval-Augmented Language Models
AU - Ram, Ori
AU - Levine, Yoav
AU - Dalmedigos, Itay
AU - Muhlgay, Dor
AU - Shashua, Amnon
AU - Leyton-Brown, Kevin
AU - Shoham, Yoav
N1 - Publisher Copyright:
© 2023, MIT Press Journals. All rights reserved.
PY - 2023
Y1 - 2023
N2 - Retrieval-Augmented Language Modeling (RALM) methods, which condition a language model (LM) on relevant documents from a grounding corpus during generation, were shown to significantly improve language modeling performance. In addition, they can mitigate the problem of factually inac-curate text generation and provide natural source attribution mechanism. Existing RALM approaches focus on modifying the LM architecture in order to facilitate the incorporation of external information, significantly complicating deployment. This paper considers a simple alternative, which we dub In-Context RALM: leaving the LM architecture unchanged and prepending grounding documents to the input, without any further training of the LM. We show that In-Context RALM that builds on off-the-shelf general purpose retrievers provides surpris-ingly large LM gains across model sizes and diverse corpora. We also demonstrate that the document retrieval and ranking mechanism can be specialized to the RALM setting to further boost performance. We conclude that In-Context RALM has considerable potential to increase the prevalence of LM grounding, particularly in settings where a pretrained LM must be used without modification or even via API access.1.
AB - Retrieval-Augmented Language Modeling (RALM) methods, which condition a language model (LM) on relevant documents from a grounding corpus during generation, were shown to significantly improve language modeling performance. In addition, they can mitigate the problem of factually inac-curate text generation and provide natural source attribution mechanism. Existing RALM approaches focus on modifying the LM architecture in order to facilitate the incorporation of external information, significantly complicating deployment. This paper considers a simple alternative, which we dub In-Context RALM: leaving the LM architecture unchanged and prepending grounding documents to the input, without any further training of the LM. We show that In-Context RALM that builds on off-the-shelf general purpose retrievers provides surpris-ingly large LM gains across model sizes and diverse corpora. We also demonstrate that the document retrieval and ranking mechanism can be specialized to the RALM setting to further boost performance. We conclude that In-Context RALM has considerable potential to increase the prevalence of LM grounding, particularly in settings where a pretrained LM must be used without modification or even via API access.1.
UR - http://www.scopus.com/inward/record.url?scp=85178620160&partnerID=8YFLogxK
U2 - 10.1162/tacl_a_00605
DO - 10.1162/tacl_a_00605
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
AN - SCOPUS:85178620160
SN - 2307-387X
VL - 11
SP - 1316
EP - 1331
JO - Transactions of the Association for Computational Linguistics
JF - Transactions of the Association for Computational Linguistics
ER -