What is: RAG?
Source | Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
Retriever-Augmented Generation, or RAG, is a type of language generation model that combines pre-trained parametric and non-parametric memory for language generation. Specifically, the parametric memory is a pre-trained seq2seq model and the non-parametric memory is a dense vector index of Wikipedia, accessed with a pre-trained neural retriever. For query , Maximum Inner Product Search (MIPS) is used to find the top-K documents . For final prediction , we treat as a latent variable and marginalize over seq2seq predictions given different documents.