
Sign up to save your podcasts
Or
The sources describe RETRO (Retrieval-Enhanced Transformer), a language model that enhances its performance by retrieving information from a large database. RETRO uses a key-value store where keys are BERT embeddings of text chunks and values are the text chunks themselves. When processing input, it retrieves similar text chunks from the database to augment the input, allowing it to perform comparably to much larger models. By incorporating this retrieved information through a chunked cross-attention mechanism, RETRO reduces the need to memorize facts and improves its performance on knowledge-intensive tasks. The database contains trillions of tokens.
5
22 ratings
The sources describe RETRO (Retrieval-Enhanced Transformer), a language model that enhances its performance by retrieving information from a large database. RETRO uses a key-value store where keys are BERT embeddings of text chunks and values are the text chunks themselves. When processing input, it retrieves similar text chunks from the database to augment the input, allowing it to perform comparably to much larger models. By incorporating this retrieved information through a chunked cross-attention mechanism, RETRO reduces the need to memorize facts and improves its performance on knowledge-intensive tasks. The database contains trillions of tokens.
272 Listeners
441 Listeners
298 Listeners
331 Listeners
217 Listeners
156 Listeners
192 Listeners
9,189 Listeners
417 Listeners
121 Listeners
75 Listeners
479 Listeners
94 Listeners
31 Listeners
43 Listeners