Toggle navigation
OpenReview
.net
Login
×
Go to
ICML 2022
homepage
Improving Language Models by Retrieving from Trillions of Tokens
Sebastian Borgeaud
,
Arthur Mensch
,
Jordan Hoffmann
,
Trevor Cai
,
Eliza Rutherford
,
Katie Millican
,
George van den Driessche
,
Jean-Baptiste Lespiau
,
Bogdan Damoc
,
Aidan Clark
,
Diego de Las Casas
,
Aurelia Guy
,
Jacob Menick
,
Roman Ring
,
Tom Hennigan
,
Saffron Huang
,
Loren Maggiore
,
Chris Jones
,
Albin Cassirer
,
Andy Brock
et al. (8 additional authors not shown)
2022 (modified: 24 Apr 2023)
ICML 2022
Readers:
Everyone
Abstract:
We enhance auto-regressive language models by conditioning on document chunks retrieved from a large corpus, based on local similarity with preceding tokens. With a 2 trillion token database, our R...
0 Replies
Loading