Ahmadzei's picture
added 3 more tables for large emb model
5fa1a76
raw
history blame contribute delete
288 Bytes
To capture knowledge in a more modular and interpretable way, we
augment language model pre-training with a latent knowledge retriever, which allows the model to retrieve and attend
over documents from a large corpus such as Wikipedia, used during pre-training, fine-tuning and inference.