Ahmadzei's picture
added 3 more tables for large emb model
5fa1a76
raw
history blame contribute delete
229 Bytes
In this work, we propose a method to pre-train a smaller general-purpose language representation
model, called DistilBERT, which can then be fine-tuned with good performances on a wide range of tasks like its larger
counterparts.