Ahmadzei's picture
added 3 more tables for large emb model
5fa1a76
Since umT5 was pre-trained in an unsupervised manner, there's no real advantage to using a task prefix during single-task
fine-tuning.