Moreover, we show that with sufficient model size, cross-lingual pretraining can outperform | |
English-only pretraining when translating English speech into other languages, a setting which favors monolingual | |
pretraining. |
Moreover, we show that with sufficient model size, cross-lingual pretraining can outperform | |
English-only pretraining when translating English speech into other languages, a setting which favors monolingual | |
pretraining. |