As a result, | |
the pre-trained BERT model can be fine-tuned with just one additional output layer to create state-of-the-art models | |
for a wide range of tasks, such as question answering and language inference, without substantial task-specific | |
architecture modifications. |