Ahmadzei's picture
added 3 more tables for large emb model
5fa1a76
It is often better to use the average instead of the default summation:
from transformers import AutoModelForCTC, TrainingArguments, Trainer
model = AutoModelForCTC.from_pretrained(
"facebook/wav2vec2-base",
ctc_loss_reduction="mean",
pad_token_id=processor.tokenizer.pad_token_id,
)
At this point, only three steps remain:
Define your training hyperparameters in [TrainingArguments].