from transformers import TrainingArguments | |
REPLACE THIS WITH YOUR REPO ID | |
repo_id = "MariaK/layoutlmv2-base-uncased_finetuned_docvqa" | |
training_args = TrainingArguments( | |
output_dir=repo_id, | |
per_device_train_batch_size=4, | |
num_train_epochs=20, | |
save_steps=200, | |
logging_steps=50, | |
evaluation_strategy="steps", | |
learning_rate=5e-5, | |
save_total_limit=2, | |
remove_unused_columns=False, | |
push_to_hub=True, | |
) | |
Define a simple data collator to batch examples together. |