Hi,
For traning of any seq2seq model say Bart :-
from transformers import DataCollatorForSeq2Seq
data_collator = DataCollatorForSeq2Seq(tokenizer, model=model)
trainer = Seq2SeqTrainer(
model,
args,
train_dataset=tokenized_datasets["train"],
eval_dataset=tokenized_datasets["validation"],
data_collator=data_collator,
tokenizer=tokenizer,
compute_metrics=compute_metrics
)
Why DataCollatorForSeq2Seq Takes model as parameter when it is also provided to Seq2SeqTrainer.
What exactly it does with the tokenizer and model as input?
I appreciate your help.
Thank You.