I would like to finetune facebook/mbart-large-cc25
on my data using pre-training tasks, in particular Masked Language Modeling (MLM).
How can I do that in HuggingFace?
Edit: rewrote the question for the sake of clarity
I would like to finetune facebook/mbart-large-cc25
on my data using pre-training tasks, in particular Masked Language Modeling (MLM).
How can I do that in HuggingFace?
Edit: rewrote the question for the sake of clarity
Since you are doing everything in HuggingFace, fine-tuning a model on pre-training tasks (assuming that pre-training task is provided in Huggingface) is pretty much the same for most models. What tasks are you interested in fine-tuning mBART on?
Hugginface provides extensive documentation for several fine-tuning tasks. For instance the links provided below will help you fine tune HF models for Language modelling, MNLI, SQuAD etc. https://huggingface.co/transformers/v2.0.0/examples.html and https://huggingface.co/transformers/training.html