I followed the fine-tuning example described in here: https://github.com/pytorch/fairseq/blob/master/examples/mbart/README.md
However I didn't manage to reproduce the results described in the paper for EN-RO translation.
Code from here: https://github.com/pytorch/fairseq/blob/master/examples/mbart/README.md
Modified training script only with these changes:
--memory-efficient-fp16
--max-sentences 8
--required-batch-size-multiple 8
Trained with described setup and got BLEU of 2.3 only.
CC @ngoyal2707
You should find #1758 useful (it’d be nice if they updated the documentation here).
Thanks @mjpost
I found that issue really useful and that's how I validated the fine-tuned model.
However the question remains, how to reproduce fine-tuning results for en-ro translation? (Meaning training the base model (mbart.CC25) on en-ro data)
cc @ngoyal2707, could you please synthesize the discussion in #1758 and update the README?
Most helpful comment
You should find #1758 useful (it’d be nice if they updated the documentation here).