A link to original question on Stack Overflow:
Yes, you can use the run_language_modeling.py script for pre-training e.g. BERT from scratch:
https://huggingface.co/transformers/examples.html#language-model-training
(Just leave the model_name_or_path parameter empty for pre-training from scratch)
@stefan-it thank u.
@xealml Curious whether you managed to train XLNet? If so, any pointers you could share?
Most helpful comment
@xealml Curious whether you managed to train XLNet? If so, any pointers you could share?