Model I am using (Bert, XLNet....): GPT2
Language I am using the model on (English, Chinese....): English
The problem arise when using:
The tasks I am working on is:
convert_tf_checkpoint_to_pytorch.py
script to PyTorch. Running run_generation.py
from the examples folder results in an error. It seems like the tokenizer is not loaded from the converted model. (Maybe it is not saved?)Steps to reproduce the behavior:
python pytorch_transformers gpt2 path/to/checkpoint path/to/save/model
python run_generation.py --model_type gpt2 --model_name_or_path path/to/saved/model --top_p 0.9 --prompt "Hello Huggingface"
This results in the following error:
Traceback (most recent call last):
File "run_generation.py", line 195, in <module>
main()
File "run_generation.py", line 175, in main
context_tokens = tokenizer.encode(raw_text)
AttributeError: 'NoneType' object has no attribute 'encode'
Text generation like using "gpt2" as model_name_or_path
.
I manged to get it working by substituting the loading of the tokenizer with "gpt2", that way the tokenizer is loaded not from my fine-tuned model, but from the cache of the 117M version. Is the tokenizer actually trained?
Right now I have 3 files in the models folder: config.json
, pytorch_model.bin
and vocab.bpe
. Am I missing a file?
Hi, no the tokenizer is not trained. You can just load the original gpt2
one.
Shouldn't the tokenizer then be loaded from args.model_type
and not args.model_name_or_path
? Or do they differ from gpt2
to gpt2-medium
?
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.
Most helpful comment
Shouldn't the tokenizer then be loaded from
args.model_type
and notargs.model_name_or_path
? Or do they differ fromgpt2
togpt2-medium
?