When I finetune Bert with simple_lm_finetuning.py, there seems an error:
"UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector."
Will it influence the performance of the finetuning process ? Thanks in advance for any suggestion.
It should be fine. There are probably your output losses.
@thomwolf Thanks for your reply. Could you explain more about why this happens ? I am still confused though
Maybe it is caused by calculating the loss in the model's forward function.
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.
I have the same problem in version 2.2.1, what can it be?
In my case, it doesn’t influence the results
On Mon, Jan 6, 2020 at 5:19 AM calusbr notifications@github.com wrote:
I have the same problem in version 2.2.1, what can it be?
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/huggingface/transformers/issues/852?email_source=notifications&email_token=AKRMVV4J4UT6HCXEEMOYBGDQ4MVV3A5CNFSM4IFUKP3KYY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEIFNJAI#issuecomment-571135105,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AKRMVV37AGO5AXJNPSGUVLTQ4MVV3ANCNFSM4IFUKP3A
.
In my case, the speed of training with 4 GPU is the same as 1 GPU. How solve the speed issues?
Most helpful comment
It should be fine. There are probably your output losses.