I'm pretty sure PyTorch does CPU parallelization by default.
@severinsimmler Thank you for your reply. Is there any way to check for making sure of CPU parallelization while training in Flair?
You could start training and check with e.g. htop if more than one CPU is used.
What about multiple GPUs? PyTorch supports that, so how much is needed to make Flair support it?
There are several ways of adding multi-GPU support. One is with the built-in DataParallel classes of PyTorch (see #848), the other is using Horovod (see #859). We are still evaluating which is the best way to go here, but currently we tend to favor horovod since this is less invasive to the rest of the code and has more functionality.
Automatic mixed precision was also added recently (see #934) and is in fact already merged to master. This gives huge speedups in training large language models on a single GPU.
This is about Flair language model or other models (e.g. classification)? Any updates?
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.
Most helpful comment
What about multiple GPUs? PyTorch supports that, so how much is needed to make Flair support it?