-
Notifications
You must be signed in to change notification settings - Fork 64
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training with multiple GPUs #164
Conversation
I've extended the support to the other scripts of the sequence labelling. For example in previous of my tests, the |
Thank you @lfoppiano ! I was not able to test with a multi GPU settings so just tested with normal single GPU, which is working fine as expected. I think this is useful as you say for larger models (keeping same batch size), but also for prediction because we can increase the batch size and process more rapidly texts. |
Doing more tests, training is fine but there is failure when writing a model with the
without Of course, using |
@kermitt2 thanks for testing it. I will add the option for the inference too. |
The |
This PR adds the support to multi-gpu, it has been tested on multiple GPU on the same node (4 x 16Gb) allowing to provide a larger batch size.
I wanted to implement it in the trainer but the processing related to the data preparation require to to be under the
with strategy.scope():
.I implemented only on the sequence labelling, once it's revised I update the classification as well.